Google Cloud Native is in preview. Google Cloud Classic is fully supported.
Google Cloud Native v0.32.0 published on Wednesday, Nov 29, 2023 by Pulumi
google-native.aiplatform/v1.getPipelineJob
Explore with Pulumi AI
Google Cloud Native is in preview. Google Cloud Classic is fully supported.
Google Cloud Native v0.32.0 published on Wednesday, Nov 29, 2023 by Pulumi
Gets a PipelineJob.
Using getPipelineJob
Two invocation forms are available. The direct form accepts plain arguments and either blocks until the result value is available, or returns a Promise-wrapped result. The output form accepts Input-wrapped arguments and returns an Output-wrapped result.
function getPipelineJob(args: GetPipelineJobArgs, opts?: InvokeOptions): Promise<GetPipelineJobResult>
function getPipelineJobOutput(args: GetPipelineJobOutputArgs, opts?: InvokeOptions): Output<GetPipelineJobResult>
def get_pipeline_job(location: Optional[str] = None,
pipeline_job_id: Optional[str] = None,
project: Optional[str] = None,
opts: Optional[InvokeOptions] = None) -> GetPipelineJobResult
def get_pipeline_job_output(location: Optional[pulumi.Input[str]] = None,
pipeline_job_id: Optional[pulumi.Input[str]] = None,
project: Optional[pulumi.Input[str]] = None,
opts: Optional[InvokeOptions] = None) -> Output[GetPipelineJobResult]
func LookupPipelineJob(ctx *Context, args *LookupPipelineJobArgs, opts ...InvokeOption) (*LookupPipelineJobResult, error)
func LookupPipelineJobOutput(ctx *Context, args *LookupPipelineJobOutputArgs, opts ...InvokeOption) LookupPipelineJobResultOutput
> Note: This function is named LookupPipelineJob
in the Go SDK.
public static class GetPipelineJob
{
public static Task<GetPipelineJobResult> InvokeAsync(GetPipelineJobArgs args, InvokeOptions? opts = null)
public static Output<GetPipelineJobResult> Invoke(GetPipelineJobInvokeArgs args, InvokeOptions? opts = null)
}
public static CompletableFuture<GetPipelineJobResult> getPipelineJob(GetPipelineJobArgs args, InvokeOptions options)
// Output-based functions aren't available in Java yet
fn::invoke:
function: google-native:aiplatform/v1:getPipelineJob
arguments:
# arguments dictionary
The following arguments are supported:
- Location string
- Pipeline
Job stringId - Project string
- Location string
- Pipeline
Job stringId - Project string
- location String
- pipeline
Job StringId - project String
- location string
- pipeline
Job stringId - project string
- location str
- pipeline_
job_ strid - project str
- location String
- pipeline
Job StringId - project String
getPipelineJob Result
The following output properties are available:
- Create
Time string - Pipeline creation time.
- Display
Name string - The display name of the Pipeline. The name can be up to 128 characters long and can consist of any UTF-8 characters.
- Encryption
Spec Pulumi.Google Native. Aiplatform. V1. Outputs. Google Cloud Aiplatform V1Encryption Spec Response - Customer-managed encryption key spec for a pipelineJob. If set, this PipelineJob and all of its sub-resources will be secured by this key.
- End
Time string - Pipeline end time.
- Error
Pulumi.
Google Native. Aiplatform. V1. Outputs. Google Rpc Status Response - The error that occurred during pipeline execution. Only populated when the pipeline's state is FAILED or CANCELLED.
- Job
Detail Pulumi.Google Native. Aiplatform. V1. Outputs. Google Cloud Aiplatform V1Pipeline Job Detail Response - The details of pipeline run. Not available in the list view.
- Labels Dictionary<string, string>
- The labels with user-defined metadata to organize PipelineJob. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels. Note there is some reserved label key for Vertex AI Pipelines. -
vertex-ai-pipelines-run-billing-id
, user set value will get overrided. - Name string
- The resource name of the PipelineJob.
- Network string
- The full name of the Compute Engine network to which the Pipeline Job's workload should be peered. For example,
projects/12345/global/networks/myVPC
. Format is of the formprojects/{project}/global/networks/{network}
. Where {project} is a project number, as in12345
, and {network} is a network name. Private services access must already be configured for the network. Pipeline job will apply the network configuration to the Google Cloud resources being launched, if applied, such as Vertex AI Training or Dataflow job. If left unspecified, the workload is not peered with any network. - Pipeline
Spec Dictionary<string, string> - The spec of the pipeline.
- Reserved
Ip List<string>Ranges - A list of names for the reserved ip ranges under the VPC network that can be used for this Pipeline Job's workload. If set, we will deploy the Pipeline Job's workload within the provided ip ranges. Otherwise, the job will be deployed to any ip ranges under the provided VPC network. Example: ['vertex-ai-ip-range'].
- Runtime
Config Pulumi.Google Native. Aiplatform. V1. Outputs. Google Cloud Aiplatform V1Pipeline Job Runtime Config Response - Runtime config of the pipeline.
- Schedule
Name string - The schedule resource name. Only returned if the Pipeline is created by Schedule API.
- Service
Account string - The service account that the pipeline workload runs as. If not specified, the Compute Engine default service account in the project will be used. See https://cloud.google.com/compute/docs/access/service-accounts#default_service_account Users starting the pipeline must have the
iam.serviceAccounts.actAs
permission on this service account. - Start
Time string - Pipeline start time.
- State string
- The detailed state of the job.
- Template
Metadata Pulumi.Google Native. Aiplatform. V1. Outputs. Google Cloud Aiplatform V1Pipeline Template Metadata Response - Pipeline template metadata. Will fill up fields if PipelineJob.template_uri is from supported template registry.
- Template
Uri string - A template uri from where the PipelineJob.pipeline_spec, if empty, will be downloaded. Currently, only uri from Vertex Template Registry & Gallery is supported. Reference to https://cloud.google.com/vertex-ai/docs/pipelines/create-pipeline-template.
- Update
Time string - Timestamp when this PipelineJob was most recently updated.
- Create
Time string - Pipeline creation time.
- Display
Name string - The display name of the Pipeline. The name can be up to 128 characters long and can consist of any UTF-8 characters.
- Encryption
Spec GoogleCloud Aiplatform V1Encryption Spec Response - Customer-managed encryption key spec for a pipelineJob. If set, this PipelineJob and all of its sub-resources will be secured by this key.
- End
Time string - Pipeline end time.
- Error
Google
Rpc Status Response - The error that occurred during pipeline execution. Only populated when the pipeline's state is FAILED or CANCELLED.
- Job
Detail GoogleCloud Aiplatform V1Pipeline Job Detail Response - The details of pipeline run. Not available in the list view.
- Labels map[string]string
- The labels with user-defined metadata to organize PipelineJob. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels. Note there is some reserved label key for Vertex AI Pipelines. -
vertex-ai-pipelines-run-billing-id
, user set value will get overrided. - Name string
- The resource name of the PipelineJob.
- Network string
- The full name of the Compute Engine network to which the Pipeline Job's workload should be peered. For example,
projects/12345/global/networks/myVPC
. Format is of the formprojects/{project}/global/networks/{network}
. Where {project} is a project number, as in12345
, and {network} is a network name. Private services access must already be configured for the network. Pipeline job will apply the network configuration to the Google Cloud resources being launched, if applied, such as Vertex AI Training or Dataflow job. If left unspecified, the workload is not peered with any network. - Pipeline
Spec map[string]string - The spec of the pipeline.
- Reserved
Ip []stringRanges - A list of names for the reserved ip ranges under the VPC network that can be used for this Pipeline Job's workload. If set, we will deploy the Pipeline Job's workload within the provided ip ranges. Otherwise, the job will be deployed to any ip ranges under the provided VPC network. Example: ['vertex-ai-ip-range'].
- Runtime
Config GoogleCloud Aiplatform V1Pipeline Job Runtime Config Response - Runtime config of the pipeline.
- Schedule
Name string - The schedule resource name. Only returned if the Pipeline is created by Schedule API.
- Service
Account string - The service account that the pipeline workload runs as. If not specified, the Compute Engine default service account in the project will be used. See https://cloud.google.com/compute/docs/access/service-accounts#default_service_account Users starting the pipeline must have the
iam.serviceAccounts.actAs
permission on this service account. - Start
Time string - Pipeline start time.
- State string
- The detailed state of the job.
- Template
Metadata GoogleCloud Aiplatform V1Pipeline Template Metadata Response - Pipeline template metadata. Will fill up fields if PipelineJob.template_uri is from supported template registry.
- Template
Uri string - A template uri from where the PipelineJob.pipeline_spec, if empty, will be downloaded. Currently, only uri from Vertex Template Registry & Gallery is supported. Reference to https://cloud.google.com/vertex-ai/docs/pipelines/create-pipeline-template.
- Update
Time string - Timestamp when this PipelineJob was most recently updated.
- create
Time String - Pipeline creation time.
- display
Name String - The display name of the Pipeline. The name can be up to 128 characters long and can consist of any UTF-8 characters.
- encryption
Spec GoogleCloud Aiplatform V1Encryption Spec Response - Customer-managed encryption key spec for a pipelineJob. If set, this PipelineJob and all of its sub-resources will be secured by this key.
- end
Time String - Pipeline end time.
- error
Google
Rpc Status Response - The error that occurred during pipeline execution. Only populated when the pipeline's state is FAILED or CANCELLED.
- job
Detail GoogleCloud Aiplatform V1Pipeline Job Detail Response - The details of pipeline run. Not available in the list view.
- labels Map<String,String>
- The labels with user-defined metadata to organize PipelineJob. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels. Note there is some reserved label key for Vertex AI Pipelines. -
vertex-ai-pipelines-run-billing-id
, user set value will get overrided. - name String
- The resource name of the PipelineJob.
- network String
- The full name of the Compute Engine network to which the Pipeline Job's workload should be peered. For example,
projects/12345/global/networks/myVPC
. Format is of the formprojects/{project}/global/networks/{network}
. Where {project} is a project number, as in12345
, and {network} is a network name. Private services access must already be configured for the network. Pipeline job will apply the network configuration to the Google Cloud resources being launched, if applied, such as Vertex AI Training or Dataflow job. If left unspecified, the workload is not peered with any network. - pipeline
Spec Map<String,String> - The spec of the pipeline.
- reserved
Ip List<String>Ranges - A list of names for the reserved ip ranges under the VPC network that can be used for this Pipeline Job's workload. If set, we will deploy the Pipeline Job's workload within the provided ip ranges. Otherwise, the job will be deployed to any ip ranges under the provided VPC network. Example: ['vertex-ai-ip-range'].
- runtime
Config GoogleCloud Aiplatform V1Pipeline Job Runtime Config Response - Runtime config of the pipeline.
- schedule
Name String - The schedule resource name. Only returned if the Pipeline is created by Schedule API.
- service
Account String - The service account that the pipeline workload runs as. If not specified, the Compute Engine default service account in the project will be used. See https://cloud.google.com/compute/docs/access/service-accounts#default_service_account Users starting the pipeline must have the
iam.serviceAccounts.actAs
permission on this service account. - start
Time String - Pipeline start time.
- state String
- The detailed state of the job.
- template
Metadata GoogleCloud Aiplatform V1Pipeline Template Metadata Response - Pipeline template metadata. Will fill up fields if PipelineJob.template_uri is from supported template registry.
- template
Uri String - A template uri from where the PipelineJob.pipeline_spec, if empty, will be downloaded. Currently, only uri from Vertex Template Registry & Gallery is supported. Reference to https://cloud.google.com/vertex-ai/docs/pipelines/create-pipeline-template.
- update
Time String - Timestamp when this PipelineJob was most recently updated.
- create
Time string - Pipeline creation time.
- display
Name string - The display name of the Pipeline. The name can be up to 128 characters long and can consist of any UTF-8 characters.
- encryption
Spec GoogleCloud Aiplatform V1Encryption Spec Response - Customer-managed encryption key spec for a pipelineJob. If set, this PipelineJob and all of its sub-resources will be secured by this key.
- end
Time string - Pipeline end time.
- error
Google
Rpc Status Response - The error that occurred during pipeline execution. Only populated when the pipeline's state is FAILED or CANCELLED.
- job
Detail GoogleCloud Aiplatform V1Pipeline Job Detail Response - The details of pipeline run. Not available in the list view.
- labels {[key: string]: string}
- The labels with user-defined metadata to organize PipelineJob. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels. Note there is some reserved label key for Vertex AI Pipelines. -
vertex-ai-pipelines-run-billing-id
, user set value will get overrided. - name string
- The resource name of the PipelineJob.
- network string
- The full name of the Compute Engine network to which the Pipeline Job's workload should be peered. For example,
projects/12345/global/networks/myVPC
. Format is of the formprojects/{project}/global/networks/{network}
. Where {project} is a project number, as in12345
, and {network} is a network name. Private services access must already be configured for the network. Pipeline job will apply the network configuration to the Google Cloud resources being launched, if applied, such as Vertex AI Training or Dataflow job. If left unspecified, the workload is not peered with any network. - pipeline
Spec {[key: string]: string} - The spec of the pipeline.
- reserved
Ip string[]Ranges - A list of names for the reserved ip ranges under the VPC network that can be used for this Pipeline Job's workload. If set, we will deploy the Pipeline Job's workload within the provided ip ranges. Otherwise, the job will be deployed to any ip ranges under the provided VPC network. Example: ['vertex-ai-ip-range'].
- runtime
Config GoogleCloud Aiplatform V1Pipeline Job Runtime Config Response - Runtime config of the pipeline.
- schedule
Name string - The schedule resource name. Only returned if the Pipeline is created by Schedule API.
- service
Account string - The service account that the pipeline workload runs as. If not specified, the Compute Engine default service account in the project will be used. See https://cloud.google.com/compute/docs/access/service-accounts#default_service_account Users starting the pipeline must have the
iam.serviceAccounts.actAs
permission on this service account. - start
Time string - Pipeline start time.
- state string
- The detailed state of the job.
- template
Metadata GoogleCloud Aiplatform V1Pipeline Template Metadata Response - Pipeline template metadata. Will fill up fields if PipelineJob.template_uri is from supported template registry.
- template
Uri string - A template uri from where the PipelineJob.pipeline_spec, if empty, will be downloaded. Currently, only uri from Vertex Template Registry & Gallery is supported. Reference to https://cloud.google.com/vertex-ai/docs/pipelines/create-pipeline-template.
- update
Time string - Timestamp when this PipelineJob was most recently updated.
- create_
time str - Pipeline creation time.
- display_
name str - The display name of the Pipeline. The name can be up to 128 characters long and can consist of any UTF-8 characters.
- encryption_
spec GoogleCloud Aiplatform V1Encryption Spec Response - Customer-managed encryption key spec for a pipelineJob. If set, this PipelineJob and all of its sub-resources will be secured by this key.
- end_
time str - Pipeline end time.
- error
Google
Rpc Status Response - The error that occurred during pipeline execution. Only populated when the pipeline's state is FAILED or CANCELLED.
- job_
detail GoogleCloud Aiplatform V1Pipeline Job Detail Response - The details of pipeline run. Not available in the list view.
- labels Mapping[str, str]
- The labels with user-defined metadata to organize PipelineJob. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels. Note there is some reserved label key for Vertex AI Pipelines. -
vertex-ai-pipelines-run-billing-id
, user set value will get overrided. - name str
- The resource name of the PipelineJob.
- network str
- The full name of the Compute Engine network to which the Pipeline Job's workload should be peered. For example,
projects/12345/global/networks/myVPC
. Format is of the formprojects/{project}/global/networks/{network}
. Where {project} is a project number, as in12345
, and {network} is a network name. Private services access must already be configured for the network. Pipeline job will apply the network configuration to the Google Cloud resources being launched, if applied, such as Vertex AI Training or Dataflow job. If left unspecified, the workload is not peered with any network. - pipeline_
spec Mapping[str, str] - The spec of the pipeline.
- reserved_
ip_ Sequence[str]ranges - A list of names for the reserved ip ranges under the VPC network that can be used for this Pipeline Job's workload. If set, we will deploy the Pipeline Job's workload within the provided ip ranges. Otherwise, the job will be deployed to any ip ranges under the provided VPC network. Example: ['vertex-ai-ip-range'].
- runtime_
config GoogleCloud Aiplatform V1Pipeline Job Runtime Config Response - Runtime config of the pipeline.
- schedule_
name str - The schedule resource name. Only returned if the Pipeline is created by Schedule API.
- service_
account str - The service account that the pipeline workload runs as. If not specified, the Compute Engine default service account in the project will be used. See https://cloud.google.com/compute/docs/access/service-accounts#default_service_account Users starting the pipeline must have the
iam.serviceAccounts.actAs
permission on this service account. - start_
time str - Pipeline start time.
- state str
- The detailed state of the job.
- template_
metadata GoogleCloud Aiplatform V1Pipeline Template Metadata Response - Pipeline template metadata. Will fill up fields if PipelineJob.template_uri is from supported template registry.
- template_
uri str - A template uri from where the PipelineJob.pipeline_spec, if empty, will be downloaded. Currently, only uri from Vertex Template Registry & Gallery is supported. Reference to https://cloud.google.com/vertex-ai/docs/pipelines/create-pipeline-template.
- update_
time str - Timestamp when this PipelineJob was most recently updated.
- create
Time String - Pipeline creation time.
- display
Name String - The display name of the Pipeline. The name can be up to 128 characters long and can consist of any UTF-8 characters.
- encryption
Spec Property Map - Customer-managed encryption key spec for a pipelineJob. If set, this PipelineJob and all of its sub-resources will be secured by this key.
- end
Time String - Pipeline end time.
- error Property Map
- The error that occurred during pipeline execution. Only populated when the pipeline's state is FAILED or CANCELLED.
- job
Detail Property Map - The details of pipeline run. Not available in the list view.
- labels Map<String>
- The labels with user-defined metadata to organize PipelineJob. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels. Note there is some reserved label key for Vertex AI Pipelines. -
vertex-ai-pipelines-run-billing-id
, user set value will get overrided. - name String
- The resource name of the PipelineJob.
- network String
- The full name of the Compute Engine network to which the Pipeline Job's workload should be peered. For example,
projects/12345/global/networks/myVPC
. Format is of the formprojects/{project}/global/networks/{network}
. Where {project} is a project number, as in12345
, and {network} is a network name. Private services access must already be configured for the network. Pipeline job will apply the network configuration to the Google Cloud resources being launched, if applied, such as Vertex AI Training or Dataflow job. If left unspecified, the workload is not peered with any network. - pipeline
Spec Map<String> - The spec of the pipeline.
- reserved
Ip List<String>Ranges - A list of names for the reserved ip ranges under the VPC network that can be used for this Pipeline Job's workload. If set, we will deploy the Pipeline Job's workload within the provided ip ranges. Otherwise, the job will be deployed to any ip ranges under the provided VPC network. Example: ['vertex-ai-ip-range'].
- runtime
Config Property Map - Runtime config of the pipeline.
- schedule
Name String - The schedule resource name. Only returned if the Pipeline is created by Schedule API.
- service
Account String - The service account that the pipeline workload runs as. If not specified, the Compute Engine default service account in the project will be used. See https://cloud.google.com/compute/docs/access/service-accounts#default_service_account Users starting the pipeline must have the
iam.serviceAccounts.actAs
permission on this service account. - start
Time String - Pipeline start time.
- state String
- The detailed state of the job.
- template
Metadata Property Map - Pipeline template metadata. Will fill up fields if PipelineJob.template_uri is from supported template registry.
- template
Uri String - A template uri from where the PipelineJob.pipeline_spec, if empty, will be downloaded. Currently, only uri from Vertex Template Registry & Gallery is supported. Reference to https://cloud.google.com/vertex-ai/docs/pipelines/create-pipeline-template.
- update
Time String - Timestamp when this PipelineJob was most recently updated.
Supporting Types
GoogleCloudAiplatformV1ContextResponse
- Create
Time string - Timestamp when this Context was created.
- Description string
- Description of the Context
- Display
Name string - User provided display name of the Context. May be up to 128 Unicode characters.
- Etag string
- An eTag used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens.
- Labels Dictionary<string, string>
- The labels with user-defined metadata to organize your Contexts. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. No more than 64 user labels can be associated with one Context (System labels are excluded).
- Metadata Dictionary<string, string>
- Properties of the Context. Top level metadata keys' heading and trailing spaces will be trimmed. The size of this field should not exceed 200KB.
- Name string
- Immutable. The resource name of the Context.
- Parent
Contexts List<string> - A list of resource names of Contexts that are parents of this Context. A Context may have at most 10 parent_contexts.
- Schema
Title string - The title of the schema describing the metadata. Schema title and version is expected to be registered in earlier Create Schema calls. And both are used together as unique identifiers to identify schemas within the local metadata store.
- Schema
Version string - The version of the schema in schema_name to use. Schema title and version is expected to be registered in earlier Create Schema calls. And both are used together as unique identifiers to identify schemas within the local metadata store.
- Update
Time string - Timestamp when this Context was last updated.
- Create
Time string - Timestamp when this Context was created.
- Description string
- Description of the Context
- Display
Name string - User provided display name of the Context. May be up to 128 Unicode characters.
- Etag string
- An eTag used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens.
- Labels map[string]string
- The labels with user-defined metadata to organize your Contexts. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. No more than 64 user labels can be associated with one Context (System labels are excluded).
- Metadata map[string]string
- Properties of the Context. Top level metadata keys' heading and trailing spaces will be trimmed. The size of this field should not exceed 200KB.
- Name string
- Immutable. The resource name of the Context.
- Parent
Contexts []string - A list of resource names of Contexts that are parents of this Context. A Context may have at most 10 parent_contexts.
- Schema
Title string - The title of the schema describing the metadata. Schema title and version is expected to be registered in earlier Create Schema calls. And both are used together as unique identifiers to identify schemas within the local metadata store.
- Schema
Version string - The version of the schema in schema_name to use. Schema title and version is expected to be registered in earlier Create Schema calls. And both are used together as unique identifiers to identify schemas within the local metadata store.
- Update
Time string - Timestamp when this Context was last updated.
- create
Time String - Timestamp when this Context was created.
- description String
- Description of the Context
- display
Name String - User provided display name of the Context. May be up to 128 Unicode characters.
- etag String
- An eTag used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens.
- labels Map<String,String>
- The labels with user-defined metadata to organize your Contexts. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. No more than 64 user labels can be associated with one Context (System labels are excluded).
- metadata Map<String,String>
- Properties of the Context. Top level metadata keys' heading and trailing spaces will be trimmed. The size of this field should not exceed 200KB.
- name String
- Immutable. The resource name of the Context.
- parent
Contexts List<String> - A list of resource names of Contexts that are parents of this Context. A Context may have at most 10 parent_contexts.
- schema
Title String - The title of the schema describing the metadata. Schema title and version is expected to be registered in earlier Create Schema calls. And both are used together as unique identifiers to identify schemas within the local metadata store.
- schema
Version String - The version of the schema in schema_name to use. Schema title and version is expected to be registered in earlier Create Schema calls. And both are used together as unique identifiers to identify schemas within the local metadata store.
- update
Time String - Timestamp when this Context was last updated.
- create
Time string - Timestamp when this Context was created.
- description string
- Description of the Context
- display
Name string - User provided display name of the Context. May be up to 128 Unicode characters.
- etag string
- An eTag used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens.
- labels {[key: string]: string}
- The labels with user-defined metadata to organize your Contexts. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. No more than 64 user labels can be associated with one Context (System labels are excluded).
- metadata {[key: string]: string}
- Properties of the Context. Top level metadata keys' heading and trailing spaces will be trimmed. The size of this field should not exceed 200KB.
- name string
- Immutable. The resource name of the Context.
- parent
Contexts string[] - A list of resource names of Contexts that are parents of this Context. A Context may have at most 10 parent_contexts.
- schema
Title string - The title of the schema describing the metadata. Schema title and version is expected to be registered in earlier Create Schema calls. And both are used together as unique identifiers to identify schemas within the local metadata store.
- schema
Version string - The version of the schema in schema_name to use. Schema title and version is expected to be registered in earlier Create Schema calls. And both are used together as unique identifiers to identify schemas within the local metadata store.
- update
Time string - Timestamp when this Context was last updated.
- create_
time str - Timestamp when this Context was created.
- description str
- Description of the Context
- display_
name str - User provided display name of the Context. May be up to 128 Unicode characters.
- etag str
- An eTag used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens.
- labels Mapping[str, str]
- The labels with user-defined metadata to organize your Contexts. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. No more than 64 user labels can be associated with one Context (System labels are excluded).
- metadata Mapping[str, str]
- Properties of the Context. Top level metadata keys' heading and trailing spaces will be trimmed. The size of this field should not exceed 200KB.
- name str
- Immutable. The resource name of the Context.
- parent_
contexts Sequence[str] - A list of resource names of Contexts that are parents of this Context. A Context may have at most 10 parent_contexts.
- schema_
title str - The title of the schema describing the metadata. Schema title and version is expected to be registered in earlier Create Schema calls. And both are used together as unique identifiers to identify schemas within the local metadata store.
- schema_
version str - The version of the schema in schema_name to use. Schema title and version is expected to be registered in earlier Create Schema calls. And both are used together as unique identifiers to identify schemas within the local metadata store.
- update_
time str - Timestamp when this Context was last updated.
- create
Time String - Timestamp when this Context was created.
- description String
- Description of the Context
- display
Name String - User provided display name of the Context. May be up to 128 Unicode characters.
- etag String
- An eTag used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens.
- labels Map<String>
- The labels with user-defined metadata to organize your Contexts. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. No more than 64 user labels can be associated with one Context (System labels are excluded).
- metadata Map<String>
- Properties of the Context. Top level metadata keys' heading and trailing spaces will be trimmed. The size of this field should not exceed 200KB.
- name String
- Immutable. The resource name of the Context.
- parent
Contexts List<String> - A list of resource names of Contexts that are parents of this Context. A Context may have at most 10 parent_contexts.
- schema
Title String - The title of the schema describing the metadata. Schema title and version is expected to be registered in earlier Create Schema calls. And both are used together as unique identifiers to identify schemas within the local metadata store.
- schema
Version String - The version of the schema in schema_name to use. Schema title and version is expected to be registered in earlier Create Schema calls. And both are used together as unique identifiers to identify schemas within the local metadata store.
- update
Time String - Timestamp when this Context was last updated.
GoogleCloudAiplatformV1EncryptionSpecResponse
- Kms
Key stringName - The Cloud KMS resource identifier of the customer managed encryption key used to protect a resource. Has the form:
projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key
. The key needs to be in the same region as where the compute resource is created.
- Kms
Key stringName - The Cloud KMS resource identifier of the customer managed encryption key used to protect a resource. Has the form:
projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key
. The key needs to be in the same region as where the compute resource is created.
- kms
Key StringName - The Cloud KMS resource identifier of the customer managed encryption key used to protect a resource. Has the form:
projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key
. The key needs to be in the same region as where the compute resource is created.
- kms
Key stringName - The Cloud KMS resource identifier of the customer managed encryption key used to protect a resource. Has the form:
projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key
. The key needs to be in the same region as where the compute resource is created.
- kms_
key_ strname - The Cloud KMS resource identifier of the customer managed encryption key used to protect a resource. Has the form:
projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key
. The key needs to be in the same region as where the compute resource is created.
- kms
Key StringName - The Cloud KMS resource identifier of the customer managed encryption key used to protect a resource. Has the form:
projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key
. The key needs to be in the same region as where the compute resource is created.
GoogleCloudAiplatformV1ExecutionResponse
- Create
Time string - Timestamp when this Execution was created.
- Description string
- Description of the Execution
- Display
Name string - User provided display name of the Execution. May be up to 128 Unicode characters.
- Etag string
- An eTag used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens.
- Labels Dictionary<string, string>
- The labels with user-defined metadata to organize your Executions. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. No more than 64 user labels can be associated with one Execution (System labels are excluded).
- Metadata Dictionary<string, string>
- Properties of the Execution. Top level metadata keys' heading and trailing spaces will be trimmed. The size of this field should not exceed 200KB.
- Name string
- The resource name of the Execution.
- Schema
Title string - The title of the schema describing the metadata. Schema title and version is expected to be registered in earlier Create Schema calls. And both are used together as unique identifiers to identify schemas within the local metadata store.
- Schema
Version string - The version of the schema in
schema_title
to use. Schema title and version is expected to be registered in earlier Create Schema calls. And both are used together as unique identifiers to identify schemas within the local metadata store. - State string
- The state of this Execution. This is a property of the Execution, and does not imply or capture any ongoing process. This property is managed by clients (such as Vertex AI Pipelines) and the system does not prescribe or check the validity of state transitions.
- Update
Time string - Timestamp when this Execution was last updated.
- Create
Time string - Timestamp when this Execution was created.
- Description string
- Description of the Execution
- Display
Name string - User provided display name of the Execution. May be up to 128 Unicode characters.
- Etag string
- An eTag used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens.
- Labels map[string]string
- The labels with user-defined metadata to organize your Executions. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. No more than 64 user labels can be associated with one Execution (System labels are excluded).
- Metadata map[string]string
- Properties of the Execution. Top level metadata keys' heading and trailing spaces will be trimmed. The size of this field should not exceed 200KB.
- Name string
- The resource name of the Execution.
- Schema
Title string - The title of the schema describing the metadata. Schema title and version is expected to be registered in earlier Create Schema calls. And both are used together as unique identifiers to identify schemas within the local metadata store.
- Schema
Version string - The version of the schema in
schema_title
to use. Schema title and version is expected to be registered in earlier Create Schema calls. And both are used together as unique identifiers to identify schemas within the local metadata store. - State string
- The state of this Execution. This is a property of the Execution, and does not imply or capture any ongoing process. This property is managed by clients (such as Vertex AI Pipelines) and the system does not prescribe or check the validity of state transitions.
- Update
Time string - Timestamp when this Execution was last updated.
- create
Time String - Timestamp when this Execution was created.
- description String
- Description of the Execution
- display
Name String - User provided display name of the Execution. May be up to 128 Unicode characters.
- etag String
- An eTag used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens.
- labels Map<String,String>
- The labels with user-defined metadata to organize your Executions. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. No more than 64 user labels can be associated with one Execution (System labels are excluded).
- metadata Map<String,String>
- Properties of the Execution. Top level metadata keys' heading and trailing spaces will be trimmed. The size of this field should not exceed 200KB.
- name String
- The resource name of the Execution.
- schema
Title String - The title of the schema describing the metadata. Schema title and version is expected to be registered in earlier Create Schema calls. And both are used together as unique identifiers to identify schemas within the local metadata store.
- schema
Version String - The version of the schema in
schema_title
to use. Schema title and version is expected to be registered in earlier Create Schema calls. And both are used together as unique identifiers to identify schemas within the local metadata store. - state String
- The state of this Execution. This is a property of the Execution, and does not imply or capture any ongoing process. This property is managed by clients (such as Vertex AI Pipelines) and the system does not prescribe or check the validity of state transitions.
- update
Time String - Timestamp when this Execution was last updated.
- create
Time string - Timestamp when this Execution was created.
- description string
- Description of the Execution
- display
Name string - User provided display name of the Execution. May be up to 128 Unicode characters.
- etag string
- An eTag used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens.
- labels {[key: string]: string}
- The labels with user-defined metadata to organize your Executions. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. No more than 64 user labels can be associated with one Execution (System labels are excluded).
- metadata {[key: string]: string}
- Properties of the Execution. Top level metadata keys' heading and trailing spaces will be trimmed. The size of this field should not exceed 200KB.
- name string
- The resource name of the Execution.
- schema
Title string - The title of the schema describing the metadata. Schema title and version is expected to be registered in earlier Create Schema calls. And both are used together as unique identifiers to identify schemas within the local metadata store.
- schema
Version string - The version of the schema in
schema_title
to use. Schema title and version is expected to be registered in earlier Create Schema calls. And both are used together as unique identifiers to identify schemas within the local metadata store. - state string
- The state of this Execution. This is a property of the Execution, and does not imply or capture any ongoing process. This property is managed by clients (such as Vertex AI Pipelines) and the system does not prescribe or check the validity of state transitions.
- update
Time string - Timestamp when this Execution was last updated.
- create_
time str - Timestamp when this Execution was created.
- description str
- Description of the Execution
- display_
name str - User provided display name of the Execution. May be up to 128 Unicode characters.
- etag str
- An eTag used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens.
- labels Mapping[str, str]
- The labels with user-defined metadata to organize your Executions. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. No more than 64 user labels can be associated with one Execution (System labels are excluded).
- metadata Mapping[str, str]
- Properties of the Execution. Top level metadata keys' heading and trailing spaces will be trimmed. The size of this field should not exceed 200KB.
- name str
- The resource name of the Execution.
- schema_
title str - The title of the schema describing the metadata. Schema title and version is expected to be registered in earlier Create Schema calls. And both are used together as unique identifiers to identify schemas within the local metadata store.
- schema_
version str - The version of the schema in
schema_title
to use. Schema title and version is expected to be registered in earlier Create Schema calls. And both are used together as unique identifiers to identify schemas within the local metadata store. - state str
- The state of this Execution. This is a property of the Execution, and does not imply or capture any ongoing process. This property is managed by clients (such as Vertex AI Pipelines) and the system does not prescribe or check the validity of state transitions.
- update_
time str - Timestamp when this Execution was last updated.
- create
Time String - Timestamp when this Execution was created.
- description String
- Description of the Execution
- display
Name String - User provided display name of the Execution. May be up to 128 Unicode characters.
- etag String
- An eTag used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens.
- labels Map<String>
- The labels with user-defined metadata to organize your Executions. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. No more than 64 user labels can be associated with one Execution (System labels are excluded).
- metadata Map<String>
- Properties of the Execution. Top level metadata keys' heading and trailing spaces will be trimmed. The size of this field should not exceed 200KB.
- name String
- The resource name of the Execution.
- schema
Title String - The title of the schema describing the metadata. Schema title and version is expected to be registered in earlier Create Schema calls. And both are used together as unique identifiers to identify schemas within the local metadata store.
- schema
Version String - The version of the schema in
schema_title
to use. Schema title and version is expected to be registered in earlier Create Schema calls. And both are used together as unique identifiers to identify schemas within the local metadata store. - state String
- The state of this Execution. This is a property of the Execution, and does not imply or capture any ongoing process. This property is managed by clients (such as Vertex AI Pipelines) and the system does not prescribe or check the validity of state transitions.
- update
Time String - Timestamp when this Execution was last updated.
GoogleCloudAiplatformV1PipelineJobDetailResponse
- Pipeline
Context Pulumi.Google Native. Aiplatform. V1. Inputs. Google Cloud Aiplatform V1Context Response - The context of the pipeline.
- Pipeline
Run Pulumi.Context Google Native. Aiplatform. V1. Inputs. Google Cloud Aiplatform V1Context Response - The context of the current pipeline run.
- Task
Details List<Pulumi.Google Native. Aiplatform. V1. Inputs. Google Cloud Aiplatform V1Pipeline Task Detail Response> - The runtime details of the tasks under the pipeline.
- Pipeline
Context GoogleCloud Aiplatform V1Context Response - The context of the pipeline.
- Pipeline
Run GoogleContext Cloud Aiplatform V1Context Response - The context of the current pipeline run.
- Task
Details []GoogleCloud Aiplatform V1Pipeline Task Detail Response - The runtime details of the tasks under the pipeline.
- pipeline
Context GoogleCloud Aiplatform V1Context Response - The context of the pipeline.
- pipeline
Run GoogleContext Cloud Aiplatform V1Context Response - The context of the current pipeline run.
- task
Details List<GoogleCloud Aiplatform V1Pipeline Task Detail Response> - The runtime details of the tasks under the pipeline.
- pipeline
Context GoogleCloud Aiplatform V1Context Response - The context of the pipeline.
- pipeline
Run GoogleContext Cloud Aiplatform V1Context Response - The context of the current pipeline run.
- task
Details GoogleCloud Aiplatform V1Pipeline Task Detail Response[] - The runtime details of the tasks under the pipeline.
- pipeline_
context GoogleCloud Aiplatform V1Context Response - The context of the pipeline.
- pipeline_
run_ Googlecontext Cloud Aiplatform V1Context Response - The context of the current pipeline run.
- task_
details Sequence[GoogleCloud Aiplatform V1Pipeline Task Detail Response] - The runtime details of the tasks under the pipeline.
- pipeline
Context Property Map - The context of the pipeline.
- pipeline
Run Property MapContext - The context of the current pipeline run.
- task
Details List<Property Map> - The runtime details of the tasks under the pipeline.
GoogleCloudAiplatformV1PipelineJobRuntimeConfigResponse
- Failure
Policy string - Represents the failure policy of a pipeline. Currently, the default of a pipeline is that the pipeline will continue to run until no more tasks can be executed, also known as PIPELINE_FAILURE_POLICY_FAIL_SLOW. However, if a pipeline is set to PIPELINE_FAILURE_POLICY_FAIL_FAST, it will stop scheduling any new tasks when a task has failed. Any scheduled tasks will continue to completion.
- Gcs
Output stringDirectory - A path in a Cloud Storage bucket, which will be treated as the root output directory of the pipeline. It is used by the system to generate the paths of output artifacts. The artifact paths are generated with a sub-path pattern
{job_id}/{task_id}/{output_key}
under the specified output directory. The service account specified in this pipeline must have thestorage.objects.get
andstorage.objects.create
permissions for this bucket. - Input
Artifacts Dictionary<string, string> - The runtime artifacts of the PipelineJob. The key will be the input artifact name and the value would be one of the InputArtifact.
- Parameter
Values Dictionary<string, string> - The runtime parameters of the PipelineJob. The parameters will be passed into PipelineJob.pipeline_spec to replace the placeholders at runtime. This field is used by pipelines built using
PipelineJob.pipeline_spec.schema_version
2.1.0, such as pipelines built using Kubeflow Pipelines SDK 1.9 or higher and the v2 DSL. - Parameters Dictionary<string, string>
- Deprecated. Use RuntimeConfig.parameter_values instead. The runtime parameters of the PipelineJob. The parameters will be passed into PipelineJob.pipeline_spec to replace the placeholders at runtime. This field is used by pipelines built using
PipelineJob.pipeline_spec.schema_version
2.0.0 or lower, such as pipelines built using Kubeflow Pipelines SDK 1.8 or lower.
- Failure
Policy string - Represents the failure policy of a pipeline. Currently, the default of a pipeline is that the pipeline will continue to run until no more tasks can be executed, also known as PIPELINE_FAILURE_POLICY_FAIL_SLOW. However, if a pipeline is set to PIPELINE_FAILURE_POLICY_FAIL_FAST, it will stop scheduling any new tasks when a task has failed. Any scheduled tasks will continue to completion.
- Gcs
Output stringDirectory - A path in a Cloud Storage bucket, which will be treated as the root output directory of the pipeline. It is used by the system to generate the paths of output artifacts. The artifact paths are generated with a sub-path pattern
{job_id}/{task_id}/{output_key}
under the specified output directory. The service account specified in this pipeline must have thestorage.objects.get
andstorage.objects.create
permissions for this bucket. - Input
Artifacts map[string]string - The runtime artifacts of the PipelineJob. The key will be the input artifact name and the value would be one of the InputArtifact.
- Parameter
Values map[string]string - The runtime parameters of the PipelineJob. The parameters will be passed into PipelineJob.pipeline_spec to replace the placeholders at runtime. This field is used by pipelines built using
PipelineJob.pipeline_spec.schema_version
2.1.0, such as pipelines built using Kubeflow Pipelines SDK 1.9 or higher and the v2 DSL. - Parameters map[string]string
- Deprecated. Use RuntimeConfig.parameter_values instead. The runtime parameters of the PipelineJob. The parameters will be passed into PipelineJob.pipeline_spec to replace the placeholders at runtime. This field is used by pipelines built using
PipelineJob.pipeline_spec.schema_version
2.0.0 or lower, such as pipelines built using Kubeflow Pipelines SDK 1.8 or lower.
- failure
Policy String - Represents the failure policy of a pipeline. Currently, the default of a pipeline is that the pipeline will continue to run until no more tasks can be executed, also known as PIPELINE_FAILURE_POLICY_FAIL_SLOW. However, if a pipeline is set to PIPELINE_FAILURE_POLICY_FAIL_FAST, it will stop scheduling any new tasks when a task has failed. Any scheduled tasks will continue to completion.
- gcs
Output StringDirectory - A path in a Cloud Storage bucket, which will be treated as the root output directory of the pipeline. It is used by the system to generate the paths of output artifacts. The artifact paths are generated with a sub-path pattern
{job_id}/{task_id}/{output_key}
under the specified output directory. The service account specified in this pipeline must have thestorage.objects.get
andstorage.objects.create
permissions for this bucket. - input
Artifacts Map<String,String> - The runtime artifacts of the PipelineJob. The key will be the input artifact name and the value would be one of the InputArtifact.
- parameter
Values Map<String,String> - The runtime parameters of the PipelineJob. The parameters will be passed into PipelineJob.pipeline_spec to replace the placeholders at runtime. This field is used by pipelines built using
PipelineJob.pipeline_spec.schema_version
2.1.0, such as pipelines built using Kubeflow Pipelines SDK 1.9 or higher and the v2 DSL. - parameters Map<String,String>
- Deprecated. Use RuntimeConfig.parameter_values instead. The runtime parameters of the PipelineJob. The parameters will be passed into PipelineJob.pipeline_spec to replace the placeholders at runtime. This field is used by pipelines built using
PipelineJob.pipeline_spec.schema_version
2.0.0 or lower, such as pipelines built using Kubeflow Pipelines SDK 1.8 or lower.
- failure
Policy string - Represents the failure policy of a pipeline. Currently, the default of a pipeline is that the pipeline will continue to run until no more tasks can be executed, also known as PIPELINE_FAILURE_POLICY_FAIL_SLOW. However, if a pipeline is set to PIPELINE_FAILURE_POLICY_FAIL_FAST, it will stop scheduling any new tasks when a task has failed. Any scheduled tasks will continue to completion.
- gcs
Output stringDirectory - A path in a Cloud Storage bucket, which will be treated as the root output directory of the pipeline. It is used by the system to generate the paths of output artifacts. The artifact paths are generated with a sub-path pattern
{job_id}/{task_id}/{output_key}
under the specified output directory. The service account specified in this pipeline must have thestorage.objects.get
andstorage.objects.create
permissions for this bucket. - input
Artifacts {[key: string]: string} - The runtime artifacts of the PipelineJob. The key will be the input artifact name and the value would be one of the InputArtifact.
- parameter
Values {[key: string]: string} - The runtime parameters of the PipelineJob. The parameters will be passed into PipelineJob.pipeline_spec to replace the placeholders at runtime. This field is used by pipelines built using
PipelineJob.pipeline_spec.schema_version
2.1.0, such as pipelines built using Kubeflow Pipelines SDK 1.9 or higher and the v2 DSL. - parameters {[key: string]: string}
- Deprecated. Use RuntimeConfig.parameter_values instead. The runtime parameters of the PipelineJob. The parameters will be passed into PipelineJob.pipeline_spec to replace the placeholders at runtime. This field is used by pipelines built using
PipelineJob.pipeline_spec.schema_version
2.0.0 or lower, such as pipelines built using Kubeflow Pipelines SDK 1.8 or lower.
- failure_
policy str - Represents the failure policy of a pipeline. Currently, the default of a pipeline is that the pipeline will continue to run until no more tasks can be executed, also known as PIPELINE_FAILURE_POLICY_FAIL_SLOW. However, if a pipeline is set to PIPELINE_FAILURE_POLICY_FAIL_FAST, it will stop scheduling any new tasks when a task has failed. Any scheduled tasks will continue to completion.
- gcs_
output_ strdirectory - A path in a Cloud Storage bucket, which will be treated as the root output directory of the pipeline. It is used by the system to generate the paths of output artifacts. The artifact paths are generated with a sub-path pattern
{job_id}/{task_id}/{output_key}
under the specified output directory. The service account specified in this pipeline must have thestorage.objects.get
andstorage.objects.create
permissions for this bucket. - input_
artifacts Mapping[str, str] - The runtime artifacts of the PipelineJob. The key will be the input artifact name and the value would be one of the InputArtifact.
- parameter_
values Mapping[str, str] - The runtime parameters of the PipelineJob. The parameters will be passed into PipelineJob.pipeline_spec to replace the placeholders at runtime. This field is used by pipelines built using
PipelineJob.pipeline_spec.schema_version
2.1.0, such as pipelines built using Kubeflow Pipelines SDK 1.9 or higher and the v2 DSL. - parameters Mapping[str, str]
- Deprecated. Use RuntimeConfig.parameter_values instead. The runtime parameters of the PipelineJob. The parameters will be passed into PipelineJob.pipeline_spec to replace the placeholders at runtime. This field is used by pipelines built using
PipelineJob.pipeline_spec.schema_version
2.0.0 or lower, such as pipelines built using Kubeflow Pipelines SDK 1.8 or lower.
- failure
Policy String - Represents the failure policy of a pipeline. Currently, the default of a pipeline is that the pipeline will continue to run until no more tasks can be executed, also known as PIPELINE_FAILURE_POLICY_FAIL_SLOW. However, if a pipeline is set to PIPELINE_FAILURE_POLICY_FAIL_FAST, it will stop scheduling any new tasks when a task has failed. Any scheduled tasks will continue to completion.
- gcs
Output StringDirectory - A path in a Cloud Storage bucket, which will be treated as the root output directory of the pipeline. It is used by the system to generate the paths of output artifacts. The artifact paths are generated with a sub-path pattern
{job_id}/{task_id}/{output_key}
under the specified output directory. The service account specified in this pipeline must have thestorage.objects.get
andstorage.objects.create
permissions for this bucket. - input
Artifacts Map<String> - The runtime artifacts of the PipelineJob. The key will be the input artifact name and the value would be one of the InputArtifact.
- parameter
Values Map<String> - The runtime parameters of the PipelineJob. The parameters will be passed into PipelineJob.pipeline_spec to replace the placeholders at runtime. This field is used by pipelines built using
PipelineJob.pipeline_spec.schema_version
2.1.0, such as pipelines built using Kubeflow Pipelines SDK 1.9 or higher and the v2 DSL. - parameters Map<String>
- Deprecated. Use RuntimeConfig.parameter_values instead. The runtime parameters of the PipelineJob. The parameters will be passed into PipelineJob.pipeline_spec to replace the placeholders at runtime. This field is used by pipelines built using
PipelineJob.pipeline_spec.schema_version
2.0.0 or lower, such as pipelines built using Kubeflow Pipelines SDK 1.8 or lower.
GoogleCloudAiplatformV1PipelineTaskDetailPipelineTaskStatusResponse
- Error
Pulumi.
Google Native. Aiplatform. V1. Inputs. Google Rpc Status Response - The error that occurred during the state. May be set when the state is any of the non-final state (PENDING/RUNNING/CANCELLING) or FAILED state. If the state is FAILED, the error here is final and not going to be retried. If the state is a non-final state, the error indicates a system-error being retried.
- State string
- The state of the task.
- Update
Time string - Update time of this status.
- Error
Google
Rpc Status Response - The error that occurred during the state. May be set when the state is any of the non-final state (PENDING/RUNNING/CANCELLING) or FAILED state. If the state is FAILED, the error here is final and not going to be retried. If the state is a non-final state, the error indicates a system-error being retried.
- State string
- The state of the task.
- Update
Time string - Update time of this status.
- error
Google
Rpc Status Response - The error that occurred during the state. May be set when the state is any of the non-final state (PENDING/RUNNING/CANCELLING) or FAILED state. If the state is FAILED, the error here is final and not going to be retried. If the state is a non-final state, the error indicates a system-error being retried.
- state String
- The state of the task.
- update
Time String - Update time of this status.
- error
Google
Rpc Status Response - The error that occurred during the state. May be set when the state is any of the non-final state (PENDING/RUNNING/CANCELLING) or FAILED state. If the state is FAILED, the error here is final and not going to be retried. If the state is a non-final state, the error indicates a system-error being retried.
- state string
- The state of the task.
- update
Time string - Update time of this status.
- error
Google
Rpc Status Response - The error that occurred during the state. May be set when the state is any of the non-final state (PENDING/RUNNING/CANCELLING) or FAILED state. If the state is FAILED, the error here is final and not going to be retried. If the state is a non-final state, the error indicates a system-error being retried.
- state str
- The state of the task.
- update_
time str - Update time of this status.
- error Property Map
- The error that occurred during the state. May be set when the state is any of the non-final state (PENDING/RUNNING/CANCELLING) or FAILED state. If the state is FAILED, the error here is final and not going to be retried. If the state is a non-final state, the error indicates a system-error being retried.
- state String
- The state of the task.
- update
Time String - Update time of this status.
GoogleCloudAiplatformV1PipelineTaskDetailResponse
- Create
Time string - Task create time.
- End
Time string - Task end time.
- Error
Pulumi.
Google Native. Aiplatform. V1. Inputs. Google Rpc Status Response - The error that occurred during task execution. Only populated when the task's state is FAILED or CANCELLED.
- Execution
Pulumi.
Google Native. Aiplatform. V1. Inputs. Google Cloud Aiplatform V1Execution Response - The execution metadata of the task.
- Executor
Detail Pulumi.Google Native. Aiplatform. V1. Inputs. Google Cloud Aiplatform V1Pipeline Task Executor Detail Response - The detailed execution info.
- Inputs Dictionary<string, string>
- The runtime input artifacts of the task.
- Outputs Dictionary<string, string>
- The runtime output artifacts of the task.
- Parent
Task stringId - The id of the parent task if the task is within a component scope. Empty if the task is at the root level.
- Pipeline
Task List<Pulumi.Status Google Native. Aiplatform. V1. Inputs. Google Cloud Aiplatform V1Pipeline Task Detail Pipeline Task Status Response> - A list of task status. This field keeps a record of task status evolving over time.
- Start
Time string - Task start time.
- State string
- State of the task.
- Task
Id string - The system generated ID of the task.
- Task
Name string - The user specified name of the task that is defined in pipeline_spec.
- Create
Time string - Task create time.
- End
Time string - Task end time.
- Error
Google
Rpc Status Response - The error that occurred during task execution. Only populated when the task's state is FAILED or CANCELLED.
- Execution
Google
Cloud Aiplatform V1Execution Response - The execution metadata of the task.
- Executor
Detail GoogleCloud Aiplatform V1Pipeline Task Executor Detail Response - The detailed execution info.
- Inputs map[string]string
- The runtime input artifacts of the task.
- Outputs map[string]string
- The runtime output artifacts of the task.
- Parent
Task stringId - The id of the parent task if the task is within a component scope. Empty if the task is at the root level.
- Pipeline
Task []GoogleStatus Cloud Aiplatform V1Pipeline Task Detail Pipeline Task Status Response - A list of task status. This field keeps a record of task status evolving over time.
- Start
Time string - Task start time.
- State string
- State of the task.
- Task
Id string - The system generated ID of the task.
- Task
Name string - The user specified name of the task that is defined in pipeline_spec.
- create
Time String - Task create time.
- end
Time String - Task end time.
- error
Google
Rpc Status Response - The error that occurred during task execution. Only populated when the task's state is FAILED or CANCELLED.
- execution
Google
Cloud Aiplatform V1Execution Response - The execution metadata of the task.
- executor
Detail GoogleCloud Aiplatform V1Pipeline Task Executor Detail Response - The detailed execution info.
- inputs Map<String,String>
- The runtime input artifacts of the task.
- outputs Map<String,String>
- The runtime output artifacts of the task.
- parent
Task StringId - The id of the parent task if the task is within a component scope. Empty if the task is at the root level.
- pipeline
Task List<GoogleStatus Cloud Aiplatform V1Pipeline Task Detail Pipeline Task Status Response> - A list of task status. This field keeps a record of task status evolving over time.
- start
Time String - Task start time.
- state String
- State of the task.
- task
Id String - The system generated ID of the task.
- task
Name String - The user specified name of the task that is defined in pipeline_spec.
- create
Time string - Task create time.
- end
Time string - Task end time.
- error
Google
Rpc Status Response - The error that occurred during task execution. Only populated when the task's state is FAILED or CANCELLED.
- execution
Google
Cloud Aiplatform V1Execution Response - The execution metadata of the task.
- executor
Detail GoogleCloud Aiplatform V1Pipeline Task Executor Detail Response - The detailed execution info.
- inputs {[key: string]: string}
- The runtime input artifacts of the task.
- outputs {[key: string]: string}
- The runtime output artifacts of the task.
- parent
Task stringId - The id of the parent task if the task is within a component scope. Empty if the task is at the root level.
- pipeline
Task GoogleStatus Cloud Aiplatform V1Pipeline Task Detail Pipeline Task Status Response[] - A list of task status. This field keeps a record of task status evolving over time.
- start
Time string - Task start time.
- state string
- State of the task.
- task
Id string - The system generated ID of the task.
- task
Name string - The user specified name of the task that is defined in pipeline_spec.
- create_
time str - Task create time.
- end_
time str - Task end time.
- error
Google
Rpc Status Response - The error that occurred during task execution. Only populated when the task's state is FAILED or CANCELLED.
- execution
Google
Cloud Aiplatform V1Execution Response - The execution metadata of the task.
- executor_
detail GoogleCloud Aiplatform V1Pipeline Task Executor Detail Response - The detailed execution info.
- inputs Mapping[str, str]
- The runtime input artifacts of the task.
- outputs Mapping[str, str]
- The runtime output artifacts of the task.
- parent_
task_ strid - The id of the parent task if the task is within a component scope. Empty if the task is at the root level.
- pipeline_
task_ Sequence[Googlestatus Cloud Aiplatform V1Pipeline Task Detail Pipeline Task Status Response] - A list of task status. This field keeps a record of task status evolving over time.
- start_
time str - Task start time.
- state str
- State of the task.
- task_
id str - The system generated ID of the task.
- task_
name str - The user specified name of the task that is defined in pipeline_spec.
- create
Time String - Task create time.
- end
Time String - Task end time.
- error Property Map
- The error that occurred during task execution. Only populated when the task's state is FAILED or CANCELLED.
- execution Property Map
- The execution metadata of the task.
- executor
Detail Property Map - The detailed execution info.
- inputs Map<String>
- The runtime input artifacts of the task.
- outputs Map<String>
- The runtime output artifacts of the task.
- parent
Task StringId - The id of the parent task if the task is within a component scope. Empty if the task is at the root level.
- pipeline
Task List<Property Map>Status - A list of task status. This field keeps a record of task status evolving over time.
- start
Time String - Task start time.
- state String
- State of the task.
- task
Id String - The system generated ID of the task.
- task
Name String - The user specified name of the task that is defined in pipeline_spec.
GoogleCloudAiplatformV1PipelineTaskExecutorDetailContainerDetailResponse
- Failed
Main List<string>Jobs - The names of the previously failed CustomJob for the main container executions. The list includes the all attempts in chronological order.
- Failed
Pre List<string>Caching Check Jobs - The names of the previously failed CustomJob for the pre-caching-check container executions. This job will be available if the PipelineJob.pipeline_spec specifies the
pre_caching_check
hook in the lifecycle events. The list includes the all attempts in chronological order. - Main
Job string - The name of the CustomJob for the main container execution.
- Pre
Caching stringCheck Job - The name of the CustomJob for the pre-caching-check container execution. This job will be available if the PipelineJob.pipeline_spec specifies the
pre_caching_check
hook in the lifecycle events.
- Failed
Main []stringJobs - The names of the previously failed CustomJob for the main container executions. The list includes the all attempts in chronological order.
- Failed
Pre []stringCaching Check Jobs - The names of the previously failed CustomJob for the pre-caching-check container executions. This job will be available if the PipelineJob.pipeline_spec specifies the
pre_caching_check
hook in the lifecycle events. The list includes the all attempts in chronological order. - Main
Job string - The name of the CustomJob for the main container execution.
- Pre
Caching stringCheck Job - The name of the CustomJob for the pre-caching-check container execution. This job will be available if the PipelineJob.pipeline_spec specifies the
pre_caching_check
hook in the lifecycle events.
- failed
Main List<String>Jobs - The names of the previously failed CustomJob for the main container executions. The list includes the all attempts in chronological order.
- failed
Pre List<String>Caching Check Jobs - The names of the previously failed CustomJob for the pre-caching-check container executions. This job will be available if the PipelineJob.pipeline_spec specifies the
pre_caching_check
hook in the lifecycle events. The list includes the all attempts in chronological order. - main
Job String - The name of the CustomJob for the main container execution.
- pre
Caching StringCheck Job - The name of the CustomJob for the pre-caching-check container execution. This job will be available if the PipelineJob.pipeline_spec specifies the
pre_caching_check
hook in the lifecycle events.
- failed
Main string[]Jobs - The names of the previously failed CustomJob for the main container executions. The list includes the all attempts in chronological order.
- failed
Pre string[]Caching Check Jobs - The names of the previously failed CustomJob for the pre-caching-check container executions. This job will be available if the PipelineJob.pipeline_spec specifies the
pre_caching_check
hook in the lifecycle events. The list includes the all attempts in chronological order. - main
Job string - The name of the CustomJob for the main container execution.
- pre
Caching stringCheck Job - The name of the CustomJob for the pre-caching-check container execution. This job will be available if the PipelineJob.pipeline_spec specifies the
pre_caching_check
hook in the lifecycle events.
- failed_
main_ Sequence[str]jobs - The names of the previously failed CustomJob for the main container executions. The list includes the all attempts in chronological order.
- failed_
pre_ Sequence[str]caching_ check_ jobs - The names of the previously failed CustomJob for the pre-caching-check container executions. This job will be available if the PipelineJob.pipeline_spec specifies the
pre_caching_check
hook in the lifecycle events. The list includes the all attempts in chronological order. - main_
job str - The name of the CustomJob for the main container execution.
- pre_
caching_ strcheck_ job - The name of the CustomJob for the pre-caching-check container execution. This job will be available if the PipelineJob.pipeline_spec specifies the
pre_caching_check
hook in the lifecycle events.
- failed
Main List<String>Jobs - The names of the previously failed CustomJob for the main container executions. The list includes the all attempts in chronological order.
- failed
Pre List<String>Caching Check Jobs - The names of the previously failed CustomJob for the pre-caching-check container executions. This job will be available if the PipelineJob.pipeline_spec specifies the
pre_caching_check
hook in the lifecycle events. The list includes the all attempts in chronological order. - main
Job String - The name of the CustomJob for the main container execution.
- pre
Caching StringCheck Job - The name of the CustomJob for the pre-caching-check container execution. This job will be available if the PipelineJob.pipeline_spec specifies the
pre_caching_check
hook in the lifecycle events.
GoogleCloudAiplatformV1PipelineTaskExecutorDetailCustomJobDetailResponse
- Failed
Jobs List<string> - The names of the previously failed CustomJob. The list includes the all attempts in chronological order.
- Job string
- The name of the CustomJob.
- Failed
Jobs []string - The names of the previously failed CustomJob. The list includes the all attempts in chronological order.
- Job string
- The name of the CustomJob.
- failed
Jobs List<String> - The names of the previously failed CustomJob. The list includes the all attempts in chronological order.
- job String
- The name of the CustomJob.
- failed
Jobs string[] - The names of the previously failed CustomJob. The list includes the all attempts in chronological order.
- job string
- The name of the CustomJob.
- failed_
jobs Sequence[str] - The names of the previously failed CustomJob. The list includes the all attempts in chronological order.
- job str
- The name of the CustomJob.
- failed
Jobs List<String> - The names of the previously failed CustomJob. The list includes the all attempts in chronological order.
- job String
- The name of the CustomJob.
GoogleCloudAiplatformV1PipelineTaskExecutorDetailResponse
- Container
Detail Pulumi.Google Native. Aiplatform. V1. Inputs. Google Cloud Aiplatform V1Pipeline Task Executor Detail Container Detail Response - The detailed info for a container executor.
- Custom
Job Pulumi.Detail Google Native. Aiplatform. V1. Inputs. Google Cloud Aiplatform V1Pipeline Task Executor Detail Custom Job Detail Response - The detailed info for a custom job executor.
- Container
Detail GoogleCloud Aiplatform V1Pipeline Task Executor Detail Container Detail Response - The detailed info for a container executor.
- Custom
Job GoogleDetail Cloud Aiplatform V1Pipeline Task Executor Detail Custom Job Detail Response - The detailed info for a custom job executor.
- container
Detail GoogleCloud Aiplatform V1Pipeline Task Executor Detail Container Detail Response - The detailed info for a container executor.
- custom
Job GoogleDetail Cloud Aiplatform V1Pipeline Task Executor Detail Custom Job Detail Response - The detailed info for a custom job executor.
- container
Detail GoogleCloud Aiplatform V1Pipeline Task Executor Detail Container Detail Response - The detailed info for a container executor.
- custom
Job GoogleDetail Cloud Aiplatform V1Pipeline Task Executor Detail Custom Job Detail Response - The detailed info for a custom job executor.
- container_
detail GoogleCloud Aiplatform V1Pipeline Task Executor Detail Container Detail Response - The detailed info for a container executor.
- custom_
job_ Googledetail Cloud Aiplatform V1Pipeline Task Executor Detail Custom Job Detail Response - The detailed info for a custom job executor.
- container
Detail Property Map - The detailed info for a container executor.
- custom
Job Property MapDetail - The detailed info for a custom job executor.
GoogleCloudAiplatformV1PipelineTemplateMetadataResponse
- Version string
- The version_name in artifact registry. Will always be presented in output if the PipelineJob.template_uri is from supported template registry. Format is "sha256:abcdef123456...".
- Version string
- The version_name in artifact registry. Will always be presented in output if the PipelineJob.template_uri is from supported template registry. Format is "sha256:abcdef123456...".
- version String
- The version_name in artifact registry. Will always be presented in output if the PipelineJob.template_uri is from supported template registry. Format is "sha256:abcdef123456...".
- version string
- The version_name in artifact registry. Will always be presented in output if the PipelineJob.template_uri is from supported template registry. Format is "sha256:abcdef123456...".
- version str
- The version_name in artifact registry. Will always be presented in output if the PipelineJob.template_uri is from supported template registry. Format is "sha256:abcdef123456...".
- version String
- The version_name in artifact registry. Will always be presented in output if the PipelineJob.template_uri is from supported template registry. Format is "sha256:abcdef123456...".
GoogleRpcStatusResponse
- Code int
- The status code, which should be an enum value of google.rpc.Code.
- Details
List<Immutable
Dictionary<string, string>> - A list of messages that carry the error details. There is a common set of message types for APIs to use.
- Message string
- A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
- Code int
- The status code, which should be an enum value of google.rpc.Code.
- Details []map[string]string
- A list of messages that carry the error details. There is a common set of message types for APIs to use.
- Message string
- A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
- code Integer
- The status code, which should be an enum value of google.rpc.Code.
- details List<Map<String,String>>
- A list of messages that carry the error details. There is a common set of message types for APIs to use.
- message String
- A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
- code number
- The status code, which should be an enum value of google.rpc.Code.
- details {[key: string]: string}[]
- A list of messages that carry the error details. There is a common set of message types for APIs to use.
- message string
- A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
- code int
- The status code, which should be an enum value of google.rpc.Code.
- details Sequence[Mapping[str, str]]
- A list of messages that carry the error details. There is a common set of message types for APIs to use.
- message str
- A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
- code Number
- The status code, which should be an enum value of google.rpc.Code.
- details List<Map<String>>
- A list of messages that carry the error details. There is a common set of message types for APIs to use.
- message String
- A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
Package Details
- Repository
- Google Cloud Native pulumi/pulumi-google-native
- License
- Apache-2.0
Google Cloud Native is in preview. Google Cloud Classic is fully supported.
Google Cloud Native v0.32.0 published on Wednesday, Nov 29, 2023 by Pulumi