Google Cloud Native is in preview. Google Cloud Classic is fully supported.
Google Cloud Native v0.32.0 published on Wednesday, Nov 29, 2023 by Pulumi
google-native.aiplatform/v1beta1.getDeploymentResourcePool
Explore with Pulumi AI
Google Cloud Native is in preview. Google Cloud Classic is fully supported.
Google Cloud Native v0.32.0 published on Wednesday, Nov 29, 2023 by Pulumi
Get a DeploymentResourcePool.
Using getDeploymentResourcePool
Two invocation forms are available. The direct form accepts plain arguments and either blocks until the result value is available, or returns a Promise-wrapped result. The output form accepts Input-wrapped arguments and returns an Output-wrapped result.
function getDeploymentResourcePool(args: GetDeploymentResourcePoolArgs, opts?: InvokeOptions): Promise<GetDeploymentResourcePoolResult>
function getDeploymentResourcePoolOutput(args: GetDeploymentResourcePoolOutputArgs, opts?: InvokeOptions): Output<GetDeploymentResourcePoolResult>
def get_deployment_resource_pool(deployment_resource_pool_id: Optional[str] = None,
location: Optional[str] = None,
project: Optional[str] = None,
opts: Optional[InvokeOptions] = None) -> GetDeploymentResourcePoolResult
def get_deployment_resource_pool_output(deployment_resource_pool_id: Optional[pulumi.Input[str]] = None,
location: Optional[pulumi.Input[str]] = None,
project: Optional[pulumi.Input[str]] = None,
opts: Optional[InvokeOptions] = None) -> Output[GetDeploymentResourcePoolResult]
func LookupDeploymentResourcePool(ctx *Context, args *LookupDeploymentResourcePoolArgs, opts ...InvokeOption) (*LookupDeploymentResourcePoolResult, error)
func LookupDeploymentResourcePoolOutput(ctx *Context, args *LookupDeploymentResourcePoolOutputArgs, opts ...InvokeOption) LookupDeploymentResourcePoolResultOutput
> Note: This function is named LookupDeploymentResourcePool
in the Go SDK.
public static class GetDeploymentResourcePool
{
public static Task<GetDeploymentResourcePoolResult> InvokeAsync(GetDeploymentResourcePoolArgs args, InvokeOptions? opts = null)
public static Output<GetDeploymentResourcePoolResult> Invoke(GetDeploymentResourcePoolInvokeArgs args, InvokeOptions? opts = null)
}
public static CompletableFuture<GetDeploymentResourcePoolResult> getDeploymentResourcePool(GetDeploymentResourcePoolArgs args, InvokeOptions options)
// Output-based functions aren't available in Java yet
fn::invoke:
function: google-native:aiplatform/v1beta1:getDeploymentResourcePool
arguments:
# arguments dictionary
The following arguments are supported:
- Deployment
Resource stringPool Id - Location string
- Project string
- Deployment
Resource stringPool Id - Location string
- Project string
- deployment
Resource StringPool Id - location String
- project String
- deployment
Resource stringPool Id - location string
- project string
- deployment_
resource_ strpool_ id - location str
- project str
- deployment
Resource StringPool Id - location String
- project String
getDeploymentResourcePool Result
The following output properties are available:
- Create
Time string - Timestamp when this DeploymentResourcePool was created.
- Dedicated
Resources Pulumi.Google Native. Aiplatform. V1Beta1. Outputs. Google Cloud Aiplatform V1beta1Dedicated Resources Response - The underlying DedicatedResources that the DeploymentResourcePool uses.
- Name string
- Immutable. The resource name of the DeploymentResourcePool. Format:
projects/{project}/locations/{location}/deploymentResourcePools/{deployment_resource_pool}
- Create
Time string - Timestamp when this DeploymentResourcePool was created.
- Dedicated
Resources GoogleCloud Aiplatform V1beta1Dedicated Resources Response - The underlying DedicatedResources that the DeploymentResourcePool uses.
- Name string
- Immutable. The resource name of the DeploymentResourcePool. Format:
projects/{project}/locations/{location}/deploymentResourcePools/{deployment_resource_pool}
- create
Time String - Timestamp when this DeploymentResourcePool was created.
- dedicated
Resources GoogleCloud Aiplatform V1beta1Dedicated Resources Response - The underlying DedicatedResources that the DeploymentResourcePool uses.
- name String
- Immutable. The resource name of the DeploymentResourcePool. Format:
projects/{project}/locations/{location}/deploymentResourcePools/{deployment_resource_pool}
- create
Time string - Timestamp when this DeploymentResourcePool was created.
- dedicated
Resources GoogleCloud Aiplatform V1beta1Dedicated Resources Response - The underlying DedicatedResources that the DeploymentResourcePool uses.
- name string
- Immutable. The resource name of the DeploymentResourcePool. Format:
projects/{project}/locations/{location}/deploymentResourcePools/{deployment_resource_pool}
- create_
time str - Timestamp when this DeploymentResourcePool was created.
- dedicated_
resources GoogleCloud Aiplatform V1beta1Dedicated Resources Response - The underlying DedicatedResources that the DeploymentResourcePool uses.
- name str
- Immutable. The resource name of the DeploymentResourcePool. Format:
projects/{project}/locations/{location}/deploymentResourcePools/{deployment_resource_pool}
- create
Time String - Timestamp when this DeploymentResourcePool was created.
- dedicated
Resources Property Map - The underlying DedicatedResources that the DeploymentResourcePool uses.
- name String
- Immutable. The resource name of the DeploymentResourcePool. Format:
projects/{project}/locations/{location}/deploymentResourcePools/{deployment_resource_pool}
Supporting Types
GoogleCloudAiplatformV1beta1AutoscalingMetricSpecResponse
- Metric
Name string - The resource metric name. Supported metrics: * For Online Prediction: *
aiplatform.googleapis.com/prediction/online/accelerator/duty_cycle
*aiplatform.googleapis.com/prediction/online/cpu/utilization
- Target int
- The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided.
- Metric
Name string - The resource metric name. Supported metrics: * For Online Prediction: *
aiplatform.googleapis.com/prediction/online/accelerator/duty_cycle
*aiplatform.googleapis.com/prediction/online/cpu/utilization
- Target int
- The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided.
- metric
Name String - The resource metric name. Supported metrics: * For Online Prediction: *
aiplatform.googleapis.com/prediction/online/accelerator/duty_cycle
*aiplatform.googleapis.com/prediction/online/cpu/utilization
- target Integer
- The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided.
- metric
Name string - The resource metric name. Supported metrics: * For Online Prediction: *
aiplatform.googleapis.com/prediction/online/accelerator/duty_cycle
*aiplatform.googleapis.com/prediction/online/cpu/utilization
- target number
- The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided.
- metric_
name str - The resource metric name. Supported metrics: * For Online Prediction: *
aiplatform.googleapis.com/prediction/online/accelerator/duty_cycle
*aiplatform.googleapis.com/prediction/online/cpu/utilization
- target int
- The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided.
- metric
Name String - The resource metric name. Supported metrics: * For Online Prediction: *
aiplatform.googleapis.com/prediction/online/accelerator/duty_cycle
*aiplatform.googleapis.com/prediction/online/cpu/utilization
- target Number
- The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided.
GoogleCloudAiplatformV1beta1DedicatedResourcesResponse
- Autoscaling
Metric List<Pulumi.Specs Google Native. Aiplatform. V1Beta1. Inputs. Google Cloud Aiplatform V1beta1Autoscaling Metric Spec Response> - Immutable. The metric specifications that overrides a resource utilization metric (CPU utilization, accelerator's duty cycle, and so on) target value (default to 60 if not set). At most one entry is allowed per metric. If machine_spec.accelerator_count is above 0, the autoscaling will be based on both CPU utilization and accelerator's duty cycle metrics and scale up when either metrics exceeds its target value while scale down if both metrics are under their target value. The default target value is 60 for both metrics. If machine_spec.accelerator_count is 0, the autoscaling will be based on CPU utilization metric only with default target value 60 if not explicitly set. For example, in the case of Online Prediction, if you want to override target CPU utilization to 80, you should set autoscaling_metric_specs.metric_name to
aiplatform.googleapis.com/prediction/online/cpu/utilization
and autoscaling_metric_specs.target to80
. - Machine
Spec Pulumi.Google Native. Aiplatform. V1Beta1. Inputs. Google Cloud Aiplatform V1beta1Machine Spec Response - Immutable. The specification of a single machine used by the prediction.
- Max
Replica intCount - Immutable. The maximum number of replicas this DeployedModel may be deployed on when the traffic against it increases. If the requested value is too large, the deployment will error, but if deployment succeeds then the ability to scale the model to that many replicas is guaranteed (barring service outages). If traffic against the DeployedModel increases beyond what its replicas at maximum may handle, a portion of the traffic will be dropped. If this value is not provided, will use min_replica_count as the default value. The value of this field impacts the charge against Vertex CPU and GPU quotas. Specifically, you will be charged for (max_replica_count * number of cores in the selected machine type) and (max_replica_count * number of GPUs per replica in the selected machine type).
- Min
Replica intCount - Immutable. The minimum number of machine replicas this DeployedModel will be always deployed on. This value must be greater than or equal to 1. If traffic against the DeployedModel increases, it may dynamically be deployed onto more replicas, and as traffic decreases, some of these extra replicas may be freed.
- Autoscaling
Metric []GoogleSpecs Cloud Aiplatform V1beta1Autoscaling Metric Spec Response - Immutable. The metric specifications that overrides a resource utilization metric (CPU utilization, accelerator's duty cycle, and so on) target value (default to 60 if not set). At most one entry is allowed per metric. If machine_spec.accelerator_count is above 0, the autoscaling will be based on both CPU utilization and accelerator's duty cycle metrics and scale up when either metrics exceeds its target value while scale down if both metrics are under their target value. The default target value is 60 for both metrics. If machine_spec.accelerator_count is 0, the autoscaling will be based on CPU utilization metric only with default target value 60 if not explicitly set. For example, in the case of Online Prediction, if you want to override target CPU utilization to 80, you should set autoscaling_metric_specs.metric_name to
aiplatform.googleapis.com/prediction/online/cpu/utilization
and autoscaling_metric_specs.target to80
. - Machine
Spec GoogleCloud Aiplatform V1beta1Machine Spec Response - Immutable. The specification of a single machine used by the prediction.
- Max
Replica intCount - Immutable. The maximum number of replicas this DeployedModel may be deployed on when the traffic against it increases. If the requested value is too large, the deployment will error, but if deployment succeeds then the ability to scale the model to that many replicas is guaranteed (barring service outages). If traffic against the DeployedModel increases beyond what its replicas at maximum may handle, a portion of the traffic will be dropped. If this value is not provided, will use min_replica_count as the default value. The value of this field impacts the charge against Vertex CPU and GPU quotas. Specifically, you will be charged for (max_replica_count * number of cores in the selected machine type) and (max_replica_count * number of GPUs per replica in the selected machine type).
- Min
Replica intCount - Immutable. The minimum number of machine replicas this DeployedModel will be always deployed on. This value must be greater than or equal to 1. If traffic against the DeployedModel increases, it may dynamically be deployed onto more replicas, and as traffic decreases, some of these extra replicas may be freed.
- autoscaling
Metric List<GoogleSpecs Cloud Aiplatform V1beta1Autoscaling Metric Spec Response> - Immutable. The metric specifications that overrides a resource utilization metric (CPU utilization, accelerator's duty cycle, and so on) target value (default to 60 if not set). At most one entry is allowed per metric. If machine_spec.accelerator_count is above 0, the autoscaling will be based on both CPU utilization and accelerator's duty cycle metrics and scale up when either metrics exceeds its target value while scale down if both metrics are under their target value. The default target value is 60 for both metrics. If machine_spec.accelerator_count is 0, the autoscaling will be based on CPU utilization metric only with default target value 60 if not explicitly set. For example, in the case of Online Prediction, if you want to override target CPU utilization to 80, you should set autoscaling_metric_specs.metric_name to
aiplatform.googleapis.com/prediction/online/cpu/utilization
and autoscaling_metric_specs.target to80
. - machine
Spec GoogleCloud Aiplatform V1beta1Machine Spec Response - Immutable. The specification of a single machine used by the prediction.
- max
Replica IntegerCount - Immutable. The maximum number of replicas this DeployedModel may be deployed on when the traffic against it increases. If the requested value is too large, the deployment will error, but if deployment succeeds then the ability to scale the model to that many replicas is guaranteed (barring service outages). If traffic against the DeployedModel increases beyond what its replicas at maximum may handle, a portion of the traffic will be dropped. If this value is not provided, will use min_replica_count as the default value. The value of this field impacts the charge against Vertex CPU and GPU quotas. Specifically, you will be charged for (max_replica_count * number of cores in the selected machine type) and (max_replica_count * number of GPUs per replica in the selected machine type).
- min
Replica IntegerCount - Immutable. The minimum number of machine replicas this DeployedModel will be always deployed on. This value must be greater than or equal to 1. If traffic against the DeployedModel increases, it may dynamically be deployed onto more replicas, and as traffic decreases, some of these extra replicas may be freed.
- autoscaling
Metric GoogleSpecs Cloud Aiplatform V1beta1Autoscaling Metric Spec Response[] - Immutable. The metric specifications that overrides a resource utilization metric (CPU utilization, accelerator's duty cycle, and so on) target value (default to 60 if not set). At most one entry is allowed per metric. If machine_spec.accelerator_count is above 0, the autoscaling will be based on both CPU utilization and accelerator's duty cycle metrics and scale up when either metrics exceeds its target value while scale down if both metrics are under their target value. The default target value is 60 for both metrics. If machine_spec.accelerator_count is 0, the autoscaling will be based on CPU utilization metric only with default target value 60 if not explicitly set. For example, in the case of Online Prediction, if you want to override target CPU utilization to 80, you should set autoscaling_metric_specs.metric_name to
aiplatform.googleapis.com/prediction/online/cpu/utilization
and autoscaling_metric_specs.target to80
. - machine
Spec GoogleCloud Aiplatform V1beta1Machine Spec Response - Immutable. The specification of a single machine used by the prediction.
- max
Replica numberCount - Immutable. The maximum number of replicas this DeployedModel may be deployed on when the traffic against it increases. If the requested value is too large, the deployment will error, but if deployment succeeds then the ability to scale the model to that many replicas is guaranteed (barring service outages). If traffic against the DeployedModel increases beyond what its replicas at maximum may handle, a portion of the traffic will be dropped. If this value is not provided, will use min_replica_count as the default value. The value of this field impacts the charge against Vertex CPU and GPU quotas. Specifically, you will be charged for (max_replica_count * number of cores in the selected machine type) and (max_replica_count * number of GPUs per replica in the selected machine type).
- min
Replica numberCount - Immutable. The minimum number of machine replicas this DeployedModel will be always deployed on. This value must be greater than or equal to 1. If traffic against the DeployedModel increases, it may dynamically be deployed onto more replicas, and as traffic decreases, some of these extra replicas may be freed.
- autoscaling_
metric_ Sequence[Googlespecs Cloud Aiplatform V1beta1Autoscaling Metric Spec Response] - Immutable. The metric specifications that overrides a resource utilization metric (CPU utilization, accelerator's duty cycle, and so on) target value (default to 60 if not set). At most one entry is allowed per metric. If machine_spec.accelerator_count is above 0, the autoscaling will be based on both CPU utilization and accelerator's duty cycle metrics and scale up when either metrics exceeds its target value while scale down if both metrics are under their target value. The default target value is 60 for both metrics. If machine_spec.accelerator_count is 0, the autoscaling will be based on CPU utilization metric only with default target value 60 if not explicitly set. For example, in the case of Online Prediction, if you want to override target CPU utilization to 80, you should set autoscaling_metric_specs.metric_name to
aiplatform.googleapis.com/prediction/online/cpu/utilization
and autoscaling_metric_specs.target to80
. - machine_
spec GoogleCloud Aiplatform V1beta1Machine Spec Response - Immutable. The specification of a single machine used by the prediction.
- max_
replica_ intcount - Immutable. The maximum number of replicas this DeployedModel may be deployed on when the traffic against it increases. If the requested value is too large, the deployment will error, but if deployment succeeds then the ability to scale the model to that many replicas is guaranteed (barring service outages). If traffic against the DeployedModel increases beyond what its replicas at maximum may handle, a portion of the traffic will be dropped. If this value is not provided, will use min_replica_count as the default value. The value of this field impacts the charge against Vertex CPU and GPU quotas. Specifically, you will be charged for (max_replica_count * number of cores in the selected machine type) and (max_replica_count * number of GPUs per replica in the selected machine type).
- min_
replica_ intcount - Immutable. The minimum number of machine replicas this DeployedModel will be always deployed on. This value must be greater than or equal to 1. If traffic against the DeployedModel increases, it may dynamically be deployed onto more replicas, and as traffic decreases, some of these extra replicas may be freed.
- autoscaling
Metric List<Property Map>Specs - Immutable. The metric specifications that overrides a resource utilization metric (CPU utilization, accelerator's duty cycle, and so on) target value (default to 60 if not set). At most one entry is allowed per metric. If machine_spec.accelerator_count is above 0, the autoscaling will be based on both CPU utilization and accelerator's duty cycle metrics and scale up when either metrics exceeds its target value while scale down if both metrics are under their target value. The default target value is 60 for both metrics. If machine_spec.accelerator_count is 0, the autoscaling will be based on CPU utilization metric only with default target value 60 if not explicitly set. For example, in the case of Online Prediction, if you want to override target CPU utilization to 80, you should set autoscaling_metric_specs.metric_name to
aiplatform.googleapis.com/prediction/online/cpu/utilization
and autoscaling_metric_specs.target to80
. - machine
Spec Property Map - Immutable. The specification of a single machine used by the prediction.
- max
Replica NumberCount - Immutable. The maximum number of replicas this DeployedModel may be deployed on when the traffic against it increases. If the requested value is too large, the deployment will error, but if deployment succeeds then the ability to scale the model to that many replicas is guaranteed (barring service outages). If traffic against the DeployedModel increases beyond what its replicas at maximum may handle, a portion of the traffic will be dropped. If this value is not provided, will use min_replica_count as the default value. The value of this field impacts the charge against Vertex CPU and GPU quotas. Specifically, you will be charged for (max_replica_count * number of cores in the selected machine type) and (max_replica_count * number of GPUs per replica in the selected machine type).
- min
Replica NumberCount - Immutable. The minimum number of machine replicas this DeployedModel will be always deployed on. This value must be greater than or equal to 1. If traffic against the DeployedModel increases, it may dynamically be deployed onto more replicas, and as traffic decreases, some of these extra replicas may be freed.
GoogleCloudAiplatformV1beta1MachineSpecResponse
- Accelerator
Count int - The number of accelerators to attach to the machine.
- Accelerator
Type string - Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count.
- Machine
Type string - Immutable. The type of the machine. See the list of machine types supported for prediction See the list of machine types supported for custom training. For DeployedModel this field is optional, and the default value is
n1-standard-2
. For BatchPredictionJob or as part of WorkerPoolSpec this field is required. - Tpu
Topology string - Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1").
- Accelerator
Count int - The number of accelerators to attach to the machine.
- Accelerator
Type string - Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count.
- Machine
Type string - Immutable. The type of the machine. See the list of machine types supported for prediction See the list of machine types supported for custom training. For DeployedModel this field is optional, and the default value is
n1-standard-2
. For BatchPredictionJob or as part of WorkerPoolSpec this field is required. - Tpu
Topology string - Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1").
- accelerator
Count Integer - The number of accelerators to attach to the machine.
- accelerator
Type String - Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count.
- machine
Type String - Immutable. The type of the machine. See the list of machine types supported for prediction See the list of machine types supported for custom training. For DeployedModel this field is optional, and the default value is
n1-standard-2
. For BatchPredictionJob or as part of WorkerPoolSpec this field is required. - tpu
Topology String - Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1").
- accelerator
Count number - The number of accelerators to attach to the machine.
- accelerator
Type string - Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count.
- machine
Type string - Immutable. The type of the machine. See the list of machine types supported for prediction See the list of machine types supported for custom training. For DeployedModel this field is optional, and the default value is
n1-standard-2
. For BatchPredictionJob or as part of WorkerPoolSpec this field is required. - tpu
Topology string - Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1").
- accelerator_
count int - The number of accelerators to attach to the machine.
- accelerator_
type str - Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count.
- machine_
type str - Immutable. The type of the machine. See the list of machine types supported for prediction See the list of machine types supported for custom training. For DeployedModel this field is optional, and the default value is
n1-standard-2
. For BatchPredictionJob or as part of WorkerPoolSpec this field is required. - tpu_
topology str - Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1").
- accelerator
Count Number - The number of accelerators to attach to the machine.
- accelerator
Type String - Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count.
- machine
Type String - Immutable. The type of the machine. See the list of machine types supported for prediction See the list of machine types supported for custom training. For DeployedModel this field is optional, and the default value is
n1-standard-2
. For BatchPredictionJob or as part of WorkerPoolSpec this field is required. - tpu
Topology String - Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1").
Package Details
- Repository
- Google Cloud Native pulumi/pulumi-google-native
- License
- Apache-2.0
Google Cloud Native is in preview. Google Cloud Classic is fully supported.
Google Cloud Native v0.32.0 published on Wednesday, Nov 29, 2023 by Pulumi