AWS Native is in preview. AWS Classic is fully supported.
AWS Native v0.109.0 published on Wednesday, Jun 26, 2024 by Pulumi
aws-native.sagemaker.getInferenceComponent
Explore with Pulumi AI
AWS Native is in preview. AWS Classic is fully supported.
AWS Native v0.109.0 published on Wednesday, Jun 26, 2024 by Pulumi
Resource Type definition for AWS::SageMaker::InferenceComponent
Using getInferenceComponent
Two invocation forms are available. The direct form accepts plain arguments and either blocks until the result value is available, or returns a Promise-wrapped result. The output form accepts Input-wrapped arguments and returns an Output-wrapped result.
function getInferenceComponent(args: GetInferenceComponentArgs, opts?: InvokeOptions): Promise<GetInferenceComponentResult>
function getInferenceComponentOutput(args: GetInferenceComponentOutputArgs, opts?: InvokeOptions): Output<GetInferenceComponentResult>
def get_inference_component(inference_component_arn: Optional[str] = None,
opts: Optional[InvokeOptions] = None) -> GetInferenceComponentResult
def get_inference_component_output(inference_component_arn: Optional[pulumi.Input[str]] = None,
opts: Optional[InvokeOptions] = None) -> Output[GetInferenceComponentResult]
func LookupInferenceComponent(ctx *Context, args *LookupInferenceComponentArgs, opts ...InvokeOption) (*LookupInferenceComponentResult, error)
func LookupInferenceComponentOutput(ctx *Context, args *LookupInferenceComponentOutputArgs, opts ...InvokeOption) LookupInferenceComponentResultOutput
> Note: This function is named LookupInferenceComponent
in the Go SDK.
public static class GetInferenceComponent
{
public static Task<GetInferenceComponentResult> InvokeAsync(GetInferenceComponentArgs args, InvokeOptions? opts = null)
public static Output<GetInferenceComponentResult> Invoke(GetInferenceComponentInvokeArgs args, InvokeOptions? opts = null)
}
public static CompletableFuture<GetInferenceComponentResult> getInferenceComponent(GetInferenceComponentArgs args, InvokeOptions options)
// Output-based functions aren't available in Java yet
fn::invoke:
function: aws-native:sagemaker:getInferenceComponent
arguments:
# arguments dictionary
The following arguments are supported:
- Inference
Component stringArn - The Amazon Resource Name (ARN) of the inference component.
- Inference
Component stringArn - The Amazon Resource Name (ARN) of the inference component.
- inference
Component StringArn - The Amazon Resource Name (ARN) of the inference component.
- inference
Component stringArn - The Amazon Resource Name (ARN) of the inference component.
- inference_
component_ strarn - The Amazon Resource Name (ARN) of the inference component.
- inference
Component StringArn - The Amazon Resource Name (ARN) of the inference component.
getInferenceComponent Result
The following output properties are available:
- Creation
Time string - The time when the inference component was created.
- Endpoint
Arn string - The Amazon Resource Name (ARN) of the endpoint that hosts the inference component.
- Endpoint
Name string - The name of the endpoint that hosts the inference component.
- Failure
Reason string - Inference
Component stringArn - The Amazon Resource Name (ARN) of the inference component.
- Inference
Component stringName - The name of the inference component.
- Inference
Component Pulumi.Status Aws Native. Sage Maker. Inference Component Status - The status of the inference component.
- Last
Modified stringTime - The time when the inference component was last updated.
- Runtime
Config Pulumi.Aws Native. Sage Maker. Outputs. Inference Component Runtime Config - Specification
Pulumi.
Aws Native. Sage Maker. Outputs. Inference Component Specification - List<Pulumi.
Aws Native. Outputs. Tag> - Variant
Name string - The name of the production variant that hosts the inference component.
- Creation
Time string - The time when the inference component was created.
- Endpoint
Arn string - The Amazon Resource Name (ARN) of the endpoint that hosts the inference component.
- Endpoint
Name string - The name of the endpoint that hosts the inference component.
- Failure
Reason string - Inference
Component stringArn - The Amazon Resource Name (ARN) of the inference component.
- Inference
Component stringName - The name of the inference component.
- Inference
Component InferenceStatus Component Status - The status of the inference component.
- Last
Modified stringTime - The time when the inference component was last updated.
- Runtime
Config InferenceComponent Runtime Config - Specification
Inference
Component Specification - Tag
- Variant
Name string - The name of the production variant that hosts the inference component.
- creation
Time String - The time when the inference component was created.
- endpoint
Arn String - The Amazon Resource Name (ARN) of the endpoint that hosts the inference component.
- endpoint
Name String - The name of the endpoint that hosts the inference component.
- failure
Reason String - inference
Component StringArn - The Amazon Resource Name (ARN) of the inference component.
- inference
Component StringName - The name of the inference component.
- inference
Component InferenceStatus Component Status - The status of the inference component.
- last
Modified StringTime - The time when the inference component was last updated.
- runtime
Config InferenceComponent Runtime Config - specification
Inference
Component Specification - List<Tag>
- variant
Name String - The name of the production variant that hosts the inference component.
- creation
Time string - The time when the inference component was created.
- endpoint
Arn string - The Amazon Resource Name (ARN) of the endpoint that hosts the inference component.
- endpoint
Name string - The name of the endpoint that hosts the inference component.
- failure
Reason string - inference
Component stringArn - The Amazon Resource Name (ARN) of the inference component.
- inference
Component stringName - The name of the inference component.
- inference
Component InferenceStatus Component Status - The status of the inference component.
- last
Modified stringTime - The time when the inference component was last updated.
- runtime
Config InferenceComponent Runtime Config - specification
Inference
Component Specification - Tag[]
- variant
Name string - The name of the production variant that hosts the inference component.
- creation_
time str - The time when the inference component was created.
- endpoint_
arn str - The Amazon Resource Name (ARN) of the endpoint that hosts the inference component.
- endpoint_
name str - The name of the endpoint that hosts the inference component.
- failure_
reason str - inference_
component_ strarn - The Amazon Resource Name (ARN) of the inference component.
- inference_
component_ strname - The name of the inference component.
- inference_
component_ Inferencestatus Component Status - The status of the inference component.
- last_
modified_ strtime - The time when the inference component was last updated.
- runtime_
config InferenceComponent Runtime Config - specification
Inference
Component Specification - Sequence[root_Tag]
- variant_
name str - The name of the production variant that hosts the inference component.
- creation
Time String - The time when the inference component was created.
- endpoint
Arn String - The Amazon Resource Name (ARN) of the endpoint that hosts the inference component.
- endpoint
Name String - The name of the endpoint that hosts the inference component.
- failure
Reason String - inference
Component StringArn - The Amazon Resource Name (ARN) of the inference component.
- inference
Component StringName - The name of the inference component.
- inference
Component "InStatus Service" | "Creating" | "Updating" | "Failed" | "Deleting" - The status of the inference component.
- last
Modified StringTime - The time when the inference component was last updated.
- runtime
Config Property Map - specification Property Map
- List<Property Map>
- variant
Name String - The name of the production variant that hosts the inference component.
Supporting Types
InferenceComponentComputeResourceRequirements
- Max
Memory intRequired In Mb - The maximum MB of memory to allocate to run a model that you assign to an inference component.
- Min
Memory intRequired In Mb - The minimum MB of memory to allocate to run a model that you assign to an inference component.
- Number
Of doubleAccelerator Devices Required - The number of accelerators to allocate to run a model that you assign to an inference component. Accelerators include GPUs and AWS Inferentia.
- Number
Of doubleCpu Cores Required - The number of CPU cores to allocate to run a model that you assign to an inference component.
- Max
Memory intRequired In Mb - The maximum MB of memory to allocate to run a model that you assign to an inference component.
- Min
Memory intRequired In Mb - The minimum MB of memory to allocate to run a model that you assign to an inference component.
- Number
Of float64Accelerator Devices Required - The number of accelerators to allocate to run a model that you assign to an inference component. Accelerators include GPUs and AWS Inferentia.
- Number
Of float64Cpu Cores Required - The number of CPU cores to allocate to run a model that you assign to an inference component.
- max
Memory IntegerRequired In Mb - The maximum MB of memory to allocate to run a model that you assign to an inference component.
- min
Memory IntegerRequired In Mb - The minimum MB of memory to allocate to run a model that you assign to an inference component.
- number
Of DoubleAccelerator Devices Required - The number of accelerators to allocate to run a model that you assign to an inference component. Accelerators include GPUs and AWS Inferentia.
- number
Of DoubleCpu Cores Required - The number of CPU cores to allocate to run a model that you assign to an inference component.
- max
Memory numberRequired In Mb - The maximum MB of memory to allocate to run a model that you assign to an inference component.
- min
Memory numberRequired In Mb - The minimum MB of memory to allocate to run a model that you assign to an inference component.
- number
Of numberAccelerator Devices Required - The number of accelerators to allocate to run a model that you assign to an inference component. Accelerators include GPUs and AWS Inferentia.
- number
Of numberCpu Cores Required - The number of CPU cores to allocate to run a model that you assign to an inference component.
- max_
memory_ intrequired_ in_ mb - The maximum MB of memory to allocate to run a model that you assign to an inference component.
- min_
memory_ intrequired_ in_ mb - The minimum MB of memory to allocate to run a model that you assign to an inference component.
- number_
of_ floataccelerator_ devices_ required - The number of accelerators to allocate to run a model that you assign to an inference component. Accelerators include GPUs and AWS Inferentia.
- number_
of_ floatcpu_ cores_ required - The number of CPU cores to allocate to run a model that you assign to an inference component.
- max
Memory NumberRequired In Mb - The maximum MB of memory to allocate to run a model that you assign to an inference component.
- min
Memory NumberRequired In Mb - The minimum MB of memory to allocate to run a model that you assign to an inference component.
- number
Of NumberAccelerator Devices Required - The number of accelerators to allocate to run a model that you assign to an inference component. Accelerators include GPUs and AWS Inferentia.
- number
Of NumberCpu Cores Required - The number of CPU cores to allocate to run a model that you assign to an inference component.
InferenceComponentContainerSpecification
- Artifact
Url string - The Amazon S3 path where the model artifacts, which result from model training, are stored. This path must point to a single gzip compressed tar archive (.tar.gz suffix).
- Deployed
Image Pulumi.Aws Native. Sage Maker. Inputs. Inference Component Deployed Image - Environment Dictionary<string, string>
- The environment variables to set in the Docker container. Each key and value in the Environment string-to-string map can have length of up to 1024. We support up to 16 entries in the map.
- Image string
- The Amazon Elastic Container Registry (Amazon ECR) path where the Docker image for the model is stored.
- Artifact
Url string - The Amazon S3 path where the model artifacts, which result from model training, are stored. This path must point to a single gzip compressed tar archive (.tar.gz suffix).
- Deployed
Image InferenceComponent Deployed Image - Environment map[string]string
- The environment variables to set in the Docker container. Each key and value in the Environment string-to-string map can have length of up to 1024. We support up to 16 entries in the map.
- Image string
- The Amazon Elastic Container Registry (Amazon ECR) path where the Docker image for the model is stored.
- artifact
Url String - The Amazon S3 path where the model artifacts, which result from model training, are stored. This path must point to a single gzip compressed tar archive (.tar.gz suffix).
- deployed
Image InferenceComponent Deployed Image - environment Map<String,String>
- The environment variables to set in the Docker container. Each key and value in the Environment string-to-string map can have length of up to 1024. We support up to 16 entries in the map.
- image String
- The Amazon Elastic Container Registry (Amazon ECR) path where the Docker image for the model is stored.
- artifact
Url string - The Amazon S3 path where the model artifacts, which result from model training, are stored. This path must point to a single gzip compressed tar archive (.tar.gz suffix).
- deployed
Image InferenceComponent Deployed Image - environment {[key: string]: string}
- The environment variables to set in the Docker container. Each key and value in the Environment string-to-string map can have length of up to 1024. We support up to 16 entries in the map.
- image string
- The Amazon Elastic Container Registry (Amazon ECR) path where the Docker image for the model is stored.
- artifact_
url str - The Amazon S3 path where the model artifacts, which result from model training, are stored. This path must point to a single gzip compressed tar archive (.tar.gz suffix).
- deployed_
image InferenceComponent Deployed Image - environment Mapping[str, str]
- The environment variables to set in the Docker container. Each key and value in the Environment string-to-string map can have length of up to 1024. We support up to 16 entries in the map.
- image str
- The Amazon Elastic Container Registry (Amazon ECR) path where the Docker image for the model is stored.
- artifact
Url String - The Amazon S3 path where the model artifacts, which result from model training, are stored. This path must point to a single gzip compressed tar archive (.tar.gz suffix).
- deployed
Image Property Map - environment Map<String>
- The environment variables to set in the Docker container. Each key and value in the Environment string-to-string map can have length of up to 1024. We support up to 16 entries in the map.
- image String
- The Amazon Elastic Container Registry (Amazon ECR) path where the Docker image for the model is stored.
InferenceComponentDeployedImage
- Resolution
Time string - The date and time when the image path for the model resolved to the
ResolvedImage
- Resolved
Image string - The specific digest path of the image hosted in this
ProductionVariant
. - Specified
Image string - The image path you specified when you created the model.
- Resolution
Time string - The date and time when the image path for the model resolved to the
ResolvedImage
- Resolved
Image string - The specific digest path of the image hosted in this
ProductionVariant
. - Specified
Image string - The image path you specified when you created the model.
- resolution
Time String - The date and time when the image path for the model resolved to the
ResolvedImage
- resolved
Image String - The specific digest path of the image hosted in this
ProductionVariant
. - specified
Image String - The image path you specified when you created the model.
- resolution
Time string - The date and time when the image path for the model resolved to the
ResolvedImage
- resolved
Image string - The specific digest path of the image hosted in this
ProductionVariant
. - specified
Image string - The image path you specified when you created the model.
- resolution_
time str - The date and time when the image path for the model resolved to the
ResolvedImage
- resolved_
image str - The specific digest path of the image hosted in this
ProductionVariant
. - specified_
image str - The image path you specified when you created the model.
- resolution
Time String - The date and time when the image path for the model resolved to the
ResolvedImage
- resolved
Image String - The specific digest path of the image hosted in this
ProductionVariant
. - specified
Image String - The image path you specified when you created the model.
InferenceComponentRuntimeConfig
- Copy
Count int - The number of runtime copies of the model container to deploy with the inference component. Each copy can serve inference requests.
- Current
Copy intCount - Desired
Copy intCount
- Copy
Count int - The number of runtime copies of the model container to deploy with the inference component. Each copy can serve inference requests.
- Current
Copy intCount - Desired
Copy intCount
- copy
Count Integer - The number of runtime copies of the model container to deploy with the inference component. Each copy can serve inference requests.
- current
Copy IntegerCount - desired
Copy IntegerCount
- copy
Count number - The number of runtime copies of the model container to deploy with the inference component. Each copy can serve inference requests.
- current
Copy numberCount - desired
Copy numberCount
- copy_
count int - The number of runtime copies of the model container to deploy with the inference component. Each copy can serve inference requests.
- current_
copy_ intcount - desired_
copy_ intcount
- copy
Count Number - The number of runtime copies of the model container to deploy with the inference component. Each copy can serve inference requests.
- current
Copy NumberCount - desired
Copy NumberCount
InferenceComponentSpecification
- Compute
Resource Pulumi.Requirements Aws Native. Sage Maker. Inputs. Inference Component Compute Resource Requirements - The compute resources allocated to run the model assigned to the inference component.
- Container
Pulumi.
Aws Native. Sage Maker. Inputs. Inference Component Container Specification - Defines a container that provides the runtime environment for a model that you deploy with an inference component.
- Model
Name string - The name of an existing SageMaker model object in your account that you want to deploy with the inference component.
- Startup
Parameters Pulumi.Aws Native. Sage Maker. Inputs. Inference Component Startup Parameters - Settings that take effect while the model container starts up.
- Compute
Resource InferenceRequirements Component Compute Resource Requirements - The compute resources allocated to run the model assigned to the inference component.
- Container
Inference
Component Container Specification - Defines a container that provides the runtime environment for a model that you deploy with an inference component.
- Model
Name string - The name of an existing SageMaker model object in your account that you want to deploy with the inference component.
- Startup
Parameters InferenceComponent Startup Parameters - Settings that take effect while the model container starts up.
- compute
Resource InferenceRequirements Component Compute Resource Requirements - The compute resources allocated to run the model assigned to the inference component.
- container
Inference
Component Container Specification - Defines a container that provides the runtime environment for a model that you deploy with an inference component.
- model
Name String - The name of an existing SageMaker model object in your account that you want to deploy with the inference component.
- startup
Parameters InferenceComponent Startup Parameters - Settings that take effect while the model container starts up.
- compute
Resource InferenceRequirements Component Compute Resource Requirements - The compute resources allocated to run the model assigned to the inference component.
- container
Inference
Component Container Specification - Defines a container that provides the runtime environment for a model that you deploy with an inference component.
- model
Name string - The name of an existing SageMaker model object in your account that you want to deploy with the inference component.
- startup
Parameters InferenceComponent Startup Parameters - Settings that take effect while the model container starts up.
- compute_
resource_ Inferencerequirements Component Compute Resource Requirements - The compute resources allocated to run the model assigned to the inference component.
- container
Inference
Component Container Specification - Defines a container that provides the runtime environment for a model that you deploy with an inference component.
- model_
name str - The name of an existing SageMaker model object in your account that you want to deploy with the inference component.
- startup_
parameters InferenceComponent Startup Parameters - Settings that take effect while the model container starts up.
- compute
Resource Property MapRequirements - The compute resources allocated to run the model assigned to the inference component.
- container Property Map
- Defines a container that provides the runtime environment for a model that you deploy with an inference component.
- model
Name String - The name of an existing SageMaker model object in your account that you want to deploy with the inference component.
- startup
Parameters Property Map - Settings that take effect while the model container starts up.
InferenceComponentStartupParameters
- Container
Startup intHealth Check Timeout In Seconds - The timeout value, in seconds, for your inference container to pass health check by Amazon S3 Hosting. For more information about health check, see How Your Container Should Respond to Health Check (Ping) Requests .
- Model
Data intDownload Timeout In Seconds - The timeout value, in seconds, to download and extract the model that you want to host from Amazon S3 to the individual inference instance associated with this inference component.
- Container
Startup intHealth Check Timeout In Seconds - The timeout value, in seconds, for your inference container to pass health check by Amazon S3 Hosting. For more information about health check, see How Your Container Should Respond to Health Check (Ping) Requests .
- Model
Data intDownload Timeout In Seconds - The timeout value, in seconds, to download and extract the model that you want to host from Amazon S3 to the individual inference instance associated with this inference component.
- container
Startup IntegerHealth Check Timeout In Seconds - The timeout value, in seconds, for your inference container to pass health check by Amazon S3 Hosting. For more information about health check, see How Your Container Should Respond to Health Check (Ping) Requests .
- model
Data IntegerDownload Timeout In Seconds - The timeout value, in seconds, to download and extract the model that you want to host from Amazon S3 to the individual inference instance associated with this inference component.
- container
Startup numberHealth Check Timeout In Seconds - The timeout value, in seconds, for your inference container to pass health check by Amazon S3 Hosting. For more information about health check, see How Your Container Should Respond to Health Check (Ping) Requests .
- model
Data numberDownload Timeout In Seconds - The timeout value, in seconds, to download and extract the model that you want to host from Amazon S3 to the individual inference instance associated with this inference component.
- container_
startup_ inthealth_ check_ timeout_ in_ seconds - The timeout value, in seconds, for your inference container to pass health check by Amazon S3 Hosting. For more information about health check, see How Your Container Should Respond to Health Check (Ping) Requests .
- model_
data_ intdownload_ timeout_ in_ seconds - The timeout value, in seconds, to download and extract the model that you want to host from Amazon S3 to the individual inference instance associated with this inference component.
- container
Startup NumberHealth Check Timeout In Seconds - The timeout value, in seconds, for your inference container to pass health check by Amazon S3 Hosting. For more information about health check, see How Your Container Should Respond to Health Check (Ping) Requests .
- model
Data NumberDownload Timeout In Seconds - The timeout value, in seconds, to download and extract the model that you want to host from Amazon S3 to the individual inference instance associated with this inference component.
InferenceComponentStatus
Tag
Package Details
- Repository
- AWS Native pulumi/pulumi-aws-native
- License
- Apache-2.0
AWS Native is in preview. AWS Classic is fully supported.
AWS Native v0.109.0 published on Wednesday, Jun 26, 2024 by Pulumi