| Package | Description |
|---|---|
| com.amazonaws.services.sagemaker |
Provides APIs for creating and managing SageMaker resources.
|
| com.amazonaws.services.sagemaker.model |
| Modifier and Type | Method and Description |
|---|---|
Future<CreateModelResult> |
AmazonSageMakerAsyncClient.createModelAsync(CreateModelRequest request,
AsyncHandler<CreateModelRequest,CreateModelResult> asyncHandler) |
Future<CreateModelResult> |
AbstractAmazonSageMakerAsync.createModelAsync(CreateModelRequest request,
AsyncHandler<CreateModelRequest,CreateModelResult> asyncHandler) |
Future<CreateModelResult> |
AmazonSageMakerAsync.createModelAsync(CreateModelRequest createModelRequest,
AsyncHandler<CreateModelRequest,CreateModelResult> asyncHandler)
Creates a model in SageMaker.
|
| Modifier and Type | Method and Description |
|---|---|
CreateModelRequest |
CreateModelRequest.clone() |
CreateModelRequest |
CreateModelRequest.withContainers(Collection<ContainerDefinition> containers)
Specifies the containers in the inference pipeline.
|
CreateModelRequest |
CreateModelRequest.withContainers(ContainerDefinition... containers)
Specifies the containers in the inference pipeline.
|
CreateModelRequest |
CreateModelRequest.withEnableNetworkIsolation(Boolean enableNetworkIsolation)
Isolates the model container.
|
CreateModelRequest |
CreateModelRequest.withExecutionRoleArn(String executionRoleArn)
The Amazon Resource Name (ARN) of the IAM role that SageMaker can assume to access model artifacts and docker
image for deployment on ML compute instances or for batch transform jobs.
|
CreateModelRequest |
CreateModelRequest.withInferenceExecutionConfig(InferenceExecutionConfig inferenceExecutionConfig)
Specifies details of how containers in a multi-container endpoint are called.
|
CreateModelRequest |
CreateModelRequest.withModelName(String modelName)
The name of the new model.
|
CreateModelRequest |
CreateModelRequest.withPrimaryContainer(ContainerDefinition primaryContainer)
The location of the primary docker image containing inference code, associated artifacts, and custom environment
map that the inference code uses when the model is deployed for predictions.
|
CreateModelRequest |
CreateModelRequest.withTags(Collection<Tag> tags)
An array of key-value pairs.
|
CreateModelRequest |
CreateModelRequest.withTags(Tag... tags)
An array of key-value pairs.
|
CreateModelRequest |
CreateModelRequest.withVpcConfig(VpcConfig vpcConfig)
A VpcConfig object
that specifies the VPC that you want your model to connect to.
|
Copyright © 2025. All rights reserved.