libZSservicesZSamazonka-sagemakerZSamazonka-sagemaker
Copyright(c) 2013-2021 Brendan Hay
LicenseMozilla Public License, v. 2.0.
MaintainerBrendan Hay <brendan.g.hay+amazonka@gmail.com>
Stabilityauto-generated
Portabilitynon-portable (GHC extensions)
Safe HaskellNone

Amazonka.SageMaker.Types.ContainerDefinition

Description

 
Synopsis

Documentation

data ContainerDefinition Source #

Describes the container, as part of model definition.

See: newContainerDefinition smart constructor.

Constructors

ContainerDefinition' 

Fields

  • multiModelConfig :: Maybe MultiModelConfig

    Specifies additional configuration for multi-model endpoints.

  • modelDataUrl :: Maybe Text

    The S3 path where the model artifacts, which result from model training, are stored. This path must point to a single gzip compressed tar archive (.tar.gz suffix). The S3 path is required for Amazon SageMaker built-in algorithms, but not if you use your own algorithms. For more information on built-in algorithms, see Common Parameters.

    The model artifacts must be in an S3 bucket that is in the same region as the model or endpoint you are creating.

    If you provide a value for this parameter, Amazon SageMaker uses Amazon Web Services Security Token Service to download model artifacts from the S3 path you provide. Amazon Web Services STS is activated in your IAM user account by default. If you previously deactivated Amazon Web Services STS for a region, you need to reactivate Amazon Web Services STS for that region. For more information, see Activating and Deactivating Amazon Web Services STS in an Amazon Web Services Region in the Amazon Web Services Identity and Access Management User Guide.

    If you use a built-in algorithm to create a model, Amazon SageMaker requires that you provide a S3 path to the model artifacts in ModelDataUrl.

  • image :: Maybe Text

    The path where inference code is stored. This can be either in Amazon EC2 Container Registry or in a Docker registry that is accessible from the same VPC that you configure for your endpoint. If you are using your own custom algorithm instead of an algorithm provided by Amazon SageMaker, the inference code must meet Amazon SageMaker requirements. Amazon SageMaker supports both registry/repository[:tag] and registry/repository[@digest] image path formats. For more information, see Using Your Own Algorithms with Amazon SageMaker

  • modelPackageName :: Maybe Text

    The name or Amazon Resource Name (ARN) of the model package to use to create the model.

  • environment :: Maybe (HashMap Text Text)

    The environment variables to set in the Docker container. Each key and value in the Environment string to string map can have length of up to 1024. We support up to 16 entries in the map.

  • imageConfig :: Maybe ImageConfig

    Specifies whether the model container is in Amazon ECR or a private Docker registry accessible from your Amazon Virtual Private Cloud (VPC). For information about storing containers in a private Docker registry, see Use a Private Docker Registry for Real-Time Inference Containers

  • mode :: Maybe ContainerMode

    Whether the container hosts a single model or multiple models.

  • containerHostname :: Maybe Text

    This parameter is ignored for models that contain only a PrimaryContainer.

    When a ContainerDefinition is part of an inference pipeline, the value of the parameter uniquely identifies the container for the purposes of logging and metrics. For information, see Use Logs and Metrics to Monitor an Inference Pipeline. If you don't specify a value for this parameter for a ContainerDefinition that is part of an inference pipeline, a unique name is automatically assigned based on the position of the ContainerDefinition in the pipeline. If you specify a value for the ContainerHostName for any ContainerDefinition that is part of an inference pipeline, you must specify a value for the ContainerHostName parameter of every ContainerDefinition in that pipeline.

Instances

Instances details
Eq ContainerDefinition Source # 
Instance details

Defined in Amazonka.SageMaker.Types.ContainerDefinition

Read ContainerDefinition Source # 
Instance details

Defined in Amazonka.SageMaker.Types.ContainerDefinition

Show ContainerDefinition Source # 
Instance details

Defined in Amazonka.SageMaker.Types.ContainerDefinition

Generic ContainerDefinition Source # 
Instance details

Defined in Amazonka.SageMaker.Types.ContainerDefinition

Associated Types

type Rep ContainerDefinition :: Type -> Type #

NFData ContainerDefinition Source # 
Instance details

Defined in Amazonka.SageMaker.Types.ContainerDefinition

Methods

rnf :: ContainerDefinition -> () #

Hashable ContainerDefinition Source # 
Instance details

Defined in Amazonka.SageMaker.Types.ContainerDefinition

ToJSON ContainerDefinition Source # 
Instance details

Defined in Amazonka.SageMaker.Types.ContainerDefinition

FromJSON ContainerDefinition Source # 
Instance details

Defined in Amazonka.SageMaker.Types.ContainerDefinition

type Rep ContainerDefinition Source # 
Instance details

Defined in Amazonka.SageMaker.Types.ContainerDefinition

type Rep ContainerDefinition = D1 ('MetaData "ContainerDefinition" "Amazonka.SageMaker.Types.ContainerDefinition" "libZSservicesZSamazonka-sagemakerZSamazonka-sagemaker" 'False) (C1 ('MetaCons "ContainerDefinition'" 'PrefixI 'True) (((S1 ('MetaSel ('Just "multiModelConfig") 'NoSourceUnpackedness 'NoSourceStrictness 'DecidedStrict) (Rec0 (Maybe MultiModelConfig)) :*: S1 ('MetaSel ('Just "modelDataUrl") 'NoSourceUnpackedness 'NoSourceStrictness 'DecidedStrict) (Rec0 (Maybe Text))) :*: (S1 ('MetaSel ('Just "image") 'NoSourceUnpackedness 'NoSourceStrictness 'DecidedStrict) (Rec0 (Maybe Text)) :*: S1 ('MetaSel ('Just "modelPackageName") 'NoSourceUnpackedness 'NoSourceStrictness 'DecidedStrict) (Rec0 (Maybe Text)))) :*: ((S1 ('MetaSel ('Just "environment") 'NoSourceUnpackedness 'NoSourceStrictness 'DecidedStrict) (Rec0 (Maybe (HashMap Text Text))) :*: S1 ('MetaSel ('Just "imageConfig") 'NoSourceUnpackedness 'NoSourceStrictness 'DecidedStrict) (Rec0 (Maybe ImageConfig))) :*: (S1 ('MetaSel ('Just "mode") 'NoSourceUnpackedness 'NoSourceStrictness 'DecidedStrict) (Rec0 (Maybe ContainerMode)) :*: S1 ('MetaSel ('Just "containerHostname") 'NoSourceUnpackedness 'NoSourceStrictness 'DecidedStrict) (Rec0 (Maybe Text))))))

newContainerDefinition :: ContainerDefinition Source #

Create a value of ContainerDefinition with all optional fields omitted.

Use generic-lens or optics to modify other optional fields.

The following record fields are available, with the corresponding lenses provided for backwards compatibility:

$sel:multiModelConfig:ContainerDefinition', containerDefinition_multiModelConfig - Specifies additional configuration for multi-model endpoints.

$sel:modelDataUrl:ContainerDefinition', containerDefinition_modelDataUrl - The S3 path where the model artifacts, which result from model training, are stored. This path must point to a single gzip compressed tar archive (.tar.gz suffix). The S3 path is required for Amazon SageMaker built-in algorithms, but not if you use your own algorithms. For more information on built-in algorithms, see Common Parameters.

The model artifacts must be in an S3 bucket that is in the same region as the model or endpoint you are creating.

If you provide a value for this parameter, Amazon SageMaker uses Amazon Web Services Security Token Service to download model artifacts from the S3 path you provide. Amazon Web Services STS is activated in your IAM user account by default. If you previously deactivated Amazon Web Services STS for a region, you need to reactivate Amazon Web Services STS for that region. For more information, see Activating and Deactivating Amazon Web Services STS in an Amazon Web Services Region in the Amazon Web Services Identity and Access Management User Guide.

If you use a built-in algorithm to create a model, Amazon SageMaker requires that you provide a S3 path to the model artifacts in ModelDataUrl.

$sel:image:ContainerDefinition', containerDefinition_image - The path where inference code is stored. This can be either in Amazon EC2 Container Registry or in a Docker registry that is accessible from the same VPC that you configure for your endpoint. If you are using your own custom algorithm instead of an algorithm provided by Amazon SageMaker, the inference code must meet Amazon SageMaker requirements. Amazon SageMaker supports both registry/repository[:tag] and registry/repository[@digest] image path formats. For more information, see Using Your Own Algorithms with Amazon SageMaker

$sel:modelPackageName:ContainerDefinition', containerDefinition_modelPackageName - The name or Amazon Resource Name (ARN) of the model package to use to create the model.

$sel:environment:ContainerDefinition', containerDefinition_environment - The environment variables to set in the Docker container. Each key and value in the Environment string to string map can have length of up to 1024. We support up to 16 entries in the map.

$sel:imageConfig:ContainerDefinition', containerDefinition_imageConfig - Specifies whether the model container is in Amazon ECR or a private Docker registry accessible from your Amazon Virtual Private Cloud (VPC). For information about storing containers in a private Docker registry, see Use a Private Docker Registry for Real-Time Inference Containers

$sel:mode:ContainerDefinition', containerDefinition_mode - Whether the container hosts a single model or multiple models.

$sel:containerHostname:ContainerDefinition', containerDefinition_containerHostname - This parameter is ignored for models that contain only a PrimaryContainer.

When a ContainerDefinition is part of an inference pipeline, the value of the parameter uniquely identifies the container for the purposes of logging and metrics. For information, see Use Logs and Metrics to Monitor an Inference Pipeline. If you don't specify a value for this parameter for a ContainerDefinition that is part of an inference pipeline, a unique name is automatically assigned based on the position of the ContainerDefinition in the pipeline. If you specify a value for the ContainerHostName for any ContainerDefinition that is part of an inference pipeline, you must specify a value for the ContainerHostName parameter of every ContainerDefinition in that pipeline.

containerDefinition_multiModelConfig :: Lens' ContainerDefinition (Maybe MultiModelConfig) Source #

Specifies additional configuration for multi-model endpoints.

containerDefinition_modelDataUrl :: Lens' ContainerDefinition (Maybe Text) Source #

The S3 path where the model artifacts, which result from model training, are stored. This path must point to a single gzip compressed tar archive (.tar.gz suffix). The S3 path is required for Amazon SageMaker built-in algorithms, but not if you use your own algorithms. For more information on built-in algorithms, see Common Parameters.

The model artifacts must be in an S3 bucket that is in the same region as the model or endpoint you are creating.

If you provide a value for this parameter, Amazon SageMaker uses Amazon Web Services Security Token Service to download model artifacts from the S3 path you provide. Amazon Web Services STS is activated in your IAM user account by default. If you previously deactivated Amazon Web Services STS for a region, you need to reactivate Amazon Web Services STS for that region. For more information, see Activating and Deactivating Amazon Web Services STS in an Amazon Web Services Region in the Amazon Web Services Identity and Access Management User Guide.

If you use a built-in algorithm to create a model, Amazon SageMaker requires that you provide a S3 path to the model artifacts in ModelDataUrl.

containerDefinition_image :: Lens' ContainerDefinition (Maybe Text) Source #

The path where inference code is stored. This can be either in Amazon EC2 Container Registry or in a Docker registry that is accessible from the same VPC that you configure for your endpoint. If you are using your own custom algorithm instead of an algorithm provided by Amazon SageMaker, the inference code must meet Amazon SageMaker requirements. Amazon SageMaker supports both registry/repository[:tag] and registry/repository[@digest] image path formats. For more information, see Using Your Own Algorithms with Amazon SageMaker

containerDefinition_modelPackageName :: Lens' ContainerDefinition (Maybe Text) Source #

The name or Amazon Resource Name (ARN) of the model package to use to create the model.

containerDefinition_environment :: Lens' ContainerDefinition (Maybe (HashMap Text Text)) Source #

The environment variables to set in the Docker container. Each key and value in the Environment string to string map can have length of up to 1024. We support up to 16 entries in the map.

containerDefinition_imageConfig :: Lens' ContainerDefinition (Maybe ImageConfig) Source #

Specifies whether the model container is in Amazon ECR or a private Docker registry accessible from your Amazon Virtual Private Cloud (VPC). For information about storing containers in a private Docker registry, see Use a Private Docker Registry for Real-Time Inference Containers

containerDefinition_mode :: Lens' ContainerDefinition (Maybe ContainerMode) Source #

Whether the container hosts a single model or multiple models.

containerDefinition_containerHostname :: Lens' ContainerDefinition (Maybe Text) Source #

This parameter is ignored for models that contain only a PrimaryContainer.

When a ContainerDefinition is part of an inference pipeline, the value of the parameter uniquely identifies the container for the purposes of logging and metrics. For information, see Use Logs and Metrics to Monitor an Inference Pipeline. If you don't specify a value for this parameter for a ContainerDefinition that is part of an inference pipeline, a unique name is automatically assigned based on the position of the ContainerDefinition in the pipeline. If you specify a value for the ContainerHostName for any ContainerDefinition that is part of an inference pipeline, you must specify a value for the ContainerHostName parameter of every ContainerDefinition in that pipeline.