From a4a0836ac36968101924691c30edfc1667befb65 Mon Sep 17 00:00:00 2001
From: awstools Creates a new Amazon ECS cluster. By default, your account receives a default
* cluster when you launch your first container instance. However, you can create your own
- * cluster with a unique name with the CreateCluster
action.
When you call the CreateCluster API operation, Amazon ECS attempts to + *
When you call the CreateCluster API operation, Amazon ECS attempts to * create the Amazon ECS service-linked role for your account. This is so that it can manage * required resources in other Amazon Web Services services on your behalf. However, if the user that * makes the call doesn't have permissions to create the service-linked role, it isn't diff --git a/clients/client-ecs/src/commands/CreateServiceCommand.ts b/clients/client-ecs/src/commands/CreateServiceCommand.ts index c5d204c34cc2..874ad0aca92d 100644 --- a/clients/client-ecs/src/commands/CreateServiceCommand.ts +++ b/clients/client-ecs/src/commands/CreateServiceCommand.ts @@ -31,7 +31,7 @@ export interface CreateServiceCommandOutput extends CreateServiceResponse, __Met *
Runs and maintains your desired number of tasks from a specified task definition. If
* the number of tasks running in a service drops below the desiredCount
,
* Amazon ECS runs another copy of the task in the specified cluster. To update an existing
- * service, see the UpdateService action.
On March 21, 2024, a change was made to resolve the task definition revision before authorization. When a task definition revision is not specified, authorization will occur using the latest revision of a task definition.
*You can optionally specify a deployment configuration for your service. The deployment
* is initiated by changing properties. For example, the deployment might be initiated by
- * the task definition or by your desired count of a service. This is done with an UpdateService operation. The default value for a replica service for
+ * the task definition or by your desired count of a service. You can use UpdateService. The default value for a replica service for
* minimumHealthyPercent
is 100%. The default value for a daemon service
* for minimumHealthyPercent
is 0%.
If a service uses the ECS
deployment controller, the minimum healthy
@@ -107,7 +107,7 @@ export interface CreateServiceCommandOutput extends CreateServiceResponse, __Met
* currently visible when describing your service.
When creating a service that uses the EXTERNAL
deployment controller, you
* can specify only parameters that aren't controlled at the task set level. The only
- * required parameter is the service name. You control your services using the CreateTaskSet operation. For more information, see Amazon ECS deployment types in the Amazon Elastic Container Service Developer Guide.
When the service scheduler launches new tasks, it determines task placement. For * information about task placement and task placement strategies, see Amazon ECS * task placement in the Amazon Elastic Container Service Developer Guide @@ -575,7 +575,7 @@ export interface CreateServiceCommandOutput extends CreateServiceResponse, __Met * * * @throws {@link ClusterNotFoundException} (client fault) - *
The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
+ *The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
* * @throws {@link InvalidParameterException} (client fault) *The specified parameter isn't valid. Review the available parameters for the API diff --git a/clients/client-ecs/src/commands/CreateTaskSetCommand.ts b/clients/client-ecs/src/commands/CreateTaskSetCommand.ts index 5b446ca24185..b794a44294b8 100644 --- a/clients/client-ecs/src/commands/CreateTaskSetCommand.ts +++ b/clients/client-ecs/src/commands/CreateTaskSetCommand.ts @@ -195,7 +195,7 @@ export interface CreateTaskSetCommandOutput extends CreateTaskSetResponse, __Met * * * @throws {@link ClusterNotFoundException} (client fault) - *
The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
+ *The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
* * @throws {@link InvalidParameterException} (client fault) *The specified parameter isn't valid. Review the available parameters for the API @@ -216,10 +216,10 @@ export interface CreateTaskSetCommandOutput extends CreateTaskSetResponse, __Met * * @throws {@link ServiceNotActiveException} (client fault) *
The specified service isn't active. You can't update a service that's inactive. If you - * have previously deleted a service, you can re-create it with CreateService.
+ * have previously deleted a service, you can re-create it with CreateService. * * @throws {@link ServiceNotFoundException} (client fault) - *The specified service wasn't found. You can view your available services with ListServices. Amazon ECS services are cluster specific and Region + *
The specified service wasn't found. You can view your available services with ListServices. Amazon ECS services are cluster specific and Region * specific.
* * @throws {@link UnsupportedFeatureException} (client fault) diff --git a/clients/client-ecs/src/commands/DeleteAttributesCommand.ts b/clients/client-ecs/src/commands/DeleteAttributesCommand.ts index a7154e1f5b07..49dc96b0927e 100644 --- a/clients/client-ecs/src/commands/DeleteAttributesCommand.ts +++ b/clients/client-ecs/src/commands/DeleteAttributesCommand.ts @@ -68,7 +68,7 @@ export interface DeleteAttributesCommandOutput extends DeleteAttributesResponse, * @see {@link ECSClientResolvedConfig | config} for ECSClient's `config` shape. * * @throws {@link ClusterNotFoundException} (client fault) - *The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
+ *The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
* * @throws {@link InvalidParameterException} (client fault) *The specified parameter isn't valid. Review the available parameters for the API diff --git a/clients/client-ecs/src/commands/DeleteCapacityProviderCommand.ts b/clients/client-ecs/src/commands/DeleteCapacityProviderCommand.ts index 8ad900577984..fe9172ef6564 100644 --- a/clients/client-ecs/src/commands/DeleteCapacityProviderCommand.ts +++ b/clients/client-ecs/src/commands/DeleteCapacityProviderCommand.ts @@ -31,18 +31,17 @@ export interface DeleteCapacityProviderCommandOutput extends DeleteCapacityProvi *
Deletes the specified capacity provider.
*The FARGATE
and FARGATE_SPOT
capacity providers are
- * reserved and can't be deleted. You can disassociate them from a cluster using either
- * the PutClusterCapacityProviders API or by deleting the
+ * reserved and can't be deleted. You can disassociate them from a cluster using either PutCapacityProviderProviders or by deleting the
* cluster.
Prior to a capacity provider being deleted, the capacity provider must be removed from
- * the capacity provider strategy from all services. The UpdateService
+ * the capacity provider strategy from all services. The UpdateService
* API can be used to remove a capacity provider from a service's capacity provider
* strategy. When updating a service, the forceNewDeployment
option can be
* used to ensure that any tasks using the Amazon EC2 instance capacity provided by the capacity
* provider are transitioned to use the capacity from the remaining capacity providers.
* Only capacity providers that aren't associated with a cluster can be deleted. To remove
- * a capacity provider from a cluster, you can either use PutClusterCapacityProviders or delete the cluster.
INACTIVE
clusters persisting.
* You must deregister all container instances from this cluster before you may delete - * it. You can list the container instances in a cluster with ListContainerInstances and deregister them with DeregisterContainerInstance.
+ * it. You can list the container instances in a cluster with ListContainerInstances and deregister them with DeregisterContainerInstance. * @example * Use a bare-bones client and the command you need to make an API call. * ```javascript @@ -145,18 +145,18 @@ export interface DeleteClusterCommandOutput extends DeleteClusterResponse, __Met * @throws {@link ClusterContainsContainerInstancesException} (client fault) *You can't delete a cluster that has registered container instances. First, deregister * the container instances before you can delete the cluster. For more information, see - * DeregisterContainerInstance.
+ * DeregisterContainerInstance. * * @throws {@link ClusterContainsServicesException} (client fault) *You can't delete a cluster that contains services. First, update the service to reduce * its desired task count to 0, and then delete the service. For more information, see - * UpdateService and DeleteService.
+ * UpdateService and DeleteService. * * @throws {@link ClusterContainsTasksException} (client fault) *You can't delete a cluster that has active tasks.
* * @throws {@link ClusterNotFoundException} (client fault) - *The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
+ *The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
* * @throws {@link InvalidParameterException} (client fault) *The specified parameter isn't valid. Review the available parameters for the API diff --git a/clients/client-ecs/src/commands/DeleteServiceCommand.ts b/clients/client-ecs/src/commands/DeleteServiceCommand.ts index c6c509129332..4acb3d7149d8 100644 --- a/clients/client-ecs/src/commands/DeleteServiceCommand.ts +++ b/clients/client-ecs/src/commands/DeleteServiceCommand.ts @@ -31,17 +31,17 @@ export interface DeleteServiceCommandOutput extends DeleteServiceResponse, __Met *
Deletes a specified service within a cluster. You can delete a service if you have no * running tasks in it and the desired task count is zero. If the service is actively * maintaining tasks, you can't delete it, and you must update the service to a desired - * task count of zero. For more information, see UpdateService.
+ * task count of zero. For more information, see UpdateService. *When you delete a service, if there are still running tasks that require cleanup,
* the service status moves from ACTIVE
to DRAINING
, and the
- * service is no longer visible in the console or in the ListServices
+ * service is no longer visible in the console or in the ListServices
* API operation. After all tasks have transitioned to either STOPPING
or
* STOPPED
status, the service status moves from DRAINING
* to INACTIVE
. Services in the DRAINING
or
- * INACTIVE
status can still be viewed with the DescribeServices API operation. However, in the future,
+ * INACTIVE
status can still be viewed with the DescribeServices API operation. However, in the future,
* INACTIVE
services may be cleaned up and purged from Amazon ECS record
- * keeping, and DescribeServices calls on those services return a
+ * keeping, and DescribeServices calls on those services return a
* ServiceNotFoundException
error.
The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
+ *The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
* * @throws {@link InvalidParameterException} (client fault) *The specified parameter isn't valid. Review the available parameters for the API @@ -370,7 +370,7 @@ export interface DeleteServiceCommandOutput extends DeleteServiceResponse, __Met *
These errors are usually caused by a server issue.
* * @throws {@link ServiceNotFoundException} (client fault) - *The specified service wasn't found. You can view your available services with ListServices. Amazon ECS services are cluster specific and Region + *
The specified service wasn't found. You can view your available services with ListServices. Amazon ECS services are cluster specific and Region * specific.
* * @throws {@link ECSServiceException} diff --git a/clients/client-ecs/src/commands/DeleteTaskSetCommand.ts b/clients/client-ecs/src/commands/DeleteTaskSetCommand.ts index 1c1ce3394214..4a86874d86db 100644 --- a/clients/client-ecs/src/commands/DeleteTaskSetCommand.ts +++ b/clients/client-ecs/src/commands/DeleteTaskSetCommand.ts @@ -141,7 +141,7 @@ export interface DeleteTaskSetCommandOutput extends DeleteTaskSetResponse, __Met * * * @throws {@link ClusterNotFoundException} (client fault) - *The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
+ *The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
* * @throws {@link InvalidParameterException} (client fault) *The specified parameter isn't valid. Review the available parameters for the API @@ -152,14 +152,14 @@ export interface DeleteTaskSetCommandOutput extends DeleteTaskSetResponse, __Met * * @throws {@link ServiceNotActiveException} (client fault) *
The specified service isn't active. You can't update a service that's inactive. If you - * have previously deleted a service, you can re-create it with CreateService.
+ * have previously deleted a service, you can re-create it with CreateService. * * @throws {@link ServiceNotFoundException} (client fault) - *The specified service wasn't found. You can view your available services with ListServices. Amazon ECS services are cluster specific and Region + *
The specified service wasn't found. You can view your available services with ListServices. Amazon ECS services are cluster specific and Region * specific.
* * @throws {@link TaskSetNotFoundException} (client fault) - *The specified task set wasn't found. You can view your available task sets with DescribeTaskSets. Task sets are specific to each cluster, service and + *
The specified task set wasn't found. You can view your available task sets with DescribeTaskSets. Task sets are specific to each cluster, service and * Region.
* * @throws {@link UnsupportedFeatureException} (client fault) diff --git a/clients/client-ecs/src/commands/DeregisterContainerInstanceCommand.ts b/clients/client-ecs/src/commands/DeregisterContainerInstanceCommand.ts index dc8a65a6822e..d5a5f2c60b95 100644 --- a/clients/client-ecs/src/commands/DeregisterContainerInstanceCommand.ts +++ b/clients/client-ecs/src/commands/DeregisterContainerInstanceCommand.ts @@ -164,7 +164,7 @@ export interface DeregisterContainerInstanceCommandOutput * * * @throws {@link ClusterNotFoundException} (client fault) - *The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
+ *The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
* * @throws {@link InvalidParameterException} (client fault) *The specified parameter isn't valid. Review the available parameters for the API diff --git a/clients/client-ecs/src/commands/DescribeContainerInstancesCommand.ts b/clients/client-ecs/src/commands/DescribeContainerInstancesCommand.ts index 8223837d8660..8688037f156f 100644 --- a/clients/client-ecs/src/commands/DescribeContainerInstancesCommand.ts +++ b/clients/client-ecs/src/commands/DescribeContainerInstancesCommand.ts @@ -163,7 +163,7 @@ export interface DescribeContainerInstancesCommandOutput extends DescribeContain * * * @throws {@link ClusterNotFoundException} (client fault) - *
The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
+ *The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
* * @throws {@link InvalidParameterException} (client fault) *The specified parameter isn't valid. Review the available parameters for the API diff --git a/clients/client-ecs/src/commands/DescribeServicesCommand.ts b/clients/client-ecs/src/commands/DescribeServicesCommand.ts index 85d3b70d7ce4..fd970491c45a 100644 --- a/clients/client-ecs/src/commands/DescribeServicesCommand.ts +++ b/clients/client-ecs/src/commands/DescribeServicesCommand.ts @@ -353,7 +353,7 @@ export interface DescribeServicesCommandOutput extends DescribeServicesResponse, * * * @throws {@link ClusterNotFoundException} (client fault) - *
The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
+ *The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
* * @throws {@link InvalidParameterException} (client fault) *The specified parameter isn't valid. Review the available parameters for the API diff --git a/clients/client-ecs/src/commands/DescribeTaskSetsCommand.ts b/clients/client-ecs/src/commands/DescribeTaskSetsCommand.ts index 97885c31ede6..3493e67a1a7a 100644 --- a/clients/client-ecs/src/commands/DescribeTaskSetsCommand.ts +++ b/clients/client-ecs/src/commands/DescribeTaskSetsCommand.ts @@ -156,7 +156,7 @@ export interface DescribeTaskSetsCommandOutput extends DescribeTaskSetsResponse, * * * @throws {@link ClusterNotFoundException} (client fault) - *
The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
+ *The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
* * @throws {@link InvalidParameterException} (client fault) *The specified parameter isn't valid. Review the available parameters for the API @@ -167,10 +167,10 @@ export interface DescribeTaskSetsCommandOutput extends DescribeTaskSetsResponse, * * @throws {@link ServiceNotActiveException} (client fault) *
The specified service isn't active. You can't update a service that's inactive. If you - * have previously deleted a service, you can re-create it with CreateService.
+ * have previously deleted a service, you can re-create it with CreateService. * * @throws {@link ServiceNotFoundException} (client fault) - *The specified service wasn't found. You can view your available services with ListServices. Amazon ECS services are cluster specific and Region + *
The specified service wasn't found. You can view your available services with ListServices. Amazon ECS services are cluster specific and Region * specific.
* * @throws {@link UnsupportedFeatureException} (client fault) diff --git a/clients/client-ecs/src/commands/DescribeTasksCommand.ts b/clients/client-ecs/src/commands/DescribeTasksCommand.ts index 17384f39ecd1..47f443d27c61 100644 --- a/clients/client-ecs/src/commands/DescribeTasksCommand.ts +++ b/clients/client-ecs/src/commands/DescribeTasksCommand.ts @@ -246,7 +246,7 @@ export interface DescribeTasksCommandOutput extends DescribeTasksResponse, __Met * * * @throws {@link ClusterNotFoundException} (client fault) - *The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
+ *The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
* * @throws {@link InvalidParameterException} (client fault) *The specified parameter isn't valid. Review the available parameters for the API diff --git a/clients/client-ecs/src/commands/ExecuteCommandCommand.ts b/clients/client-ecs/src/commands/ExecuteCommandCommand.ts index 71bea0bfead7..32aabce2c0e4 100644 --- a/clients/client-ecs/src/commands/ExecuteCommandCommand.ts +++ b/clients/client-ecs/src/commands/ExecuteCommandCommand.ts @@ -95,7 +95,7 @@ export interface ExecuteCommandCommandOutput extends ExecuteCommandResponse, __M * * * @throws {@link ClusterNotFoundException} (client fault) - *
The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
+ *The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
* * @throws {@link InvalidParameterException} (client fault) *The specified parameter isn't valid. Review the available parameters for the API diff --git a/clients/client-ecs/src/commands/GetTaskProtectionCommand.ts b/clients/client-ecs/src/commands/GetTaskProtectionCommand.ts index 2e993c5f38fa..454c13e2373e 100644 --- a/clients/client-ecs/src/commands/GetTaskProtectionCommand.ts +++ b/clients/client-ecs/src/commands/GetTaskProtectionCommand.ts @@ -87,7 +87,7 @@ export interface GetTaskProtectionCommandOutput extends GetTaskProtectionRespons * * * @throws {@link ClusterNotFoundException} (client fault) - *
The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
+ *The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
* * @throws {@link InvalidParameterException} (client fault) *The specified parameter isn't valid. Review the available parameters for the API diff --git a/clients/client-ecs/src/commands/ListAttributesCommand.ts b/clients/client-ecs/src/commands/ListAttributesCommand.ts index 0ac9ee6f3dd1..cda5462c84e3 100644 --- a/clients/client-ecs/src/commands/ListAttributesCommand.ts +++ b/clients/client-ecs/src/commands/ListAttributesCommand.ts @@ -72,7 +72,7 @@ export interface ListAttributesCommandOutput extends ListAttributesResponse, __M * @see {@link ECSClientResolvedConfig | config} for ECSClient's `config` shape. * * @throws {@link ClusterNotFoundException} (client fault) - *
The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
+ *The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
* * @throws {@link InvalidParameterException} (client fault) *The specified parameter isn't valid. Review the available parameters for the API diff --git a/clients/client-ecs/src/commands/ListContainerInstancesCommand.ts b/clients/client-ecs/src/commands/ListContainerInstancesCommand.ts index 837a7433d1ea..f24e4fce3efa 100644 --- a/clients/client-ecs/src/commands/ListContainerInstancesCommand.ts +++ b/clients/client-ecs/src/commands/ListContainerInstancesCommand.ts @@ -77,7 +77,7 @@ export interface ListContainerInstancesCommandOutput extends ListContainerInstan * * * @throws {@link ClusterNotFoundException} (client fault) - *
The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
+ *The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
* * @throws {@link InvalidParameterException} (client fault) *The specified parameter isn't valid. Review the available parameters for the API diff --git a/clients/client-ecs/src/commands/ListServicesCommand.ts b/clients/client-ecs/src/commands/ListServicesCommand.ts index 994d47829991..9a019891f841 100644 --- a/clients/client-ecs/src/commands/ListServicesCommand.ts +++ b/clients/client-ecs/src/commands/ListServicesCommand.ts @@ -76,7 +76,7 @@ export interface ListServicesCommandOutput extends ListServicesResponse, __Metad * * * @throws {@link ClusterNotFoundException} (client fault) - *
The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
+ *The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
* * @throws {@link InvalidParameterException} (client fault) *The specified parameter isn't valid. Review the available parameters for the API diff --git a/clients/client-ecs/src/commands/ListTagsForResourceCommand.ts b/clients/client-ecs/src/commands/ListTagsForResourceCommand.ts index 0c552979b27a..4e2055cecd58 100644 --- a/clients/client-ecs/src/commands/ListTagsForResourceCommand.ts +++ b/clients/client-ecs/src/commands/ListTagsForResourceCommand.ts @@ -73,7 +73,7 @@ export interface ListTagsForResourceCommandOutput extends ListTagsForResourceRes * * * @throws {@link ClusterNotFoundException} (client fault) - *
The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
+ *The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
* * @throws {@link InvalidParameterException} (client fault) *The specified parameter isn't valid. Review the available parameters for the API diff --git a/clients/client-ecs/src/commands/ListTasksCommand.ts b/clients/client-ecs/src/commands/ListTasksCommand.ts index 164c24934301..7397a0df32a6 100644 --- a/clients/client-ecs/src/commands/ListTasksCommand.ts +++ b/clients/client-ecs/src/commands/ListTasksCommand.ts @@ -82,7 +82,7 @@ export interface ListTasksCommandOutput extends ListTasksResponse, __MetadataBea * * * @throws {@link ClusterNotFoundException} (client fault) - *
The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
+ *The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
* * @throws {@link InvalidParameterException} (client fault) *The specified parameter isn't valid. Review the available parameters for the API @@ -92,7 +92,7 @@ export interface ListTasksCommandOutput extends ListTasksResponse, __MetadataBea *
These errors are usually caused by a server issue.
* * @throws {@link ServiceNotFoundException} (client fault) - *The specified service wasn't found. You can view your available services with ListServices. Amazon ECS services are cluster specific and Region + *
The specified service wasn't found. You can view your available services with ListServices. Amazon ECS services are cluster specific and Region * specific.
* * @throws {@link ECSServiceException} diff --git a/clients/client-ecs/src/commands/PutAttributesCommand.ts b/clients/client-ecs/src/commands/PutAttributesCommand.ts index 97e7086e9d85..fc7ed524eb7c 100644 --- a/clients/client-ecs/src/commands/PutAttributesCommand.ts +++ b/clients/client-ecs/src/commands/PutAttributesCommand.ts @@ -30,7 +30,7 @@ export interface PutAttributesCommandOutput extends PutAttributesResponse, __Met /** *Create or update an attribute on an Amazon ECS resource. If the attribute doesn't exist, * it's created. If the attribute exists, its value is replaced with the specified value. - * To delete an attribute, use DeleteAttributes. For more information, + * To delete an attribute, use DeleteAttributes. For more information, * see Attributes in the Amazon Elastic Container Service Developer Guide.
* @example * Use a bare-bones client and the command you need to make an API call. @@ -72,11 +72,11 @@ export interface PutAttributesCommandOutput extends PutAttributesResponse, __Met * * @throws {@link AttributeLimitExceededException} (client fault) *You can apply up to 10 custom attributes for each resource. You can view the - * attributes of a resource with ListAttributes. You can remove existing - * attributes on a resource with DeleteAttributes.
+ * attributes of a resource with ListAttributes. You can remove existing + * attributes on a resource with DeleteAttributes. * * @throws {@link ClusterNotFoundException} (client fault) - *The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
+ *The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
* * @throws {@link InvalidParameterException} (client fault) *The specified parameter isn't valid. Review the available parameters for the API diff --git a/clients/client-ecs/src/commands/PutClusterCapacityProvidersCommand.ts b/clients/client-ecs/src/commands/PutClusterCapacityProvidersCommand.ts index 6540eba5601f..a08f09850cd6 100644 --- a/clients/client-ecs/src/commands/PutClusterCapacityProvidersCommand.ts +++ b/clients/client-ecs/src/commands/PutClusterCapacityProvidersCommand.ts @@ -36,7 +36,7 @@ export interface PutClusterCapacityProvidersCommandOutput * strategy for the cluster. If the specified cluster has existing capacity providers * associated with it, you must specify all existing capacity providers in addition to any * new ones you want to add. Any existing capacity providers that are associated with a - * cluster that are omitted from a PutClusterCapacityProviders API call + * cluster that are omitted from a PutClusterCapacityProviders API call * will be disassociated with the cluster. You can only disassociate an existing capacity * provider from a cluster if it's not being used by any existing tasks.
*When creating a service or running a task on a cluster, if no capacity provider or @@ -163,7 +163,7 @@ export interface PutClusterCapacityProvidersCommandOutput * * * @throws {@link ClusterNotFoundException} (client fault) - *
The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
+ *The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
* * @throws {@link InvalidParameterException} (client fault) *The specified parameter isn't valid. Review the available parameters for the API diff --git a/clients/client-ecs/src/commands/RegisterTaskDefinitionCommand.ts b/clients/client-ecs/src/commands/RegisterTaskDefinitionCommand.ts index 74ad0dd4d959..cd62c0962fe3 100644 --- a/clients/client-ecs/src/commands/RegisterTaskDefinitionCommand.ts +++ b/clients/client-ecs/src/commands/RegisterTaskDefinitionCommand.ts @@ -40,8 +40,7 @@ export interface RegisterTaskDefinitionCommandOutput extends RegisterTaskDefinit * Roles for Tasks in the Amazon Elastic Container Service Developer Guide.
*You can specify a Docker networking mode for the containers in your task definition
* with the networkMode
parameter. If you specify the awsvpc
- * network mode, the task is allocated an elastic network interface, and you must specify a
- * NetworkConfiguration when you create a service or run a task with
+ * network mode, the task is allocated an elastic network interface, and you must specify a NetworkConfiguration when you create a service or run a task with
* the task definition. For more information, see Task Networking
* in the Amazon Elastic Container Service Developer Guide.
You can allow Amazon ECS to place tasks for you, or you can customize how Amazon ECS places * tasks using placement constraints and placement strategies. For more information, see * Scheduling Tasks in the Amazon Elastic Container Service Developer Guide.
- *Alternatively, you can use StartTask to use your own scheduler or + *
Alternatively, you can use StartTask
to use your own scheduler or
* place tasks manually on specific container instances.
Starting April 15, 2023, Amazon Web Services will not onboard new customers to Amazon Elastic Inference (EI), and will help current customers migrate their workloads to options that offer better price and performance. After April 15, 2023, new customers will not be able to launch instances with Amazon EI accelerators in Amazon SageMaker, Amazon ECS, or Amazon EC2. However, customers who have used Amazon EI at least once during the past 30-day period are considered current customers and will be able to continue using the service.
*You can attach Amazon EBS volumes to Amazon ECS tasks by configuring the volume when creating or @@ -398,7 +398,7 @@ export interface RunTaskCommandOutput extends RunTaskResponse, __MetadataBearer * * * @throws {@link ClusterNotFoundException} (client fault) - *
The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
+ *The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
* * @throws {@link ConflictException} (client fault) *The RunTask
request could not be processed due to conflicts. The provided
diff --git a/clients/client-ecs/src/commands/StartTaskCommand.ts b/clients/client-ecs/src/commands/StartTaskCommand.ts
index 7d0794305f25..233c236bb3a1 100644
--- a/clients/client-ecs/src/commands/StartTaskCommand.ts
+++ b/clients/client-ecs/src/commands/StartTaskCommand.ts
@@ -34,7 +34,7 @@ export interface StartTaskCommandOutput extends StartTaskResponse, __MetadataBea
*
On March 21, 2024, a change was made to resolve the task definition revision before authorization. When a task definition revision is not specified, authorization will occur using the latest revision of a task definition.
*Starting April 15, 2023, Amazon Web Services will not onboard new customers to Amazon Elastic Inference (EI), and will help current customers migrate their workloads to options that offer better price and performance. After April 15, 2023, new customers will not be able to launch instances with Amazon EI accelerators in Amazon SageMaker, Amazon ECS, or Amazon EC2. However, customers who have used Amazon EI at least once during the past 30-day period are considered current customers and will be able to continue using the service.
- *Alternatively, you can use RunTask to place tasks for you. For more + *
Alternatively, you can useRunTask
to place tasks for you. For more
* information, see Scheduling Tasks in the Amazon Elastic Container Service Developer Guide.
You can attach Amazon EBS volumes to Amazon ECS tasks by configuring the volume when creating or * updating a service. For more infomation, see Amazon EBS volumes in the Amazon Elastic Container Service Developer Guide.
@@ -345,7 +345,7 @@ export interface StartTaskCommandOutput extends StartTaskResponse, __MetadataBea * * * @throws {@link ClusterNotFoundException} (client fault) - *The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
+ *The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
* * @throws {@link InvalidParameterException} (client fault) *The specified parameter isn't valid. Review the available parameters for the API diff --git a/clients/client-ecs/src/commands/StopTaskCommand.ts b/clients/client-ecs/src/commands/StopTaskCommand.ts index 99ba27e5e055..765a4bf8b74d 100644 --- a/clients/client-ecs/src/commands/StopTaskCommand.ts +++ b/clients/client-ecs/src/commands/StopTaskCommand.ts @@ -29,7 +29,7 @@ export interface StopTaskCommandOutput extends StopTaskResponse, __MetadataBeare /** *
Stops a running task. Any tags associated with the task will be deleted.
- *When StopTask is called on a task, the equivalent of When you call The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific. The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific. The specified parameter isn't valid. Review the available parameters for the API
diff --git a/clients/client-ecs/src/commands/TagResourceCommand.ts b/clients/client-ecs/src/commands/TagResourceCommand.ts
index c478cd833d67..3a12b67385ee 100644
--- a/clients/client-ecs/src/commands/TagResourceCommand.ts
+++ b/clients/client-ecs/src/commands/TagResourceCommand.ts
@@ -75,7 +75,7 @@ export interface TagResourceCommandOutput extends TagResourceResponse, __Metadat
*
*
* @throws {@link ClusterNotFoundException} (client fault)
- * The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific. The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific. The specified parameter isn't valid. Review the available parameters for the API
diff --git a/clients/client-ecs/src/commands/UntagResourceCommand.ts b/clients/client-ecs/src/commands/UntagResourceCommand.ts
index 44ca0c7f9a0e..eaa96e74ff46 100644
--- a/clients/client-ecs/src/commands/UntagResourceCommand.ts
+++ b/clients/client-ecs/src/commands/UntagResourceCommand.ts
@@ -69,7 +69,7 @@ export interface UntagResourceCommandOutput extends UntagResourceResponse, __Met
*
*
* @throws {@link ClusterNotFoundException} (client fault)
- * The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific. The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific. The specified parameter isn't valid. Review the available parameters for the API
diff --git a/clients/client-ecs/src/commands/UpdateClusterCommand.ts b/clients/client-ecs/src/commands/UpdateClusterCommand.ts
index 3f164d0118b2..d574b798561b 100644
--- a/clients/client-ecs/src/commands/UpdateClusterCommand.ts
+++ b/clients/client-ecs/src/commands/UpdateClusterCommand.ts
@@ -164,7 +164,7 @@ export interface UpdateClusterCommandOutput extends UpdateClusterResponse, __Met
*
*
* @throws {@link ClusterNotFoundException} (client fault)
- * The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific. The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific. The specified parameter isn't valid. Review the available parameters for the API
diff --git a/clients/client-ecs/src/commands/UpdateClusterSettingsCommand.ts b/clients/client-ecs/src/commands/UpdateClusterSettingsCommand.ts
index 4fdf0630a742..0cd16da20cab 100644
--- a/clients/client-ecs/src/commands/UpdateClusterSettingsCommand.ts
+++ b/clients/client-ecs/src/commands/UpdateClusterSettingsCommand.ts
@@ -144,7 +144,7 @@ export interface UpdateClusterSettingsCommandOutput extends UpdateClusterSetting
*
*
* @throws {@link ClusterNotFoundException} (client fault)
- * The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific. The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific. The specified parameter isn't valid. Review the available parameters for the API
diff --git a/clients/client-ecs/src/commands/UpdateContainerAgentCommand.ts b/clients/client-ecs/src/commands/UpdateContainerAgentCommand.ts
index c9cb8adfb9d0..8b993670936f 100644
--- a/clients/client-ecs/src/commands/UpdateContainerAgentCommand.ts
+++ b/clients/client-ecs/src/commands/UpdateContainerAgentCommand.ts
@@ -166,7 +166,7 @@ export interface UpdateContainerAgentCommandOutput extends UpdateContainerAgentR
*
*
* @throws {@link ClusterNotFoundException} (client fault)
- * The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific. The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific. The specified parameter isn't valid. Review the available parameters for the API
diff --git a/clients/client-ecs/src/commands/UpdateContainerInstancesStateCommand.ts b/clients/client-ecs/src/commands/UpdateContainerInstancesStateCommand.ts
index 52ba5580c629..90b257d23c1e 100644
--- a/clients/client-ecs/src/commands/UpdateContainerInstancesStateCommand.ts
+++ b/clients/client-ecs/src/commands/UpdateContainerInstancesStateCommand.ts
@@ -51,7 +51,7 @@ export interface UpdateContainerInstancesStateCommandOutput
* Service tasks on the container instance that are in the docker
+ *
StopTask
on a task, the equivalent of docker
* stop
is issued to the containers running in the task. This results in a
* SIGTERM
value and a default 30-second timeout, after which the
* SIGKILL
value is sent and the containers are forcibly stopped. If the
@@ -244,7 +244,7 @@ export interface StopTaskCommandOutput extends StopTaskResponse, __MetadataBeare
*
*
* @throws {@link ClusterNotFoundException} (client fault)
- * RUNNING
state are
* stopped and replaced according to the service's deployment configuration parameters,
* minimumHealthyPercent
and maximumPercent
. You can change
- * the deployment configuration of your service using UpdateService.
If minimumHealthyPercent
is below 100%, the scheduler can ignore
@@ -213,7 +213,7 @@ export interface UpdateContainerInstancesStateCommandOutput
*
The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
+ *The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
* * @throws {@link InvalidParameterException} (client fault) *The specified parameter isn't valid. Review the available parameters for the API diff --git a/clients/client-ecs/src/commands/UpdateServiceCommand.ts b/clients/client-ecs/src/commands/UpdateServiceCommand.ts index 9416699a2241..a527bfda2273 100644 --- a/clients/client-ecs/src/commands/UpdateServiceCommand.ts +++ b/clients/client-ecs/src/commands/UpdateServiceCommand.ts @@ -54,7 +54,7 @@ export interface UpdateServiceCommandOutput extends UpdateServiceResponse, __Met * count, task placement constraints and strategies, health check grace period, enable ECS * managed tags option, and propagate tags option, using this API. If the launch type, load * balancer, network configuration, platform version, or task definition need to be - * updated, create a new task set For more information, see CreateTaskSet.
+ * updated, create a new task set For more information, see CreateTaskSet. *You can add to or subtract from the number of instantiations of a task definition in a
* service by specifying the cluster that the service is running in and a new
* desiredCount
parameter.
When UpdateService stops a task during a deployment, the equivalent + *
When UpdateService stops a task during a deployment, the equivalent
* of docker stop
is issued to the containers running in the task. This
* results in a SIGTERM
and a 30-second timeout. After this,
* SIGKILL
is sent and the containers are forcibly stopped. If the
@@ -614,7 +614,7 @@ export interface UpdateServiceCommandOutput extends UpdateServiceResponse, __Met
*
*
* @throws {@link ClusterNotFoundException} (client fault)
- *
The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
+ *The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
* * @throws {@link InvalidParameterException} (client fault) *The specified parameter isn't valid. Review the available parameters for the API @@ -635,10 +635,10 @@ export interface UpdateServiceCommandOutput extends UpdateServiceResponse, __Met * * @throws {@link ServiceNotActiveException} (client fault) *
The specified service isn't active. You can't update a service that's inactive. If you - * have previously deleted a service, you can re-create it with CreateService.
+ * have previously deleted a service, you can re-create it with CreateService. * * @throws {@link ServiceNotFoundException} (client fault) - *The specified service wasn't found. You can view your available services with ListServices. Amazon ECS services are cluster specific and Region + *
The specified service wasn't found. You can view your available services with ListServices. Amazon ECS services are cluster specific and Region * specific.
* * @throws {@link UnsupportedFeatureException} (client fault) diff --git a/clients/client-ecs/src/commands/UpdateServicePrimaryTaskSetCommand.ts b/clients/client-ecs/src/commands/UpdateServicePrimaryTaskSetCommand.ts index 6794cba10a78..b12275305886 100644 --- a/clients/client-ecs/src/commands/UpdateServicePrimaryTaskSetCommand.ts +++ b/clients/client-ecs/src/commands/UpdateServicePrimaryTaskSetCommand.ts @@ -145,7 +145,7 @@ export interface UpdateServicePrimaryTaskSetCommandOutput * * * @throws {@link ClusterNotFoundException} (client fault) - *The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
+ *The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
* * @throws {@link InvalidParameterException} (client fault) *The specified parameter isn't valid. Review the available parameters for the API @@ -156,14 +156,14 @@ export interface UpdateServicePrimaryTaskSetCommandOutput * * @throws {@link ServiceNotActiveException} (client fault) *
The specified service isn't active. You can't update a service that's inactive. If you - * have previously deleted a service, you can re-create it with CreateService.
+ * have previously deleted a service, you can re-create it with CreateService. * * @throws {@link ServiceNotFoundException} (client fault) - *The specified service wasn't found. You can view your available services with ListServices. Amazon ECS services are cluster specific and Region + *
The specified service wasn't found. You can view your available services with ListServices. Amazon ECS services are cluster specific and Region * specific.
* * @throws {@link TaskSetNotFoundException} (client fault) - *The specified task set wasn't found. You can view your available task sets with DescribeTaskSets. Task sets are specific to each cluster, service and + *
The specified task set wasn't found. You can view your available task sets with DescribeTaskSets. Task sets are specific to each cluster, service and * Region.
* * @throws {@link UnsupportedFeatureException} (client fault) diff --git a/clients/client-ecs/src/commands/UpdateTaskProtectionCommand.ts b/clients/client-ecs/src/commands/UpdateTaskProtectionCommand.ts index 8c8af456b1ce..c32b011132cb 100644 --- a/clients/client-ecs/src/commands/UpdateTaskProtectionCommand.ts +++ b/clients/client-ecs/src/commands/UpdateTaskProtectionCommand.ts @@ -115,7 +115,7 @@ export interface UpdateTaskProtectionCommandOutput extends UpdateTaskProtectionR * * * @throws {@link ClusterNotFoundException} (client fault) - *The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
+ *The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
* * @throws {@link InvalidParameterException} (client fault) *The specified parameter isn't valid. Review the available parameters for the API diff --git a/clients/client-ecs/src/commands/UpdateTaskSetCommand.ts b/clients/client-ecs/src/commands/UpdateTaskSetCommand.ts index 5dac762f1c12..34a1413ed125 100644 --- a/clients/client-ecs/src/commands/UpdateTaskSetCommand.ts +++ b/clients/client-ecs/src/commands/UpdateTaskSetCommand.ts @@ -146,7 +146,7 @@ export interface UpdateTaskSetCommandOutput extends UpdateTaskSetResponse, __Met * * * @throws {@link ClusterNotFoundException} (client fault) - *
The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
+ *The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
* * @throws {@link InvalidParameterException} (client fault) *The specified parameter isn't valid. Review the available parameters for the API @@ -157,14 +157,14 @@ export interface UpdateTaskSetCommandOutput extends UpdateTaskSetResponse, __Met * * @throws {@link ServiceNotActiveException} (client fault) *
The specified service isn't active. You can't update a service that's inactive. If you - * have previously deleted a service, you can re-create it with CreateService.
+ * have previously deleted a service, you can re-create it with CreateService. * * @throws {@link ServiceNotFoundException} (client fault) - *The specified service wasn't found. You can view your available services with ListServices. Amazon ECS services are cluster specific and Region + *
The specified service wasn't found. You can view your available services with ListServices. Amazon ECS services are cluster specific and Region * specific.
* * @throws {@link TaskSetNotFoundException} (client fault) - *The specified task set wasn't found. You can view your available task sets with DescribeTaskSets. Task sets are specific to each cluster, service and + *
The specified task set wasn't found. You can view your available task sets with DescribeTaskSets. Task sets are specific to each cluster, service and * Region.
* * @throws {@link UnsupportedFeatureException} (client fault) diff --git a/clients/client-ecs/src/models/models_0.ts b/clients/client-ecs/src/models/models_0.ts index d59c3f8d774b..b5b1bae70acb 100644 --- a/clients/client-ecs/src/models/models_0.ts +++ b/clients/client-ecs/src/models/models_0.ts @@ -145,8 +145,7 @@ export interface ManagedScaling { minimumScalingStepSize?: number; /** - *The maximum number of Amazon EC2 instances that Amazon ECS will scale out at one time. The scale - * in process is not affected by this parameter. If this parameter is omitted, the default + *
The maximum number of Amazon EC2 instances that Amazon ECS will scale out at one time. If this parameter is omitted, the default
* value of 10000
is used.
The details of a capacity provider strategy. A capacity provider strategy can be set - * when using the RunTask or CreateCluster APIs or as - * the default capacity provider strategy for a cluster with the CreateCluster API.
+ * when using the RunTaskor CreateCluster APIs or as + * the default capacity provider strategy for a cluster with theCreateCluster
API.
* Only capacity providers that are already associated with a cluster and have an
* ACTIVE
or UPDATING
status can be used in a capacity
- * provider strategy. The PutClusterCapacityProviders API is used to
+ * provider strategy. The PutClusterCapacityProviders API is used to
* associate a capacity provider with a cluster.
If specifying a capacity provider that uses an Auto Scaling group, the capacity * provider must already be created. New Auto Scaling group capacity providers can be - * created with the CreateCapacityProvider API operation.
+ * created with the CreateClusterCapacityProvider API operation. *To use a Fargate capacity provider, specify either the FARGATE
or
* FARGATE_SPOT
capacity providers. The Fargate capacity providers are
* available to all accounts and only need to be associated with a cluster to be used in a
@@ -894,8 +893,8 @@ export interface CreateClusterRequest {
/**
*
The setting to use when creating a cluster. This parameter is used to turn on CloudWatch
* Container Insights for a cluster. If this value is specified, it overrides the
- * containerInsights
value set with PutAccountSetting or
- * PutAccountSettingDefault.
containerInsights
value set with PutAccountSetting or
+ * PutAccountSettingDefault.
* @public
*/
settings?: ClusterSetting[];
@@ -931,7 +930,7 @@ export interface CreateClusterRequest {
* capacity provider strategy or launch type specified, the default capacity provider
* strategy for the cluster is used.
* If a default capacity provider strategy isn't defined for a cluster when it was - * created, it can be defined later with the PutClusterCapacityProviders + * created, it can be defined later with the PutClusterCapacityProviders * API operation.
* @public */ @@ -1126,7 +1125,7 @@ export interface Cluster { /** *The number of services that are running on the cluster in an ACTIVE
- * state. You can view these services with ListServices.
The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
+ *The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
* @public */ export class ClusterNotFoundException extends __BaseException { @@ -1544,7 +1543,7 @@ export interface DeploymentController { * the current running version of the container with the latest version. The * number of containers Amazon ECS adds or removes from the service during a rolling * update is controlled by adjusting the minimum and maximum number of healthy - * tasks allowed during a service deployment, as specified in the DeploymentConfiguration. + * tasks allowed during a service deployment, as specified in the DeploymentConfiguration. * *The IDs of the subnets associated with the task or service. There's a limit of 16
- * subnets that can be specified per AwsVpcConfiguration
.
awsvpcConfiguration
.
* All specified subnets must be from the same VPC.
*The IDs of the security groups associated with the task or service. If you don't
* specify a security group, the default security group for the VPC is used. There's a
* limit of 5 security groups that can be specified per
- * AwsVpcConfiguration
.
awsvpcConfiguration
.
* All specified security groups must be from the same VPC.
*The log configuration for the container. This parameter maps to LogConfig
- * in the docker create-container command and the
+ * in the docker conainer create command and the
* --log-driver
option to docker
* run.
By default, containers use the same logging driver that the Docker daemon uses. @@ -2184,7 +2183,7 @@ export interface ServiceConnectConfiguration { /** *
The log configuration for the container. This parameter maps to LogConfig
- * in the docker create-container command and the
+ * in the docker conainer create command and the
* --log-driver
option to docker
* run.
By default, containers use the same logging driver that the Docker daemon uses.
@@ -3076,7 +3075,7 @@ export interface Deployment {
* deployment transitions to a COMPLETED
state. If the service fails to reach
* a steady state and circuit breaker is turned on, the deployment transitions to a
* FAILED
state. A deployment in FAILED
state doesn't launch
- * any new tasks. For more information, see DeploymentCircuitBreaker.
The desired number of instantiations of the task definition to keep running on the - * service. This value is specified when the service is created with CreateService, and it can be modified with UpdateService.
+ * service. This value is specified when the service is created with CreateService , and it can be modified with UpdateService. * @public */ desiredCount?: number; @@ -3550,8 +3549,8 @@ export interface Service { /** *The task definition to use for tasks in the service. This value is specified when the - * service is created with CreateService, and it can be modified with - * UpdateService.
+ * service is created with CreateService, and it can be modified with + * UpdateService. * @public */ taskDefinition?: string; @@ -3876,8 +3875,8 @@ export interface CreateTaskSetRequest { *The capacity provider strategy to use for the task set.
*A capacity provider strategy consists of one or more capacity providers along with the
* base
and weight
to assign to them. A capacity provider
- * must be associated with the cluster to be used in a capacity provider strategy. The
- * PutClusterCapacityProviders API is used to associate a capacity
+ * must be associated with the cluster to be used in a capacity provider strategy. The PutClusterCapacityProviders
+ * API is used to associate a capacity
* provider with a cluster. Only capacity providers with an ACTIVE
or
* UPDATING
status can be used.
If a capacityProviderStrategy
is specified, the launchType
@@ -3885,12 +3884,12 @@ export interface CreateTaskSetRequest {
* launchType
is specified, the
* defaultCapacityProviderStrategy
for the cluster is used.
If specifying a capacity provider that uses an Auto Scaling group, the capacity - * provider must already be created. New capacity providers can be created with the CreateCapacityProvider API operation.
+ * provider must already be created. New capacity providers can be created with the CreateCapacityProviderProviderAPI operation. *To use a Fargate capacity provider, specify either the FARGATE
or
* FARGATE_SPOT
capacity providers. The Fargate capacity providers are
* available to all accounts and only need to be associated with a cluster to be
* used.
The PutClusterCapacityProviders API operation is used to update the + *
The PutClusterCapacityProviders API operation is used to update the * list of available capacity providers for a cluster after the cluster is created.
* @public */ @@ -3974,7 +3973,7 @@ export interface CreateTaskSetResponse { /** *The specified service isn't active. You can't update a service that's inactive. If you - * have previously deleted a service, you can re-create it with CreateService.
+ * have previously deleted a service, you can re-create it with CreateService. * @public */ export class ServiceNotActiveException extends __BaseException { @@ -3994,7 +3993,7 @@ export class ServiceNotActiveException extends __BaseException { } /** - *The specified service wasn't found. You can view your available services with ListServices. Amazon ECS services are cluster specific and Region + *
The specified service wasn't found. You can view your available services with ListServices. Amazon ECS services are cluster specific and Region * specific.
* @public */ @@ -4252,7 +4251,7 @@ export interface DeleteCapacityProviderResponse { /** *You can't delete a cluster that has registered container instances. First, deregister * the container instances before you can delete the cluster. For more information, see - * DeregisterContainerInstance.
+ * DeregisterContainerInstance. * @public */ export class ClusterContainsContainerInstancesException extends __BaseException { @@ -4274,7 +4273,7 @@ export class ClusterContainsContainerInstancesException extends __BaseException /** *You can't delete a cluster that contains services. First, update the service to reduce * its desired task count to 0, and then delete the service. For more information, see - * UpdateService and DeleteService.
+ * UpdateService and DeleteService. * @public */ export class ClusterContainsServicesException extends __BaseException { @@ -4583,7 +4582,7 @@ export interface EnvironmentFile { /** *Hostnames and IP address entries that are added to the /etc/hosts
file of
- * a container via the extraHosts
parameter of its ContainerDefinition.
extraHosts
parameter of its ContainerDefinition.
* @public
*/
export interface HostEntry {
@@ -4806,7 +4805,7 @@ export interface HealthCheck {
* CMD-SHELL, curl -f http://localhost/ || exit 1
*
* An exit code of 0 indicates success, and non-zero exit code indicates failure. For
- * more information, see HealthCheck
in tthe docker create-container command
HealthCheck
in tthe docker conainer create command
* @public
*/
command: string[] | undefined;
@@ -4855,7 +4854,7 @@ export interface HealthCheck {
export interface KernelCapabilities {
/**
* The Linux capabilities for the container that have been added to the default
- * configuration provided by Docker. This parameter maps to CapAdd
in the docker create-container command and the
+ * configuration provided by Docker. This parameter maps to CapAdd
in the docker conainer create command and the
* --cap-add
option to docker
* run.
The Linux capabilities for the container that have been removed from the default
- * configuration provided by Docker. This parameter maps to CapDrop
in the docker create-container command and the
+ * configuration provided by Docker. This parameter maps to CapDrop
in the docker conainer create command and the
* --cap-drop
option to docker
* run.
Valid values: Any host devices to expose to the container. This parameter maps to
- * "ALL" | "AUDIT_CONTROL" | "AUDIT_WRITE" | "BLOCK_SUSPEND" |
@@ -4986,7 +4985,7 @@ export interface LinuxParameters {
/**
*
Devices
in tthe docker create-container command and the --device
option to docker run.Devices
in tthe docker conainer create command and the --device
option to docker run.
If you're using tasks that use the Fargate launch type, the
* devices
parameter isn't supported.
Run an init
process inside the container that forwards signals and reaps
- * processes. This parameter maps to the --init
option to docker run. This parameter requires version 1.25 of the Docker Remote API or greater on your container instance. To check the Docker Remote API version on your container instance, log in to your container instance and run the following command: sudo docker version --format '\{\{.Server.APIVersion\}\}'
+ * processes. This parameter maps to the --init
option to docker run.
+ * This parameter requires version 1.25 of the Docker Remote API or greater on your container instance. To check the Docker Remote API version on your container instance, log in to your container instance and run the following command: sudo docker version --format '\{\{.Server.APIVersion\}\}'
*
The total amount of swap memory (in MiB) a container can use. This parameter will be
- * translated to the --memory-swap
option to docker run where the value would be the sum of
+ * translated to the --memory-swap
option to docker run where the value would be the sum of
* the container memory plus the maxSwap
value.
If a maxSwap
value of 0
is specified, the container will not
* use swap. Accepted values are 0
or any positive integer. If the
@@ -5130,7 +5130,7 @@ export type TransportProtocol = (typeof TransportProtocol)[keyof typeof Transpor
* hostPort
can be left blank or it must be the same value as the
* containerPort
.
Most fields of this parameter (containerPort
, hostPort
,
- * protocol
) maps to PortBindings
in the docker create-container command and the
+ * protocol
) maps to PortBindings
in the docker conainer create command and the
* --publish
option to docker
* run
. If the network mode of a task definition is set to
* host
, host ports must either be undefined or match the container port
@@ -5141,7 +5141,7 @@ export type TransportProtocol = (typeof TransportProtocol)[keyof typeof Transpor
*
After a task reaches the RUNNING
status, manual and automatic host and
* container port assignments are visible in the networkBindings
section of
- * DescribeTasks API responses.
A list of namespaced kernel parameters to set in the container. This parameter maps to
- * Sysctls
in tthe docker create-container command and the --sysctl
option to docker run. For example, you can configure
+ * Sysctls
in tthe docker conainer create command and the --sysctl
option to docker run. For example, you can configure
* net.ipv4.tcp_keepalive_time
setting to maintain longer lived
* connections.
We don't recommend that you specify network-related systemControls
@@ -5562,7 +5562,7 @@ export interface ContainerDefinition {
*
The name of a container. If you're linking multiple containers together in a task
* definition, the name
of one container can be entered in the
* links
of another container to connect the containers.
- * Up to 255 letters (uppercase and lowercase), numbers, underscores, and hyphens are allowed. This parameter maps to name
in tthe docker create-container command and the
+ * Up to 255 letters (uppercase and lowercase), numbers, underscores, and hyphens are allowed. This parameter maps to name
in tthe docker conainer create command and the
* --name
option to docker
* run.
* repository-url/image@digest
- *
. Up to 255 letters (uppercase and lowercase), numbers, hyphens, underscores, colons, periods, forward slashes, and number signs are allowed. This parameter maps to Image
in the docker create-container command and the
+ * . Up to 255 letters (uppercase and lowercase), numbers, hyphens, underscores, colons, periods, forward slashes, and number signs are allowed. This parameter maps to Image
in the docker conainer create command and the
* IMAGE
parameter of docker
* run.
* The number of cpu
units reserved for the container. This parameter maps
- * to CpuShares
in the docker create-container commandand the --cpu-shares
option to docker run.
CpuShares
in the docker conainer create commandand the --cpu-shares
option to docker run.
* This field is optional for tasks using the Fargate launch type, and the
* only requirement is that the total amount of CPU reserved for all containers within a
* task be lower than the task-level cpu
value.
memory
value, if one is specified. This parameter maps to
- * Memory
in thethe docker create-container command and the --memory
option to docker run.
+ * Memory
in thethe docker conainer create command and the --memory
option to docker run.
* If using the Fargate launch type, this parameter is optional.
*If using the EC2 launch type, you must specify either a task-level
* memory value or a container-level memory value. If you specify both a container-level
@@ -5700,7 +5700,7 @@ export interface ContainerDefinition {
* However, your container can consume more memory when it needs to, up to either the hard
* limit specified with the memory
parameter (if applicable), or all of the
* available memory on the container instance, whichever comes first. This parameter maps
- * to MemoryReservation
in the the docker create-container command and the --memory-reservation
option to docker run.
MemoryReservation
in the the docker conainer create command and the --memory-reservation
option to docker run.
* If a task-level memory value is not specified, you must specify a non-zero integer for
* one or both of memory
or memoryReservation
in a container
* definition. If you specify both, memory
must be greater than
@@ -5727,7 +5727,7 @@ export interface ContainerDefinition {
* without the need for port mappings. This parameter is only supported if the network mode
* of a task definition is bridge
. The name:internalName
* construct is analogous to name:alias
in Docker links.
- * Up to 255 letters (uppercase and lowercase), numbers, underscores, and hyphens are allowed.. This parameter maps to Links
in the docker create-container command and the
+ * Up to 255 letters (uppercase and lowercase), numbers, underscores, and hyphens are allowed.. This parameter maps to Links
in the docker conainer create command and the
* --link
option to docker
* run.
localhost
. There's no loopback for port mappings on Windows, so you
* can't access a container's mapped port from the host itself.
* This parameter maps to PortBindings
in the
- * the docker create-container command and the
+ * the docker conainer create command and the
* --publish
option to docker
* run. If the network mode of a task definition is set to none
,
* then you can't specify port mappings. If the network mode of a task definition is set to
@@ -5764,7 +5764,7 @@ export interface ContainerDefinition {
* and container port assignments are visible in the Network
* Bindings section of a container description for a selected task in
* the Amazon ECS console. The assignments are also visible in the
- * networkBindings
section DescribeTasks
+ * networkBindings
section DescribeTasks
* responses.
command
array items instead.
*
* The entry point that's passed to the container. This parameter maps to
- * Entrypoint
in tthe docker create-container command and the --entrypoint
option to docker run.
Entrypoint
in tthe docker conainer create command and the --entrypoint
option to docker run.
* @public
*/
entryPoint?: string[];
/**
* The command that's passed to the container. This parameter maps to Cmd
in
- * the docker create-container command and the
+ * the docker conainer create command and the
* COMMAND
parameter to docker
* run. If there are multiple arguments, each
* argument is a separated string in the array.
The environment variables to pass to a container. This parameter maps to
- * Env
in the docker create-container command and the --env
option to docker run.
Env
in the docker conainer create command and the --env
option to docker run.
* We don't recommend that you use plaintext environment variables for sensitive * information, such as credential data.
@@ -5846,7 +5846,7 @@ export interface ContainerDefinition { /** *The mount points for data volumes in your container.
- *This parameter maps to Volumes
in the the docker create-container command and the --volume
option to docker run.
This parameter maps to Volumes
in the the docker conainer create command and the --volume
option to docker run.
Windows containers can mount whole directories on the same drive as
* $env:ProgramData
. Windows containers can't mount directories on a
* different drive, and mount point can't be across drives.
Data volumes to mount from another container. This parameter maps to
- * VolumesFrom
in tthe docker create-container command and the --volumes-from
option to docker run.
VolumesFrom
in tthe docker conainer create command and the --volumes-from
option to docker run.
* @public
*/
volumesFrom?: VolumeFrom[];
/**
* Linux-specific modifications that are applied to the container, such as Linux kernel - * capabilities. For more information see KernelCapabilities.
+ * capabilities. For more information see KernelCapabilities. *This parameter is not supported for Windows containers.
*The hostname to use for your container. This parameter maps to Hostname
- * in thethe docker create-container command and the
+ * in thethe docker conainer create command and the
* --hostname
option to docker
* run.
The user to use inside the container. This parameter maps to User
in the docker create-container command and the
+ *
The user to use inside the container. This parameter maps to User
in the docker conainer create command and the
* --user
option to docker
* run.
The working directory to run commands inside the container in. This parameter maps to
- * WorkingDir
in the docker create-container command and the --workdir
option to docker run.
WorkingDir
in the docker conainer create command and the --workdir
option to docker run.
* @public
*/
workingDirectory?: string;
/**
* When this parameter is true, networking is off within the container. This parameter
- * maps to NetworkDisabled
in the docker create-container command.
NetworkDisabled
in the docker conainer create command.
* This parameter is not supported for Windows containers.
*When this parameter is true, the container is given elevated privileges on the host
* container instance (similar to the root
user). This parameter maps to
- * Privileged
in the the docker create-container command and the --privileged
option to docker run
Privileged
in the the docker conainer create command and the --privileged
option to docker run
* This parameter is not supported for Windows containers or tasks run on Fargate.
*When this parameter is true, the container is given read-only access to its root file
- * system. This parameter maps to ReadonlyRootfs
in the docker create-container command and the
+ * system. This parameter maps to ReadonlyRootfs
in the docker conainer create command and the
* --read-only
option to docker
* run.
A list of DNS servers that are presented to the container. This parameter maps to
- * Dns
in the the docker create-container command and the --dns
option to docker run.
Dns
in the the docker conainer create command and the --dns
option to docker run.
* This parameter is not supported for Windows containers.
*A list of DNS search domains that are presented to the container. This parameter maps
- * to DnsSearch
in the docker create-container command and the --dns-search
option to docker run.
DnsSearch
in the docker conainer create command and the --dns-search
option to docker run.
* This parameter is not supported for Windows containers.
*A list of hostnames and IP address mappings to append to the /etc/hosts
- * file on the container. This parameter maps to ExtraHosts
in the docker create-container command and the
+ * file on the container. This parameter maps to ExtraHosts
in the docker conainer create command and the
* --add-host
option to docker
* run.
This parameter maps to SecurityOpt
in the docker create-container command and the
+ *
This parameter maps to SecurityOpt
in the docker conainer create command and the
* --security-opt
option to docker
* run.
When this parameter is true
, you can deploy containerized applications
* that require stdin
or a tty
to be allocated. This parameter
- * maps to OpenStdin
in the docker create-container command and the --interactive
option to docker run.
OpenStdin
in the docker conainer create command and the --interactive
option to docker run.
* @public
*/
interactive?: boolean;
/**
* When this parameter is true
, a TTY is allocated. This parameter maps to
- * Tty
in tthe docker create-container command and the --tty
option to docker run.
Tty
in tthe docker conainer create command and the --tty
option to docker run.
* @public
*/
pseudoTerminal?: boolean;
/**
* A key/value map of labels to add to the container. This parameter maps to
- * Labels
in the docker create-container command and the --label
option to docker run. This parameter requires version 1.18 of the Docker Remote API or greater on your container instance. To check the Docker Remote API version on your container instance, log in to your container instance and run the following command: sudo docker version --format '\{\{.Server.APIVersion\}\}'
+ * Labels
in the docker conainer create command and the --label
option to docker run. This parameter requires version 1.18 of the Docker Remote API or greater on your container instance. To check the Docker Remote API version on your container instance, log in to your container instance and run the following command: sudo docker version --format '\{\{.Server.APIVersion\}\}'
*
A list of ulimits
to set in the container. If a ulimit
value
* is specified in a task definition, it overrides the default values set by Docker. This
- * parameter maps to Ulimits
in tthe docker create-container command and the --ulimit
option to docker run. Valid naming values are displayed
- * in the Ulimit data type.
Ulimits
in tthe docker conainer create command and the --ulimit
option to docker run. Valid naming values are displayed
+ * in the Ulimit data type.
* Amazon ECS tasks hosted on Fargate use the default
* resource limit values set by the operating system with the exception of
* the nofile
resource limit parameter which Fargate
@@ -6185,7 +6185,7 @@ export interface ContainerDefinition {
/**
*
The log configuration specification for the container.
- *This parameter maps to LogConfig
in the docker create-container command and the
+ *
This parameter maps to LogConfig
in the docker conainer create command and the
* --log-driver
option to docker
* run. By default, containers use the same logging driver that the Docker
* daemon uses. However the container can use a different logging driver than the Docker
@@ -6195,7 +6195,7 @@ export interface ContainerDefinition {
* options).
Amazon ECS currently supports a subset of the logging drivers available to the Docker - * daemon (shown in the LogConfiguration data type). Additional log + * daemon (shown in the LogConfiguration data type). Additional log * drivers may be available in future releases of the Amazon ECS container agent.
*This parameter requires version 1.18 of the Docker Remote API or greater on your container instance. To check the Docker Remote API version on your container instance, log in to your container instance and run the following command: sudo docker version --format '\{\{.Server.APIVersion\}\}'
@@ -6214,7 +6214,7 @@ export interface ContainerDefinition {
/**
*
The container health check command and associated configuration parameters for the
- * container. This parameter maps to HealthCheck
in the docker create-container command and the
+ * container. This parameter maps to HealthCheck
in the docker conainer create command and the
* HEALTHCHECK
parameter of docker
* run.
A list of namespaced kernel parameters to set in the container. This parameter maps to
- * Sysctls
in tthe docker create-container command and the --sysctl
option to docker run. For example, you can configure
+ * Sysctls
in tthe docker conainer create command and the --sysctl
option to docker run. For example, you can configure
* net.ipv4.tcp_keepalive_time
setting to maintain longer lived
* connections.
docker plugin ls
to retrieve the driver name from
* your container instance. If the driver was installed using another method, use Docker
- * plugin discovery to retrieve the driver name. This parameter maps to Driver
in the docker create-container command and the
+ * plugin discovery to retrieve the driver name. This parameter maps to Driver
in the docker conainer create command and the
* xxdriver
option to docker
* volume create.
* @public
@@ -6641,7 +6641,7 @@ export interface DockerVolumeConfiguration {
/**
* Custom metadata to add to your Docker volume. This parameter maps to
- * Labels
in the docker create-container command and the xxlabel
option to docker
+ * Labels
in the docker conainer create command and the xxlabel
option to docker
* volume create.
The container instance attributes required by your task. When an Amazon EC2 instance is * registered to your cluster, the Amazon ECS container agent assigns some standard attributes * to the instance. You can apply custom attributes. These are specified as key-value pairs - * using the Amazon ECS console or the PutAttributes API. These attributes are + * using the Amazon ECS console or the PutAttributes API. These attributes are * used when determining task placement for tasks hosted on Amazon EC2 instances. For more * information, see Attributes in the Amazon Elastic Container Service Developer Guide.
*The specified task set wasn't found. You can view your available task sets with DescribeTaskSets. Task sets are specific to each cluster, service and + *
The specified task set wasn't found. You can view your available task sets with DescribeTaskSets. Task sets are specific to each cluster, service and * Region.
* @public */ @@ -7650,7 +7650,7 @@ export interface ContainerInstance { /** *The attributes set for the container instance, either by the Amazon ECS container agent at - * instance registration or manually with the PutAttributes + * instance registration or manually with the PutAttributes * operation.
* @public */ @@ -8221,7 +8221,7 @@ export interface ManagedAgent { *Details on the network bindings between a container and its host container instance.
* After a task reaches the RUNNING
status, manual and automatic host and
* container port assignments are visible in the networkBindings
section of
- * DescribeTasks API responses.
An object representing the protection status details for a task. You can set the - * protection status with the UpdateTaskProtection API and get the status - * of tasks with the GetTaskProtection API.
+ * protection status with the UpdateTaskProtection API and get the status + * of tasks with the GetTaskProtection API. * @public */ export interface ProtectedTask { @@ -9680,7 +9680,7 @@ export interface ListContainerInstancesRequest { /** *Filters the container instances by status. For example, if you specify the
* DRAINING
status, the results include only container instances that have
- * been set to DRAINING
using UpdateContainerInstancesState.
+ * been set to DRAINING
using UpdateContainerInstancesState.
* If you don't specify this parameter, the default is to include container instances set
* to all states other than INACTIVE
.
You can apply up to 10 custom attributes for each resource. You can view the - * attributes of a resource with ListAttributes. You can remove existing - * attributes on a resource with DeleteAttributes.
+ * attributes of a resource with ListAttributes. You can remove existing + * attributes on a resource with DeleteAttributes. * @public */ export class AttributeLimitExceededException extends __BaseException { @@ -10558,7 +10558,7 @@ export interface PutClusterCapacityProvidersRequest { /** *The name of one or more capacity providers to associate with the cluster.
*If specifying a capacity provider that uses an Auto Scaling group, the capacity - * provider must already be created. New capacity providers can be created with the CreateCapacityProvider API operation.
+ * provider must already be created. New capacity providers can be created with the CreateCapacityProvider API operation. *To use a Fargate capacity provider, specify either the FARGATE
or
* FARGATE_SPOT
capacity providers. The Fargate capacity providers are
* available to all accounts and only need to be associated with a cluster to be
@@ -10575,11 +10575,11 @@ export interface PutClusterCapacityProvidersRequest {
*
A capacity provider strategy consists of one or more capacity providers along with the
* base
and weight
to assign to them. A capacity provider
* must be associated with the cluster to be used in a capacity provider strategy. The
- * PutClusterCapacityProviders API is used to associate a capacity
+ * PutClusterCapacityProviders API is used to associate a capacity
* provider with a cluster. Only capacity providers with an ACTIVE
or
* UPDATING
status can be used.
If specifying a capacity provider that uses an Auto Scaling group, the capacity - * provider must already be created. New capacity providers can be created with the CreateCapacityProvider API operation.
+ * provider must already be created. New capacity providers can be created with the CreateCapacityProvider API operation. *To use a Fargate capacity provider, specify either the FARGATE
or
* FARGATE_SPOT
capacity providers. The Fargate capacity providers are
* available to all accounts and only need to be associated with a cluster to be
@@ -11529,7 +11529,7 @@ export interface RunTaskRequest {
/**
*
Specifies whether to propagate the tags from the task definition to the task. If no * value is specified, the tags aren't propagated. Tags can only be propagated to the task - * during task creation. To add tags to a task after task creation, use the TagResource API action.
+ * during task creation. To add tags to a task after task creation, use theTagResource API action. *An error will be received if you specify the SERVICE
option when
* running a task.
An optional message specified when a task is stopped. For example, if you're using a * custom scheduler, you can use this parameter to specify the reason for stopping the task - * here, and the message appears in subsequent DescribeTasks API + * here, and the message appears in subsequent DescribeTasks> API * operations on this task.
* @public */ @@ -12326,8 +12326,8 @@ export interface UpdateClusterSettingsRequest { /** *The setting to use by default for a cluster. This parameter is used to turn on CloudWatch
* Container Insights for a cluster. If this value is specified, it overrides the
- * containerInsights
value set with PutAccountSetting or
- * PutAccountSettingDefault.
containerInsights
value set with PutAccountSetting or
+ * PutAccountSettingDefault.
* Currently, if you delete an existing cluster that does not have Container Insights * turned on, and then create a new cluster with the same name with Container Insights @@ -12516,16 +12516,16 @@ export interface UpdateServiceRequest { *
A capacity provider strategy consists of one or more capacity providers along with the
* base
and weight
to assign to them. A capacity provider
* must be associated with the cluster to be used in a capacity provider strategy. The
- * PutClusterCapacityProviders API is used to associate a capacity
+ * PutClusterCapacityProviders API is used to associate a capacity
* provider with a cluster. Only capacity providers with an ACTIVE
or
* UPDATING
status can be used.
If specifying a capacity provider that uses an Auto Scaling group, the capacity - * provider must already be created. New capacity providers can be created with the CreateCapacityProvider API operation.
+ * provider must already be created. New capacity providers can be created with the CreateClusterCapacityProvider API operation. *To use a Fargate capacity provider, specify either the FARGATE
or
* FARGATE_SPOT
capacity providers. The Fargate capacity providers are
* available to all accounts and only need to be associated with a cluster to be
* used.
The PutClusterCapacityProviders API operation is used to update the + *
The PutClusterCapacityProvidersAPI operation is used to update the * list of available capacity providers for a cluster after the cluster is created.
* * @public diff --git a/codegen/sdk-codegen/aws-models/ecs.json b/codegen/sdk-codegen/aws-models/ecs.json index bfcc1fe5e498..854049db5bb2 100644 --- a/codegen/sdk-codegen/aws-models/ecs.json +++ b/codegen/sdk-codegen/aws-models/ecs.json @@ -1425,7 +1425,7 @@ } }, "traits": { - "smithy.api#documentation": "You can apply up to 10 custom attributes for each resource. You can view the\n\t\t\tattributes of a resource with ListAttributes. You can remove existing\n\t\t\tattributes on a resource with DeleteAttributes.
", + "smithy.api#documentation": "You can apply up to 10 custom attributes for each resource. You can view the\n\t\t\tattributes of a resource with ListAttributes. You can remove existing\n\t\t\tattributes on a resource with DeleteAttributes.
", "smithy.api#error": "client" } }, @@ -1500,14 +1500,14 @@ "subnets": { "target": "com.amazonaws.ecs#StringList", "traits": { - "smithy.api#documentation": "The IDs of the subnets associated with the task or service. There's a limit of 16\n\t\t\tsubnets that can be specified per AwsVpcConfiguration
.
All specified subnets must be from the same VPC.
\nThe IDs of the subnets associated with the task or service. There's a limit of 16\n\t\t\tsubnets that can be specified per awsvpcConfiguration
.
All specified subnets must be from the same VPC.
\nThe IDs of the security groups associated with the task or service. If you don't\n\t\t\tspecify a security group, the default security group for the VPC is used. There's a\n\t\t\tlimit of 5 security groups that can be specified per\n\t\t\tAwsVpcConfiguration
.
All specified security groups must be from the same VPC.
\nThe IDs of the security groups associated with the task or service. If you don't\n\t\t\tspecify a security group, the default security group for the VPC is used. There's a\n\t\t\tlimit of 5 security groups that can be specified per\n\t\t\tawsvpcConfiguration
.
All specified security groups must be from the same VPC.
\nThe details of a capacity provider strategy. A capacity provider strategy can be set\n\t\t\twhen using the RunTask or CreateCluster APIs or as\n\t\t\tthe default capacity provider strategy for a cluster with the CreateCluster API.
\nOnly capacity providers that are already associated with a cluster and have an\n\t\t\t\tACTIVE
or UPDATING
status can be used in a capacity\n\t\t\tprovider strategy. The PutClusterCapacityProviders API is used to\n\t\t\tassociate a capacity provider with a cluster.
If specifying a capacity provider that uses an Auto Scaling group, the capacity\n\t\t\tprovider must already be created. New Auto Scaling group capacity providers can be\n\t\t\tcreated with the CreateCapacityProvider API operation.
\nTo use a Fargate capacity provider, specify either the FARGATE
or\n\t\t\t\tFARGATE_SPOT
capacity providers. The Fargate capacity providers are\n\t\t\tavailable to all accounts and only need to be associated with a cluster to be used in a\n\t\t\tcapacity provider strategy.
With FARGATE_SPOT
, you can run interruption tolerant tasks at a rate\n\t\t\tthat's discounted compared to the FARGATE
price. FARGATE_SPOT
\n\t\t\truns tasks on spare compute capacity. When Amazon Web Services needs the capacity back, your tasks are\n\t\t\tinterrupted with a two-minute warning. FARGATE_SPOT
only supports Linux\n\t\t\ttasks with the X86_64 architecture on platform version 1.3.0 or later.
A capacity provider strategy may contain a maximum of 6 capacity providers.
" + "smithy.api#documentation": "The details of a capacity provider strategy. A capacity provider strategy can be set\n\t\t\twhen using the RunTaskor CreateCluster APIs or as\n\t\t\tthe default capacity provider strategy for a cluster with the CreateCluster
API.
Only capacity providers that are already associated with a cluster and have an\n\t\t\t\tACTIVE
or UPDATING
status can be used in a capacity\n\t\t\tprovider strategy. The PutClusterCapacityProviders API is used to\n\t\t\tassociate a capacity provider with a cluster.
If specifying a capacity provider that uses an Auto Scaling group, the capacity\n\t\t\tprovider must already be created. New Auto Scaling group capacity providers can be\n\t\t\tcreated with the CreateClusterCapacityProvider API operation.
\nTo use a Fargate capacity provider, specify either the FARGATE
or\n\t\t\t\tFARGATE_SPOT
capacity providers. The Fargate capacity providers are\n\t\t\tavailable to all accounts and only need to be associated with a cluster to be used in a\n\t\t\tcapacity provider strategy.
With FARGATE_SPOT
, you can run interruption tolerant tasks at a rate\n\t\t\tthat's discounted compared to the FARGATE
price. FARGATE_SPOT
\n\t\t\truns tasks on spare compute capacity. When Amazon Web Services needs the capacity back, your tasks are\n\t\t\tinterrupted with a two-minute warning. FARGATE_SPOT
only supports Linux\n\t\t\ttasks with the X86_64 architecture on platform version 1.3.0 or later.
A capacity provider strategy may contain a maximum of 6 capacity providers.
" } }, "com.amazonaws.ecs#CapacityProviderStrategyItemBase": { @@ -1818,7 +1818,7 @@ "target": "com.amazonaws.ecs#Integer", "traits": { "smithy.api#default": 0, - "smithy.api#documentation": "The number of services that are running on the cluster in an ACTIVE
\n\t\t\tstate. You can view these services with ListServices.
The number of services that are running on the cluster in an ACTIVE
\n\t\t\tstate. You can view these services with PListServices.
You can't delete a cluster that has registered container instances. First, deregister\n\t\t\tthe container instances before you can delete the cluster. For more information, see\n\t\t\t\tDeregisterContainerInstance.
", + "smithy.api#documentation": "You can't delete a cluster that has registered container instances. First, deregister\n\t\t\tthe container instances before you can delete the cluster. For more information, see\n\t\t\tDeregisterContainerInstance.
", "smithy.api#error": "client" } }, @@ -1920,7 +1920,7 @@ } }, "traits": { - "smithy.api#documentation": "You can't delete a cluster that contains services. First, update the service to reduce\n\t\t\tits desired task count to 0, and then delete the service. For more information, see\n\t\t\t\tUpdateService and DeleteService.
", + "smithy.api#documentation": "You can't delete a cluster that contains services. First, update the service to reduce\n\t\t\tits desired task count to 0, and then delete the service. For more information, see\n\t\t\tUpdateService and DeleteService.
", "smithy.api#error": "client" } }, @@ -1991,7 +1991,7 @@ } }, "traits": { - "smithy.api#documentation": "The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
", + "smithy.api#documentation": "The specified cluster wasn't found. You can view your available clusters with ListClusters. Amazon ECS clusters are Region specific.
", "smithy.api#error": "client" } }, @@ -2279,13 +2279,13 @@ "name": { "target": "com.amazonaws.ecs#String", "traits": { - "smithy.api#documentation": "The name of a container. If you're linking multiple containers together in a task\n\t\t\tdefinition, the name
of one container can be entered in the\n\t\t\t\tlinks
of another container to connect the containers.\n\t\t\tUp to 255 letters (uppercase and lowercase), numbers, underscores, and hyphens are allowed. This parameter maps to name
in tthe docker create-container command and the\n\t\t\t\t--name
option to docker\n\t\t\trun.
The name of a container. If you're linking multiple containers together in a task\n\t\t\tdefinition, the name
of one container can be entered in the\n\t\t\t\tlinks
of another container to connect the containers.\n\t\t\tUp to 255 letters (uppercase and lowercase), numbers, underscores, and hyphens are allowed. This parameter maps to name
in tthe docker conainer create command and the\n\t\t\t\t--name
option to docker\n\t\t\trun.
The image used to start a container. This string is passed directly to the Docker\n\t\t\tdaemon. By default, images in the Docker Hub registry are available. Other repositories\n\t\t\tare specified with either \n repository-url/image:tag\n
or \n repository-url/image@digest\n
. Up to 255 letters (uppercase and lowercase), numbers, hyphens, underscores, colons, periods, forward slashes, and number signs are allowed. This parameter maps to Image
in the docker create-container command and the\n\t\t\t\tIMAGE
parameter of docker\n\t\t\t\trun.
When a new task starts, the Amazon ECS container agent pulls the latest version of\n\t\t\t\t\tthe specified image and tag for the container to use. However, subsequent\n\t\t\t\t\tupdates to a repository image aren't propagated to already running tasks.
\nImages in Amazon ECR repositories can be specified by either using the full\n\t\t\t\t\t\tregistry/repository:tag
or\n\t\t\t\t\t\tregistry/repository@digest
. For example,\n\t\t\t\t\t\t012345678910.dkr.ecr.
\n\t\t\t\t\tor\n\t\t\t\t\t\t012345678910.dkr.ecr.
.\n\t\t\t\t
Images in official repositories on Docker Hub use a single name (for example,\n\t\t\t\t\t\tubuntu
or mongo
).
Images in other repositories on Docker Hub are qualified with an organization\n\t\t\t\t\tname (for example, amazon/amazon-ecs-agent
).
Images in other online repositories are qualified further by a domain name\n\t\t\t\t\t(for example, quay.io/assemblyline/ubuntu
).
The image used to start a container. This string is passed directly to the Docker\n\t\t\tdaemon. By default, images in the Docker Hub registry are available. Other repositories\n\t\t\tare specified with either \n repository-url/image:tag\n
or \n repository-url/image@digest\n
. Up to 255 letters (uppercase and lowercase), numbers, hyphens, underscores, colons, periods, forward slashes, and number signs are allowed. This parameter maps to Image
in the docker conainer create command and the\n\t\t\t\tIMAGE
parameter of docker\n\t\t\t\trun.
When a new task starts, the Amazon ECS container agent pulls the latest version of\n\t\t\t\t\tthe specified image and tag for the container to use. However, subsequent\n\t\t\t\t\tupdates to a repository image aren't propagated to already running tasks.
\nImages in Amazon ECR repositories can be specified by either using the full\n\t\t\t\t\t\tregistry/repository:tag
or\n\t\t\t\t\t\tregistry/repository@digest
. For example,\n\t\t\t\t\t\t012345678910.dkr.ecr.
\n\t\t\t\t\tor\n\t\t\t\t\t\t012345678910.dkr.ecr.
.\n\t\t\t\t
Images in official repositories on Docker Hub use a single name (for example,\n\t\t\t\t\t\tubuntu
or mongo
).
Images in other repositories on Docker Hub are qualified with an organization\n\t\t\t\t\tname (for example, amazon/amazon-ecs-agent
).
Images in other online repositories are qualified further by a domain name\n\t\t\t\t\t(for example, quay.io/assemblyline/ubuntu
).
The number of cpu
units reserved for the container. This parameter maps\n\t\t\tto CpuShares
in the docker create-container commandand the --cpu-shares
option to docker run.
This field is optional for tasks using the Fargate launch type, and the\n\t\t\tonly requirement is that the total amount of CPU reserved for all containers within a\n\t\t\ttask be lower than the task-level cpu
value.
You can determine the number of CPU units that are available per EC2 instance type\n\t\t\t\tby multiplying the vCPUs listed for that instance type on the Amazon EC2 Instances detail page\n\t\t\t\tby 1,024.
\nLinux containers share unallocated CPU units with other containers on the container\n\t\t\tinstance with the same ratio as their allocated amount. For example, if you run a\n\t\t\tsingle-container task on a single-core instance type with 512 CPU units specified for\n\t\t\tthat container, and that's the only task running on the container instance, that\n\t\t\tcontainer could use the full 1,024 CPU unit share at any given time. However, if you\n\t\t\tlaunched another copy of the same task on that container instance, each task is\n\t\t\tguaranteed a minimum of 512 CPU units when needed. Moreover, each container could float\n\t\t\tto higher CPU usage if the other container was not using it. If both tasks were 100%\n\t\t\tactive all of the time, they would be limited to 512 CPU units.
\nOn Linux container instances, the Docker daemon on the container instance uses the CPU\n\t\t\tvalue to calculate the relative CPU share ratios for running containers. The minimum valid CPU share value\n\t\t\tthat the Linux kernel allows is 2, and the\n\t\t\tmaximum valid CPU share value that the Linux kernel allows is 262144. However, the CPU parameter isn't required, and you\n\t\t\tcan use CPU values below 2 or above 262144 in your container definitions. For CPU values below 2\n\t\t\t(including null) or above 262144, the behavior varies based on your Amazon ECS container agent\n\t\t\tversion:
\n\n Agent versions less than or equal to 1.1.0:\n\t\t\t\t\tNull and zero CPU values are passed to Docker as 0, which Docker then converts\n\t\t\t\t\tto 1,024 CPU shares. CPU values of 1 are passed to Docker as 1, which the Linux\n\t\t\t\t\tkernel converts to two CPU shares.
\n\n Agent versions greater than or equal to 1.2.0:\n\t\t\t\t\tNull, zero, and CPU values of 1 are passed to Docker as 2.
\n\n Agent versions greater than or equal to\n\t\t\t\t\t\t1.84.0: CPU values greater than 256 vCPU are passed to Docker as\n\t\t\t\t\t256, which is equivalent to 262144 CPU shares.
\nOn Windows container instances, the CPU limit is enforced as an absolute limit, or a\n\t\t\tquota. Windows containers only have access to the specified amount of CPU that's\n\t\t\tdescribed in the task definition. A null or zero CPU value is passed to Docker as\n\t\t\t\t0
, which Windows interprets as 1% of one CPU.
The number of cpu
units reserved for the container. This parameter maps\n\t\t\tto CpuShares
in the docker conainer create commandand the --cpu-shares
option to docker run.
This field is optional for tasks using the Fargate launch type, and the\n\t\t\tonly requirement is that the total amount of CPU reserved for all containers within a\n\t\t\ttask be lower than the task-level cpu
value.
You can determine the number of CPU units that are available per EC2 instance type\n\t\t\t\tby multiplying the vCPUs listed for that instance type on the Amazon EC2 Instances detail page\n\t\t\t\tby 1,024.
\nLinux containers share unallocated CPU units with other containers on the container\n\t\t\tinstance with the same ratio as their allocated amount. For example, if you run a\n\t\t\tsingle-container task on a single-core instance type with 512 CPU units specified for\n\t\t\tthat container, and that's the only task running on the container instance, that\n\t\t\tcontainer could use the full 1,024 CPU unit share at any given time. However, if you\n\t\t\tlaunched another copy of the same task on that container instance, each task is\n\t\t\tguaranteed a minimum of 512 CPU units when needed. Moreover, each container could float\n\t\t\tto higher CPU usage if the other container was not using it. If both tasks were 100%\n\t\t\tactive all of the time, they would be limited to 512 CPU units.
\nOn Linux container instances, the Docker daemon on the container instance uses the CPU\n\t\t\tvalue to calculate the relative CPU share ratios for running containers. The minimum valid CPU share value\n\t\t\tthat the Linux kernel allows is 2, and the\n\t\t\tmaximum valid CPU share value that the Linux kernel allows is 262144. However, the CPU parameter isn't required, and you\n\t\t\tcan use CPU values below 2 or above 262144 in your container definitions. For CPU values below 2\n\t\t\t(including null) or above 262144, the behavior varies based on your Amazon ECS container agent\n\t\t\tversion:
\n\n Agent versions less than or equal to 1.1.0:\n\t\t\t\t\tNull and zero CPU values are passed to Docker as 0, which Docker then converts\n\t\t\t\t\tto 1,024 CPU shares. CPU values of 1 are passed to Docker as 1, which the Linux\n\t\t\t\t\tkernel converts to two CPU shares.
\n\n Agent versions greater than or equal to 1.2.0:\n\t\t\t\t\tNull, zero, and CPU values of 1 are passed to Docker as 2.
\n\n Agent versions greater than or equal to\n\t\t\t\t\t\t1.84.0: CPU values greater than 256 vCPU are passed to Docker as\n\t\t\t\t\t256, which is equivalent to 262144 CPU shares.
\nOn Windows container instances, the CPU limit is enforced as an absolute limit, or a\n\t\t\tquota. Windows containers only have access to the specified amount of CPU that's\n\t\t\tdescribed in the task definition. A null or zero CPU value is passed to Docker as\n\t\t\t\t0
, which Windows interprets as 1% of one CPU.
The amount (in MiB) of memory to present to the container. If your container attempts\n\t\t\tto exceed the memory specified here, the container is killed. The total amount of memory\n\t\t\treserved for all containers within a task must be lower than the task\n\t\t\t\tmemory
value, if one is specified. This parameter maps to\n\t\t\tMemory
in thethe docker create-container command and the --memory
option to docker run.
If using the Fargate launch type, this parameter is optional.
\nIf using the EC2 launch type, you must specify either a task-level\n\t\t\tmemory value or a container-level memory value. If you specify both a container-level\n\t\t\t\tmemory
and memoryReservation
value, memory
\n\t\t\tmust be greater than memoryReservation
. If you specify\n\t\t\t\tmemoryReservation
, then that value is subtracted from the available\n\t\t\tmemory resources for the container instance where the container is placed. Otherwise,\n\t\t\tthe value of memory
is used.
The Docker 20.10.0 or later daemon reserves a minimum of 6 MiB of memory for a\n\t\t\tcontainer. So, don't specify less than 6 MiB of memory for your containers.
\nThe Docker 19.03.13-ce or earlier daemon reserves a minimum of 4 MiB of memory for a\n\t\t\tcontainer. So, don't specify less than 4 MiB of memory for your containers.
" + "smithy.api#documentation": "The amount (in MiB) of memory to present to the container. If your container attempts\n\t\t\tto exceed the memory specified here, the container is killed. The total amount of memory\n\t\t\treserved for all containers within a task must be lower than the task\n\t\t\t\tmemory
value, if one is specified. This parameter maps to\n\t\t\tMemory
in thethe docker conainer create command and the --memory
option to docker run.
If using the Fargate launch type, this parameter is optional.
\nIf using the EC2 launch type, you must specify either a task-level\n\t\t\tmemory value or a container-level memory value. If you specify both a container-level\n\t\t\t\tmemory
and memoryReservation
value, memory
\n\t\t\tmust be greater than memoryReservation
. If you specify\n\t\t\t\tmemoryReservation
, then that value is subtracted from the available\n\t\t\tmemory resources for the container instance where the container is placed. Otherwise,\n\t\t\tthe value of memory
is used.
The Docker 20.10.0 or later daemon reserves a minimum of 6 MiB of memory for a\n\t\t\tcontainer. So, don't specify less than 6 MiB of memory for your containers.
\nThe Docker 19.03.13-ce or earlier daemon reserves a minimum of 4 MiB of memory for a\n\t\t\tcontainer. So, don't specify less than 4 MiB of memory for your containers.
" } }, "memoryReservation": { "target": "com.amazonaws.ecs#BoxedInteger", "traits": { - "smithy.api#documentation": "The soft limit (in MiB) of memory to reserve for the container. When system memory is\n\t\t\tunder heavy contention, Docker attempts to keep the container memory to this soft limit.\n\t\t\tHowever, your container can consume more memory when it needs to, up to either the hard\n\t\t\tlimit specified with the memory
parameter (if applicable), or all of the\n\t\t\tavailable memory on the container instance, whichever comes first. This parameter maps\n\t\t\tto MemoryReservation
in the the docker create-container command and the --memory-reservation
option to docker run.
If a task-level memory value is not specified, you must specify a non-zero integer for\n\t\t\tone or both of memory
or memoryReservation
in a container\n\t\t\tdefinition. If you specify both, memory
must be greater than\n\t\t\t\tmemoryReservation
. If you specify memoryReservation
, then\n\t\t\tthat value is subtracted from the available memory resources for the container instance\n\t\t\twhere the container is placed. Otherwise, the value of memory
is\n\t\t\tused.
For example, if your container normally uses 128 MiB of memory, but occasionally\n\t\t\tbursts to 256 MiB of memory for short periods of time, you can set a\n\t\t\t\tmemoryReservation
of 128 MiB, and a memory
hard limit of\n\t\t\t300 MiB. This configuration would allow the container to only reserve 128 MiB of memory\n\t\t\tfrom the remaining resources on the container instance, but also allow the container to\n\t\t\tconsume more memory resources when needed.
The Docker 20.10.0 or later daemon reserves a minimum of 6 MiB of memory for a\n\t\t\tcontainer. So, don't specify less than 6 MiB of memory for your containers.
\nThe Docker 19.03.13-ce or earlier daemon reserves a minimum of 4 MiB of memory for a\n\t\t\tcontainer. So, don't specify less than 4 MiB of memory for your containers.
" + "smithy.api#documentation": "The soft limit (in MiB) of memory to reserve for the container. When system memory is\n\t\t\tunder heavy contention, Docker attempts to keep the container memory to this soft limit.\n\t\t\tHowever, your container can consume more memory when it needs to, up to either the hard\n\t\t\tlimit specified with the memory
parameter (if applicable), or all of the\n\t\t\tavailable memory on the container instance, whichever comes first. This parameter maps\n\t\t\tto MemoryReservation
in the the docker conainer create command and the --memory-reservation
option to docker run.
If a task-level memory value is not specified, you must specify a non-zero integer for\n\t\t\tone or both of memory
or memoryReservation
in a container\n\t\t\tdefinition. If you specify both, memory
must be greater than\n\t\t\t\tmemoryReservation
. If you specify memoryReservation
, then\n\t\t\tthat value is subtracted from the available memory resources for the container instance\n\t\t\twhere the container is placed. Otherwise, the value of memory
is\n\t\t\tused.
For example, if your container normally uses 128 MiB of memory, but occasionally\n\t\t\tbursts to 256 MiB of memory for short periods of time, you can set a\n\t\t\t\tmemoryReservation
of 128 MiB, and a memory
hard limit of\n\t\t\t300 MiB. This configuration would allow the container to only reserve 128 MiB of memory\n\t\t\tfrom the remaining resources on the container instance, but also allow the container to\n\t\t\tconsume more memory resources when needed.
The Docker 20.10.0 or later daemon reserves a minimum of 6 MiB of memory for a\n\t\t\tcontainer. So, don't specify less than 6 MiB of memory for your containers.
\nThe Docker 19.03.13-ce or earlier daemon reserves a minimum of 4 MiB of memory for a\n\t\t\tcontainer. So, don't specify less than 4 MiB of memory for your containers.
" } }, "links": { "target": "com.amazonaws.ecs#StringList", "traits": { - "smithy.api#documentation": "The links
parameter allows containers to communicate with each other\n\t\t\twithout the need for port mappings. This parameter is only supported if the network mode\n\t\t\tof a task definition is bridge
. The name:internalName
\n\t\t\tconstruct is analogous to name:alias
in Docker links.\n\t\t\tUp to 255 letters (uppercase and lowercase), numbers, underscores, and hyphens are allowed.. This parameter maps to Links
in the docker create-container command and the\n\t\t\t\t--link
option to docker\n\t\t\trun.
This parameter is not supported for Windows containers.
\nContainers that are collocated on a single container instance may be able to\n\t\t\t\tcommunicate with each other without requiring links or host port mappings. Network\n\t\t\t\tisolation is achieved on the container instance using security groups and VPC\n\t\t\t\tsettings.
\nThe links
parameter allows containers to communicate with each other\n\t\t\twithout the need for port mappings. This parameter is only supported if the network mode\n\t\t\tof a task definition is bridge
. The name:internalName
\n\t\t\tconstruct is analogous to name:alias
in Docker links.\n\t\t\tUp to 255 letters (uppercase and lowercase), numbers, underscores, and hyphens are allowed.. This parameter maps to Links
in the docker conainer create command and the\n\t\t\t\t--link
option to docker\n\t\t\trun.
This parameter is not supported for Windows containers.
\nContainers that are collocated on a single container instance may be able to\n\t\t\t\tcommunicate with each other without requiring links or host port mappings. Network\n\t\t\t\tisolation is achieved on the container instance using security groups and VPC\n\t\t\t\tsettings.
\nThe list of port mappings for the container. Port mappings allow containers to access\n\t\t\tports on the host container instance to send or receive traffic.
\nFor task definitions that use the awsvpc
network mode, only specify the\n\t\t\t\tcontainerPort
. The hostPort
can be left blank or it must\n\t\t\tbe the same value as the containerPort
.
Port mappings on Windows use the NetNAT
gateway address rather than\n\t\t\t\tlocalhost
. There's no loopback for port mappings on Windows, so you\n\t\t\tcan't access a container's mapped port from the host itself.
This parameter maps to PortBindings
in the\n\t\t\tthe docker create-container command and the\n\t\t\t\t--publish
option to docker\n\t\t\t\trun. If the network mode of a task definition is set to none
,\n\t\t\tthen you can't specify port mappings. If the network mode of a task definition is set to\n\t\t\t\thost
, then host ports must either be undefined or they must match the\n\t\t\tcontainer port in the port mapping.
After a task reaches the RUNNING
status, manual and automatic host\n\t\t\t\tand container port assignments are visible in the Network\n\t\t\t\t\tBindings section of a container description for a selected task in\n\t\t\t\tthe Amazon ECS console. The assignments are also visible in the\n\t\t\t\t\tnetworkBindings
section DescribeTasks\n\t\t\t\tresponses.
The list of port mappings for the container. Port mappings allow containers to access\n\t\t\tports on the host container instance to send or receive traffic.
\nFor task definitions that use the awsvpc
network mode, only specify the\n\t\t\t\tcontainerPort
. The hostPort
can be left blank or it must\n\t\t\tbe the same value as the containerPort
.
Port mappings on Windows use the NetNAT
gateway address rather than\n\t\t\t\tlocalhost
. There's no loopback for port mappings on Windows, so you\n\t\t\tcan't access a container's mapped port from the host itself.
This parameter maps to PortBindings
in the\n\t\t\tthe docker conainer create command and the\n\t\t\t\t--publish
option to docker\n\t\t\t\trun. If the network mode of a task definition is set to none
,\n\t\t\tthen you can't specify port mappings. If the network mode of a task definition is set to\n\t\t\t\thost
, then host ports must either be undefined or they must match the\n\t\t\tcontainer port in the port mapping.
After a task reaches the RUNNING
status, manual and automatic host\n\t\t\t\tand container port assignments are visible in the Network\n\t\t\t\t\tBindings section of a container description for a selected task in\n\t\t\t\tthe Amazon ECS console. The assignments are also visible in the\n\t\t\t\tnetworkBindings
section DescribeTasks\n\t\t\t\tresponses.
Early versions of the Amazon ECS container agent don't properly handle\n\t\t\t\t\tentryPoint
parameters. If you have problems using\n\t\t\t\t\tentryPoint
, update your container agent or enter your commands and\n\t\t\t\targuments as command
array items instead.
The entry point that's passed to the container. This parameter maps to\n\t\t\tEntrypoint
in tthe docker create-container command and the --entrypoint
option to docker run.
Early versions of the Amazon ECS container agent don't properly handle\n\t\t\t\t\tentryPoint
parameters. If you have problems using\n\t\t\t\t\tentryPoint
, update your container agent or enter your commands and\n\t\t\t\targuments as command
array items instead.
The entry point that's passed to the container. This parameter maps to\n\t\t\tEntrypoint
in tthe docker conainer create command and the --entrypoint
option to docker run.
The command that's passed to the container. This parameter maps to Cmd
in\n\t\t\tthe docker create-container command and the\n\t\t\t\tCOMMAND
parameter to docker\n\t\t\t\trun. If there are multiple arguments, each\n\t\t\targument is a separated string in the array.
The command that's passed to the container. This parameter maps to Cmd
in\n\t\t\tthe docker conainer create command and the\n\t\t\t\tCOMMAND
parameter to docker\n\t\t\t\trun. If there are multiple arguments, each\n\t\t\targument is a separated string in the array.
The environment variables to pass to a container. This parameter maps to\n\t\t\tEnv
in the docker create-container command and the --env
option to docker run.
We don't recommend that you use plaintext environment variables for sensitive\n\t\t\t\tinformation, such as credential data.
\nThe environment variables to pass to a container. This parameter maps to\n\t\t\tEnv
in the docker conainer create command and the --env
option to docker run.
We don't recommend that you use plaintext environment variables for sensitive\n\t\t\t\tinformation, such as credential data.
\nThe mount points for data volumes in your container.
\nThis parameter maps to Volumes
in the the docker create-container command and the --volume
option to docker run.
Windows containers can mount whole directories on the same drive as\n\t\t\t\t$env:ProgramData
. Windows containers can't mount directories on a\n\t\t\tdifferent drive, and mount point can't be across drives.
The mount points for data volumes in your container.
\nThis parameter maps to Volumes
in the the docker conainer create command and the --volume
option to docker run.
Windows containers can mount whole directories on the same drive as\n\t\t\t\t$env:ProgramData
. Windows containers can't mount directories on a\n\t\t\tdifferent drive, and mount point can't be across drives.
Data volumes to mount from another container. This parameter maps to\n\t\t\tVolumesFrom
in tthe docker create-container command and the --volumes-from
option to docker run.
Data volumes to mount from another container. This parameter maps to\n\t\t\tVolumesFrom
in tthe docker conainer create command and the --volumes-from
option to docker run.
Linux-specific modifications that are applied to the container, such as Linux kernel\n\t\t\tcapabilities. For more information see KernelCapabilities.
\nThis parameter is not supported for Windows containers.
\nLinux-specific modifications that are applied to the container, such as Linux kernel\n\t\t\tcapabilities. For more information see KernelCapabilities.
\nThis parameter is not supported for Windows containers.
\nThe hostname to use for your container. This parameter maps to Hostname
\n\t\t\tin thethe docker create-container command and the\n\t\t\t\t--hostname
option to docker\n\t\t\t\trun.
The hostname
parameter is not supported if you're using the\n\t\t\t\t\tawsvpc
network mode.
The hostname to use for your container. This parameter maps to Hostname
\n\t\t\tin thethe docker conainer create command and the\n\t\t\t\t--hostname
option to docker\n\t\t\t\trun.
The hostname
parameter is not supported if you're using the\n\t\t\t\t\tawsvpc
network mode.
The user to use inside the container. This parameter maps to User
in the docker create-container command and the\n\t\t\t\t--user
option to docker\n\t\t\trun.
When running tasks using the host
network mode, don't run containers\n\t\t\t\tusing the root user (UID 0). We recommend using a non-root user for better\n\t\t\t\tsecurity.
You can specify the user
using the following formats. If specifying a UID\n\t\t\tor GID, you must specify it as a positive integer.
\n user
\n
\n user:group
\n
\n uid
\n
\n uid:gid
\n
\n user:gid
\n
\n uid:group
\n
This parameter is not supported for Windows containers.
\nThe user to use inside the container. This parameter maps to User
in the docker conainer create command and the\n\t\t\t\t--user
option to docker\n\t\t\trun.
When running tasks using the host
network mode, don't run containers\n\t\t\t\tusing the root user (UID 0). We recommend using a non-root user for better\n\t\t\t\tsecurity.
You can specify the user
using the following formats. If specifying a UID\n\t\t\tor GID, you must specify it as a positive integer.
\n user
\n
\n user:group
\n
\n uid
\n
\n uid:gid
\n
\n user:gid
\n
\n uid:group
\n
This parameter is not supported for Windows containers.
\nThe working directory to run commands inside the container in. This parameter maps to\n\t\t\tWorkingDir
in the docker create-container command and the --workdir
option to docker run.
The working directory to run commands inside the container in. This parameter maps to\n\t\t\tWorkingDir
in the docker conainer create command and the --workdir
option to docker run.
When this parameter is true, networking is off within the container. This parameter\n\t\t\tmaps to NetworkDisabled
in the docker create-container command.
This parameter is not supported for Windows containers.
\nWhen this parameter is true, networking is off within the container. This parameter\n\t\t\tmaps to NetworkDisabled
in the docker conainer create command.
This parameter is not supported for Windows containers.
\nWhen this parameter is true, the container is given elevated privileges on the host\n\t\t\tcontainer instance (similar to the root
user). This parameter maps to\n\t\t\tPrivileged
in the the docker create-container command and the --privileged
option to docker run
This parameter is not supported for Windows containers or tasks run on Fargate.
\nWhen this parameter is true, the container is given elevated privileges on the host\n\t\t\tcontainer instance (similar to the root
user). This parameter maps to\n\t\t\tPrivileged
in the the docker conainer create command and the --privileged
option to docker run
This parameter is not supported for Windows containers or tasks run on Fargate.
\nWhen this parameter is true, the container is given read-only access to its root file\n\t\t\tsystem. This parameter maps to ReadonlyRootfs
in the docker create-container command and the\n\t\t\t\t--read-only
option to docker\n\t\t\t\trun.
This parameter is not supported for Windows containers.
\nWhen this parameter is true, the container is given read-only access to its root file\n\t\t\tsystem. This parameter maps to ReadonlyRootfs
in the docker conainer create command and the\n\t\t\t\t--read-only
option to docker\n\t\t\t\trun.
This parameter is not supported for Windows containers.
\nA list of DNS servers that are presented to the container. This parameter maps to\n\t\t\tDns
in the the docker create-container command and the --dns
option to docker run.
This parameter is not supported for Windows containers.
\nA list of DNS servers that are presented to the container. This parameter maps to\n\t\t\tDns
in the the docker conainer create command and the --dns
option to docker run.
This parameter is not supported for Windows containers.
\nA list of DNS search domains that are presented to the container. This parameter maps\n\t\t\tto DnsSearch
in the docker create-container command and the --dns-search
option to docker run.
This parameter is not supported for Windows containers.
\nA list of DNS search domains that are presented to the container. This parameter maps\n\t\t\tto DnsSearch
in the docker conainer create command and the --dns-search
option to docker run.
This parameter is not supported for Windows containers.
\nA list of hostnames and IP address mappings to append to the /etc/hosts
\n\t\t\tfile on the container. This parameter maps to ExtraHosts
in the docker create-container command and the\n\t\t\t\t--add-host
option to docker\n\t\t\t\trun.
This parameter isn't supported for Windows containers or tasks that use the\n\t\t\t\t\tawsvpc
network mode.
A list of hostnames and IP address mappings to append to the /etc/hosts
\n\t\t\tfile on the container. This parameter maps to ExtraHosts
in the docker conainer create command and the\n\t\t\t\t--add-host
option to docker\n\t\t\t\trun.
This parameter isn't supported for Windows containers or tasks that use the\n\t\t\t\t\tawsvpc
network mode.
A list of strings to provide custom configuration for multiple security systems. This field isn't valid for containers in tasks\n\t\t\tusing the Fargate launch type.
\nFor Linux tasks on EC2, this parameter can be used to reference custom\n\t\t\tlabels for SELinux and AppArmor multi-level security systems.
\nFor any tasks on EC2, this parameter can be used to reference a\n\t\t\tcredential spec file that configures a container for Active Directory authentication.\n\t\t\tFor more information, see Using gMSAs for Windows\n\t\t\t\tContainers and Using gMSAs for Linux\n\t\t\t\tContainers in the Amazon Elastic Container Service Developer Guide.
\nThis parameter maps to SecurityOpt
in the docker create-container command and the\n\t\t\t\t--security-opt
option to docker\n\t\t\t\trun.
The Amazon ECS container agent running on a container instance must register with the\n\t\t\t\t\tECS_SELINUX_CAPABLE=true
or ECS_APPARMOR_CAPABLE=true
\n\t\t\t\tenvironment variables before containers placed on that instance can use these\n\t\t\t\tsecurity options. For more information, see Amazon ECS Container\n\t\t\t\t\tAgent Configuration in the Amazon Elastic Container Service Developer Guide.
Valid values: \"no-new-privileges\" | \"apparmor:PROFILE\" | \"label:value\" |\n\t\t\t\"credentialspec:CredentialSpecFilePath\"
" + "smithy.api#documentation": "A list of strings to provide custom configuration for multiple security systems. This field isn't valid for containers in tasks\n\t\t\tusing the Fargate launch type.
\nFor Linux tasks on EC2, this parameter can be used to reference custom\n\t\t\tlabels for SELinux and AppArmor multi-level security systems.
\nFor any tasks on EC2, this parameter can be used to reference a\n\t\t\tcredential spec file that configures a container for Active Directory authentication.\n\t\t\tFor more information, see Using gMSAs for Windows\n\t\t\t\tContainers and Using gMSAs for Linux\n\t\t\t\tContainers in the Amazon Elastic Container Service Developer Guide.
\nThis parameter maps to SecurityOpt
in the docker conainer create command and the\n\t\t\t\t--security-opt
option to docker\n\t\t\t\trun.
The Amazon ECS container agent running on a container instance must register with the\n\t\t\t\t\tECS_SELINUX_CAPABLE=true
or ECS_APPARMOR_CAPABLE=true
\n\t\t\t\tenvironment variables before containers placed on that instance can use these\n\t\t\t\tsecurity options. For more information, see Amazon ECS Container\n\t\t\t\t\tAgent Configuration in the Amazon Elastic Container Service Developer Guide.
Valid values: \"no-new-privileges\" | \"apparmor:PROFILE\" | \"label:value\" |\n\t\t\t\"credentialspec:CredentialSpecFilePath\"
" } }, "interactive": { "target": "com.amazonaws.ecs#BoxedBoolean", "traits": { - "smithy.api#documentation": "When this parameter is true
, you can deploy containerized applications\n\t\t\tthat require stdin
or a tty
to be allocated. This parameter\n\t\t\tmaps to OpenStdin
in the docker create-container command and the --interactive
option to docker run.
When this parameter is true
, you can deploy containerized applications\n\t\t\tthat require stdin
or a tty
to be allocated. This parameter\n\t\t\tmaps to OpenStdin
in the docker conainer create command and the --interactive
option to docker run.
When this parameter is true
, a TTY is allocated. This parameter maps to\n\t\t\tTty
in tthe docker create-container command and the --tty
option to docker run.
When this parameter is true
, a TTY is allocated. This parameter maps to\n\t\t\tTty
in tthe docker conainer create command and the --tty
option to docker run.
A key/value map of labels to add to the container. This parameter maps to\n\t\t\tLabels
in the docker create-container command and the --label
option to docker run. This parameter requires version 1.18 of the Docker Remote API or greater on your container instance. To check the Docker Remote API version on your container instance, log in to your container instance and run the following command: sudo docker version --format '{{.Server.APIVersion}}'
\n
A key/value map of labels to add to the container. This parameter maps to\n\t\t\tLabels
in the docker conainer create command and the --label
option to docker run. This parameter requires version 1.18 of the Docker Remote API or greater on your container instance. To check the Docker Remote API version on your container instance, log in to your container instance and run the following command: sudo docker version --format '{{.Server.APIVersion}}'
\n
A list of ulimits
to set in the container. If a ulimit
value\n\t\t\tis specified in a task definition, it overrides the default values set by Docker. This\n\t\t\tparameter maps to Ulimits
in tthe docker create-container command and the --ulimit
option to docker run. Valid naming values are displayed\n\t\t\tin the Ulimit data type.
Amazon ECS tasks hosted on Fargate use the default\n\t\t\t\t\t\t\tresource limit values set by the operating system with the exception of\n\t\t\t\t\t\t\tthe nofile
resource limit parameter which Fargate\n\t\t\t\t\t\t\toverrides. The nofile
resource limit sets a restriction on\n\t\t\t\t\t\t\tthe number of open files that a container can use. The default\n\t\t\t\t\t\t\t\tnofile
soft limit is 65535
and the default hard limit\n\t\t\t\t\t\t\tis 65535
.
This parameter requires version 1.18 of the Docker Remote API or greater on your container instance. To check the Docker Remote API version on your container instance, log in to your container instance and run the following command: sudo docker version --format '{{.Server.APIVersion}}'
\n
This parameter is not supported for Windows containers.
\nA list of ulimits
to set in the container. If a ulimit
value\n\t\t\tis specified in a task definition, it overrides the default values set by Docker. This\n\t\t\tparameter maps to Ulimits
in tthe docker conainer create command and the --ulimit
option to docker run. Valid naming values are displayed\n\t\t\tin the Ulimit data type.
Amazon ECS tasks hosted on Fargate use the default\n\t\t\t\t\t\t\tresource limit values set by the operating system with the exception of\n\t\t\t\t\t\t\tthe nofile
resource limit parameter which Fargate\n\t\t\t\t\t\t\toverrides. The nofile
resource limit sets a restriction on\n\t\t\t\t\t\t\tthe number of open files that a container can use. The default\n\t\t\t\t\t\t\t\tnofile
soft limit is 65535
and the default hard limit\n\t\t\t\t\t\t\tis 65535
.
This parameter requires version 1.18 of the Docker Remote API or greater on your container instance. To check the Docker Remote API version on your container instance, log in to your container instance and run the following command: sudo docker version --format '{{.Server.APIVersion}}'
\n
This parameter is not supported for Windows containers.
\nThe log configuration specification for the container.
\nThis parameter maps to LogConfig
in the docker create-container command and the\n\t\t\t\t--log-driver
option to docker\n\t\t\t\trun. By default, containers use the same logging driver that the Docker\n\t\t\tdaemon uses. However the container can use a different logging driver than the Docker\n\t\t\tdaemon by specifying a log driver with this parameter in the container definition. To\n\t\t\tuse a different logging driver for a container, the log system must be configured\n\t\t\tproperly on the container instance (or on a different log server for remote logging\n\t\t\toptions).
Amazon ECS currently supports a subset of the logging drivers available to the Docker\n\t\t\t\tdaemon (shown in the LogConfiguration data type). Additional log\n\t\t\t\tdrivers may be available in future releases of the Amazon ECS container agent.
\nThis parameter requires version 1.18 of the Docker Remote API or greater on your container instance. To check the Docker Remote API version on your container instance, log in to your container instance and run the following command: sudo docker version --format '{{.Server.APIVersion}}'
\n
The Amazon ECS container agent running on a container instance must register the\n\t\t\t\tlogging drivers available on that instance with the\n\t\t\t\t\tECS_AVAILABLE_LOGGING_DRIVERS
environment variable before\n\t\t\t\tcontainers placed on that instance can use these log configuration options. For more\n\t\t\t\tinformation, see Amazon ECS Container\n\t\t\t\t\tAgent Configuration in the Amazon Elastic Container Service Developer Guide.
The log configuration specification for the container.
\nThis parameter maps to LogConfig
in the docker conainer create command and the\n\t\t\t\t--log-driver
option to docker\n\t\t\t\trun. By default, containers use the same logging driver that the Docker\n\t\t\tdaemon uses. However the container can use a different logging driver than the Docker\n\t\t\tdaemon by specifying a log driver with this parameter in the container definition. To\n\t\t\tuse a different logging driver for a container, the log system must be configured\n\t\t\tproperly on the container instance (or on a different log server for remote logging\n\t\t\toptions).
Amazon ECS currently supports a subset of the logging drivers available to the Docker\n\t\t\t\tdaemon (shown in the LogConfiguration data type). Additional log\n\t\t\t\tdrivers may be available in future releases of the Amazon ECS container agent.
\nThis parameter requires version 1.18 of the Docker Remote API or greater on your container instance. To check the Docker Remote API version on your container instance, log in to your container instance and run the following command: sudo docker version --format '{{.Server.APIVersion}}'
\n
The Amazon ECS container agent running on a container instance must register the\n\t\t\t\tlogging drivers available on that instance with the\n\t\t\t\t\tECS_AVAILABLE_LOGGING_DRIVERS
environment variable before\n\t\t\t\tcontainers placed on that instance can use these log configuration options. For more\n\t\t\t\tinformation, see Amazon ECS Container\n\t\t\t\t\tAgent Configuration in the Amazon Elastic Container Service Developer Guide.
The container health check command and associated configuration parameters for the\n\t\t\tcontainer. This parameter maps to HealthCheck
in the docker create-container command and the\n\t\t\t\tHEALTHCHECK
parameter of docker\n\t\t\t\trun.
The container health check command and associated configuration parameters for the\n\t\t\tcontainer. This parameter maps to HealthCheck
in the docker conainer create command and the\n\t\t\t\tHEALTHCHECK
parameter of docker\n\t\t\t\trun.
A list of namespaced kernel parameters to set in the container. This parameter maps to\n\t\t\tSysctls
in tthe docker create-container command and the --sysctl
option to docker run. For example, you can configure\n\t\t\t\tnet.ipv4.tcp_keepalive_time
setting to maintain longer lived\n\t\t\tconnections.
A list of namespaced kernel parameters to set in the container. This parameter maps to\n\t\t\tSysctls
in tthe docker conainer create command and the --sysctl
option to docker run. For example, you can configure\n\t\t\t\tnet.ipv4.tcp_keepalive_time
setting to maintain longer lived\n\t\t\tconnections.
The attributes set for the container instance, either by the Amazon ECS container agent at\n\t\t\tinstance registration or manually with the PutAttributes\n\t\t\toperation.
" + "smithy.api#documentation": "The attributes set for the container instance, either by the Amazon ECS container agent at\n\t\t\tinstance registration or manually with the PutAttributes\n\t\t\toperation.
" } }, "registeredAt": { @@ -3011,7 +3011,7 @@ } ], "traits": { - "smithy.api#documentation": "Creates a new Amazon ECS cluster. By default, your account receives a default
\n\t\t\tcluster when you launch your first container instance. However, you can create your own\n\t\t\tcluster with a unique name with the CreateCluster
action.
When you call the CreateCluster API operation, Amazon ECS attempts to\n\t\t\t\tcreate the Amazon ECS service-linked role for your account. This is so that it can manage\n\t\t\t\trequired resources in other Amazon Web Services services on your behalf. However, if the user that\n\t\t\t\tmakes the call doesn't have permissions to create the service-linked role, it isn't\n\t\t\t\tcreated. For more information, see Using\n\t\t\t\t\tservice-linked roles for Amazon ECS in the Amazon Elastic Container Service Developer Guide.
\nCreates a new Amazon ECS cluster. By default, your account receives a default
\n\t\t\tcluster when you launch your first container instance. However, you can create your own\n\t\t\tcluster with a unique name.
When you call the CreateCluster API operation, Amazon ECS attempts to\n\t\t\t\tcreate the Amazon ECS service-linked role for your account. This is so that it can manage\n\t\t\t\trequired resources in other Amazon Web Services services on your behalf. However, if the user that\n\t\t\t\tmakes the call doesn't have permissions to create the service-linked role, it isn't\n\t\t\t\tcreated. For more information, see Using\n\t\t\t\t\tservice-linked roles for Amazon ECS in the Amazon Elastic Container Service Developer Guide.
\nThe setting to use when creating a cluster. This parameter is used to turn on CloudWatch\n\t\t\tContainer Insights for a cluster. If this value is specified, it overrides the\n\t\t\t\tcontainerInsights
value set with PutAccountSetting or\n\t\t\t\tPutAccountSettingDefault.
The setting to use when creating a cluster. This parameter is used to turn on CloudWatch\n\t\t\tContainer Insights for a cluster. If this value is specified, it overrides the\n\t\t\tcontainerInsights
value set with PutAccountSetting or\n\t\t\tPutAccountSettingDefault.
The capacity provider strategy to set as the default for the cluster. After a default\n\t\t\tcapacity provider strategy is set for a cluster, when you call the CreateService or RunTask APIs with no\n\t\t\tcapacity provider strategy or launch type specified, the default capacity provider\n\t\t\tstrategy for the cluster is used.
\nIf a default capacity provider strategy isn't defined for a cluster when it was\n\t\t\tcreated, it can be defined later with the PutClusterCapacityProviders\n\t\t\tAPI operation.
" + "smithy.api#documentation": "The capacity provider strategy to set as the default for the cluster. After a default\n\t\t\tcapacity provider strategy is set for a cluster, when you call the CreateService or RunTask APIs with no\n\t\t\tcapacity provider strategy or launch type specified, the default capacity provider\n\t\t\tstrategy for the cluster is used.
\nIf a default capacity provider strategy isn't defined for a cluster when it was\n\t\t\tcreated, it can be defined later with the PutClusterCapacityProviders\n\t\t\tAPI operation.
" } }, "serviceConnectDefaults": { @@ -3136,7 +3136,7 @@ } ], "traits": { - "smithy.api#documentation": "Runs and maintains your desired number of tasks from a specified task definition. If\n\t\t\tthe number of tasks running in a service drops below the desiredCount
,\n\t\t\tAmazon ECS runs another copy of the task in the specified cluster. To update an existing\n\t\t\tservice, see the UpdateService action.
On March 21, 2024, a change was made to resolve the task definition revision before authorization. When a task definition revision is not specified, authorization will occur using the latest revision of a task definition.
\nIn addition to maintaining the desired count of tasks in your service, you can\n\t\t\toptionally run your service behind one or more load balancers. The load balancers\n\t\t\tdistribute traffic across the tasks that are associated with the service. For more\n\t\t\tinformation, see Service load balancing in the Amazon Elastic Container Service Developer Guide.
\nYou can attach Amazon EBS volumes to Amazon ECS tasks by configuring the volume when creating or\n\t\t\tupdating a service. volumeConfigurations
is only supported for REPLICA\n\t\t\tservice and not DAEMON service. For more infomation, see Amazon EBS volumes in the Amazon Elastic Container Service Developer Guide.
Tasks for services that don't use a load balancer are considered healthy if they're in\n\t\t\tthe RUNNING
state. Tasks for services that use a load balancer are\n\t\t\tconsidered healthy if they're in the RUNNING
state and are reported as\n\t\t\thealthy by the load balancer.
There are two service scheduler strategies available:
\n\n REPLICA
- The replica scheduling strategy places and\n\t\t\t\t\tmaintains your desired number of tasks across your cluster. By default, the\n\t\t\t\t\tservice scheduler spreads tasks across Availability Zones. You can use task\n\t\t\t\t\tplacement strategies and constraints to customize task placement decisions. For\n\t\t\t\t\tmore information, see Service scheduler concepts in the Amazon Elastic Container Service Developer Guide.
\n DAEMON
- The daemon scheduling strategy deploys exactly one\n\t\t\t\t\ttask on each active container instance that meets all of the task placement\n\t\t\t\t\tconstraints that you specify in your cluster. The service scheduler also\n\t\t\t\t\tevaluates the task placement constraints for running tasks. It also stops tasks\n\t\t\t\t\tthat don't meet the placement constraints. When using this strategy, you don't\n\t\t\t\t\tneed to specify a desired number of tasks, a task placement strategy, or use\n\t\t\t\t\tService Auto Scaling policies. For more information, see Service scheduler concepts in the Amazon Elastic Container Service Developer Guide.
You can optionally specify a deployment configuration for your service. The deployment\n\t\t\tis initiated by changing properties. For example, the deployment might be initiated by\n\t\t\tthe task definition or by your desired count of a service. This is done with an UpdateService operation. The default value for a replica service for\n\t\t\t\tminimumHealthyPercent
is 100%. The default value for a daemon service\n\t\t\tfor minimumHealthyPercent
is 0%.
If a service uses the ECS
deployment controller, the minimum healthy\n\t\t\tpercent represents a lower limit on the number of tasks in a service that must remain in\n\t\t\tthe RUNNING
state during a deployment. Specifically, it represents it as a\n\t\t\tpercentage of your desired number of tasks (rounded up to the nearest integer). This\n\t\t\thappens when any of your container instances are in the DRAINING
state if\n\t\t\tthe service contains tasks using the EC2 launch type. Using this\n\t\t\tparameter, you can deploy without using additional cluster capacity. For example, if you\n\t\t\tset your service to have desired number of four tasks and a minimum healthy percent of\n\t\t\t50%, the scheduler might stop two existing tasks to free up cluster capacity before\n\t\t\tstarting two new tasks. If they're in the RUNNING
state, tasks for services\n\t\t\tthat don't use a load balancer are considered healthy . If they're in the\n\t\t\t\tRUNNING
state and reported as healthy by the load balancer, tasks for\n\t\t\tservices that do use a load balancer are considered healthy . The\n\t\t\tdefault value for minimum healthy percent is 100%.
If a service uses the ECS
deployment controller, the maximum percent parameter represents an upper limit on the\n\t\t\tnumber of tasks in a service that are allowed in the RUNNING
or\n\t\t\t\tPENDING
state during a deployment. Specifically, it represents it as a\n\t\t\tpercentage of the desired number of tasks (rounded down to the nearest integer). This\n\t\t\thappens when any of your container instances are in the DRAINING
state if\n\t\t\tthe service contains tasks using the EC2 launch type. Using this\n\t\t\tparameter, you can define the deployment batch size. For example, if your service has a\n\t\t\tdesired number of four tasks and a maximum percent value of 200%, the scheduler may\n\t\t\tstart four new tasks before stopping the four older tasks (provided that the cluster\n\t\t\tresources required to do this are available). The default value for maximum percent is\n\t\t\t200%.
If a service uses either the CODE_DEPLOY
or EXTERNAL
\n\t\t\tdeployment controller types and tasks that use the EC2 launch type, the\n\t\t\t\tminimum healthy percent and maximum percent values are used only to define the lower and upper limit\n\t\t\ton the number of the tasks in the service that remain in the RUNNING
state.\n\t\t\tThis is while the container instances are in the DRAINING
state. If the\n\t\t\ttasks in the service use the Fargate launch type, the minimum healthy\n\t\t\tpercent and maximum percent values aren't used. This is the case even if they're\n\t\t\tcurrently visible when describing your service.
When creating a service that uses the EXTERNAL
deployment controller, you\n\t\t\tcan specify only parameters that aren't controlled at the task set level. The only\n\t\t\trequired parameter is the service name. You control your services using the CreateTaskSet operation. For more information, see Amazon ECS deployment types in the Amazon Elastic Container Service Developer Guide.
When the service scheduler launches new tasks, it determines task placement. For\n\t\t\tinformation about task placement and task placement strategies, see Amazon ECS\n\t\t\t\ttask placement in the Amazon Elastic Container Service Developer Guide\n
\nStarting April 15, 2023, Amazon Web Services will not onboard new customers to Amazon Elastic Inference (EI), and will help current customers migrate their workloads to options that offer better price and performance. After April 15, 2023, new customers will not be able to launch instances with Amazon EI accelerators in Amazon SageMaker, Amazon ECS, or Amazon EC2. However, customers who have used Amazon EI at least once during the past 30-day period are considered current customers and will be able to continue using the service.
", + "smithy.api#documentation": "Runs and maintains your desired number of tasks from a specified task definition. If\n\t\t\tthe number of tasks running in a service drops below the desiredCount
,\n\t\t\tAmazon ECS runs another copy of the task in the specified cluster. To update an existing\n\t\t\tservice, use UpdateService.
On March 21, 2024, a change was made to resolve the task definition revision before authorization. When a task definition revision is not specified, authorization will occur using the latest revision of a task definition.
\nIn addition to maintaining the desired count of tasks in your service, you can\n\t\t\toptionally run your service behind one or more load balancers. The load balancers\n\t\t\tdistribute traffic across the tasks that are associated with the service. For more\n\t\t\tinformation, see Service load balancing in the Amazon Elastic Container Service Developer Guide.
\nYou can attach Amazon EBS volumes to Amazon ECS tasks by configuring the volume when creating or\n\t\t\tupdating a service. volumeConfigurations
is only supported for REPLICA\n\t\t\tservice and not DAEMON service. For more infomation, see Amazon EBS volumes in the Amazon Elastic Container Service Developer Guide.
Tasks for services that don't use a load balancer are considered healthy if they're in\n\t\t\tthe RUNNING
state. Tasks for services that use a load balancer are\n\t\t\tconsidered healthy if they're in the RUNNING
state and are reported as\n\t\t\thealthy by the load balancer.
There are two service scheduler strategies available:
\n\n REPLICA
- The replica scheduling strategy places and\n\t\t\t\t\tmaintains your desired number of tasks across your cluster. By default, the\n\t\t\t\t\tservice scheduler spreads tasks across Availability Zones. You can use task\n\t\t\t\t\tplacement strategies and constraints to customize task placement decisions. For\n\t\t\t\t\tmore information, see Service scheduler concepts in the Amazon Elastic Container Service Developer Guide.
\n DAEMON
- The daemon scheduling strategy deploys exactly one\n\t\t\t\t\ttask on each active container instance that meets all of the task placement\n\t\t\t\t\tconstraints that you specify in your cluster. The service scheduler also\n\t\t\t\t\tevaluates the task placement constraints for running tasks. It also stops tasks\n\t\t\t\t\tthat don't meet the placement constraints. When using this strategy, you don't\n\t\t\t\t\tneed to specify a desired number of tasks, a task placement strategy, or use\n\t\t\t\t\tService Auto Scaling policies. For more information, see Service scheduler concepts in the Amazon Elastic Container Service Developer Guide.
You can optionally specify a deployment configuration for your service. The deployment\n\t\t\tis initiated by changing properties. For example, the deployment might be initiated by\n\t\t\tthe task definition or by your desired count of a service. You can use UpdateService. The default value for a replica service for\n\t\t\t\tminimumHealthyPercent
is 100%. The default value for a daemon service\n\t\t\tfor minimumHealthyPercent
is 0%.
If a service uses the ECS
deployment controller, the minimum healthy\n\t\t\tpercent represents a lower limit on the number of tasks in a service that must remain in\n\t\t\tthe RUNNING
state during a deployment. Specifically, it represents it as a\n\t\t\tpercentage of your desired number of tasks (rounded up to the nearest integer). This\n\t\t\thappens when any of your container instances are in the DRAINING
state if\n\t\t\tthe service contains tasks using the EC2 launch type. Using this\n\t\t\tparameter, you can deploy without using additional cluster capacity. For example, if you\n\t\t\tset your service to have desired number of four tasks and a minimum healthy percent of\n\t\t\t50%, the scheduler might stop two existing tasks to free up cluster capacity before\n\t\t\tstarting two new tasks. If they're in the RUNNING
state, tasks for services\n\t\t\tthat don't use a load balancer are considered healthy . If they're in the\n\t\t\t\tRUNNING
state and reported as healthy by the load balancer, tasks for\n\t\t\tservices that do use a load balancer are considered healthy . The\n\t\t\tdefault value for minimum healthy percent is 100%.
If a service uses the ECS
deployment controller, the maximum percent parameter represents an upper limit on the\n\t\t\tnumber of tasks in a service that are allowed in the RUNNING
or\n\t\t\t\tPENDING
state during a deployment. Specifically, it represents it as a\n\t\t\tpercentage of the desired number of tasks (rounded down to the nearest integer). This\n\t\t\thappens when any of your container instances are in the DRAINING
state if\n\t\t\tthe service contains tasks using the EC2 launch type. Using this\n\t\t\tparameter, you can define the deployment batch size. For example, if your service has a\n\t\t\tdesired number of four tasks and a maximum percent value of 200%, the scheduler may\n\t\t\tstart four new tasks before stopping the four older tasks (provided that the cluster\n\t\t\tresources required to do this are available). The default value for maximum percent is\n\t\t\t200%.
If a service uses either the CODE_DEPLOY
or EXTERNAL
\n\t\t\tdeployment controller types and tasks that use the EC2 launch type, the\n\t\t\t\tminimum healthy percent and maximum percent values are used only to define the lower and upper limit\n\t\t\ton the number of the tasks in the service that remain in the RUNNING
state.\n\t\t\tThis is while the container instances are in the DRAINING
state. If the\n\t\t\ttasks in the service use the Fargate launch type, the minimum healthy\n\t\t\tpercent and maximum percent values aren't used. This is the case even if they're\n\t\t\tcurrently visible when describing your service.
When creating a service that uses the EXTERNAL
deployment controller, you\n\t\t\tcan specify only parameters that aren't controlled at the task set level. The only\n\t\t\trequired parameter is the service name. You control your services using the CreateTaskSet. For more information, see Amazon ECS deployment types in the Amazon Elastic Container Service Developer Guide.
When the service scheduler launches new tasks, it determines task placement. For\n\t\t\tinformation about task placement and task placement strategies, see Amazon ECS\n\t\t\t\ttask placement in the Amazon Elastic Container Service Developer Guide\n
\nStarting April 15, 2023, Amazon Web Services will not onboard new customers to Amazon Elastic Inference (EI), and will help current customers migrate their workloads to options that offer better price and performance. After April 15, 2023, new customers will not be able to launch instances with Amazon EI accelerators in Amazon SageMaker, Amazon ECS, or Amazon EC2. However, customers who have used Amazon EI at least once during the past 30-day period are considered current customers and will be able to continue using the service.
", "smithy.api#examples": [ { "title": "To create a new service", @@ -3519,7 +3519,7 @@ "capacityProviderStrategy": { "target": "com.amazonaws.ecs#CapacityProviderStrategy", "traits": { - "smithy.api#documentation": "The capacity provider strategy to use for the task set.
\nA capacity provider strategy consists of one or more capacity providers along with the\n\t\t\t\tbase
and weight
to assign to them. A capacity provider\n\t\t\tmust be associated with the cluster to be used in a capacity provider strategy. The\n\t\t\t\tPutClusterCapacityProviders API is used to associate a capacity\n\t\t\tprovider with a cluster. Only capacity providers with an ACTIVE
or\n\t\t\t\tUPDATING
status can be used.
If a capacityProviderStrategy
is specified, the launchType
\n\t\t\tparameter must be omitted. If no capacityProviderStrategy
or\n\t\t\t\tlaunchType
is specified, the\n\t\t\t\tdefaultCapacityProviderStrategy
for the cluster is used.
If specifying a capacity provider that uses an Auto Scaling group, the capacity\n\t\t\tprovider must already be created. New capacity providers can be created with the CreateCapacityProvider API operation.
\nTo use a Fargate capacity provider, specify either the FARGATE
or\n\t\t\t\tFARGATE_SPOT
capacity providers. The Fargate capacity providers are\n\t\t\tavailable to all accounts and only need to be associated with a cluster to be\n\t\t\tused.
The PutClusterCapacityProviders API operation is used to update the\n\t\t\tlist of available capacity providers for a cluster after the cluster is created.
" + "smithy.api#documentation": "The capacity provider strategy to use for the task set.
\nA capacity provider strategy consists of one or more capacity providers along with the\n\t\t\t\tbase
and weight
to assign to them. A capacity provider\n\t\t\tmust be associated with the cluster to be used in a capacity provider strategy. The PutClusterCapacityProviders\n\t\t\t\t API is used to associate a capacity\n\t\t\tprovider with a cluster. Only capacity providers with an ACTIVE
or\n\t\t\t\tUPDATING
status can be used.
If a capacityProviderStrategy
is specified, the launchType
\n\t\t\tparameter must be omitted. If no capacityProviderStrategy
or\n\t\t\t\tlaunchType
is specified, the\n\t\t\t\tdefaultCapacityProviderStrategy
for the cluster is used.
If specifying a capacity provider that uses an Auto Scaling group, the capacity\n\t\t\tprovider must already be created. New capacity providers can be created with the CreateCapacityProviderProviderAPI operation.
\nTo use a Fargate capacity provider, specify either the FARGATE
or\n\t\t\t\tFARGATE_SPOT
capacity providers. The Fargate capacity providers are\n\t\t\tavailable to all accounts and only need to be associated with a cluster to be\n\t\t\tused.
The PutClusterCapacityProviders API operation is used to update the\n\t\t\tlist of available capacity providers for a cluster after the cluster is created.
" } }, "platformVersion": { @@ -3732,7 +3732,7 @@ } ], "traits": { - "smithy.api#documentation": "Deletes the specified capacity provider.
\nThe FARGATE
and FARGATE_SPOT
capacity providers are\n\t\t\t\treserved and can't be deleted. You can disassociate them from a cluster using either\n\t\t\t\tthe PutClusterCapacityProviders API or by deleting the\n\t\t\t\tcluster.
Prior to a capacity provider being deleted, the capacity provider must be removed from\n\t\t\tthe capacity provider strategy from all services. The UpdateService\n\t\t\tAPI can be used to remove a capacity provider from a service's capacity provider\n\t\t\tstrategy. When updating a service, the forceNewDeployment
option can be\n\t\t\tused to ensure that any tasks using the Amazon EC2 instance capacity provided by the capacity\n\t\t\tprovider are transitioned to use the capacity from the remaining capacity providers.\n\t\t\tOnly capacity providers that aren't associated with a cluster can be deleted. To remove\n\t\t\ta capacity provider from a cluster, you can either use PutClusterCapacityProviders or delete the cluster.
Deletes the specified capacity provider.
\nThe FARGATE
and FARGATE_SPOT
capacity providers are\n\t\t\t\treserved and can't be deleted. You can disassociate them from a cluster using either PutCapacityProviderProviders or by deleting the\n\t\t\t\tcluster.
Prior to a capacity provider being deleted, the capacity provider must be removed from\n\t\t\tthe capacity provider strategy from all services. The UpdateService\n\t\t\tAPI can be used to remove a capacity provider from a service's capacity provider\n\t\t\tstrategy. When updating a service, the forceNewDeployment
option can be\n\t\t\tused to ensure that any tasks using the Amazon EC2 instance capacity provided by the capacity\n\t\t\tprovider are transitioned to use the capacity from the remaining capacity providers.\n\t\t\tOnly capacity providers that aren't associated with a cluster can be deleted. To remove\n\t\t\ta capacity provider from a cluster, you can either use PutCapacityProviderProviders or delete the cluster.
Deletes the specified cluster. The cluster transitions to the INACTIVE
\n\t\t\tstate. Clusters with an INACTIVE
status might remain discoverable in your\n\t\t\taccount for a period of time. However, this behavior is subject to change in the future.\n\t\t\tWe don't recommend that you rely on INACTIVE
clusters persisting.
You must deregister all container instances from this cluster before you may delete\n\t\t\tit. You can list the container instances in a cluster with ListContainerInstances and deregister them with DeregisterContainerInstance.
", + "smithy.api#documentation": "Deletes the specified cluster. The cluster transitions to the INACTIVE
\n\t\t\tstate. Clusters with an INACTIVE
status might remain discoverable in your\n\t\t\taccount for a period of time. However, this behavior is subject to change in the future.\n\t\t\tWe don't recommend that you rely on INACTIVE
clusters persisting.
You must deregister all container instances from this cluster before you may delete\n\t\t\tit. You can list the container instances in a cluster with ListContainerInstances and deregister them with DeregisterContainerInstance.
", "smithy.api#examples": [ { "title": "To delete an empty cluster", @@ -3877,7 +3877,7 @@ } ], "traits": { - "smithy.api#documentation": "Deletes a specified service within a cluster. You can delete a service if you have no\n\t\t\trunning tasks in it and the desired task count is zero. If the service is actively\n\t\t\tmaintaining tasks, you can't delete it, and you must update the service to a desired\n\t\t\ttask count of zero. For more information, see UpdateService.
\nWhen you delete a service, if there are still running tasks that require cleanup,\n\t\t\t\tthe service status moves from ACTIVE
to DRAINING
, and the\n\t\t\t\tservice is no longer visible in the console or in the ListServices\n\t\t\t\tAPI operation. After all tasks have transitioned to either STOPPING
or\n\t\t\t\t\tSTOPPED
status, the service status moves from DRAINING
\n\t\t\t\tto INACTIVE
. Services in the DRAINING
or\n\t\t\t\t\tINACTIVE
status can still be viewed with the DescribeServices API operation. However, in the future,\n\t\t\t\t\tINACTIVE
services may be cleaned up and purged from Amazon ECS record\n\t\t\t\tkeeping, and DescribeServices calls on those services return a\n\t\t\t\t\tServiceNotFoundException
error.
If you attempt to create a new service with the same name as an existing service\n\t\t\t\tin either ACTIVE
or DRAINING
status, you receive an\n\t\t\t\terror.
Deletes a specified service within a cluster. You can delete a service if you have no\n\t\t\trunning tasks in it and the desired task count is zero. If the service is actively\n\t\t\tmaintaining tasks, you can't delete it, and you must update the service to a desired\n\t\t\ttask count of zero. For more information, see UpdateService.
\nWhen you delete a service, if there are still running tasks that require cleanup,\n\t\t\t\tthe service status moves from ACTIVE
to DRAINING
, and the\n\t\t\t\tservice is no longer visible in the console or in the ListServices\n\t\t\t\tAPI operation. After all tasks have transitioned to either STOPPING
or\n\t\t\t\t\tSTOPPED
status, the service status moves from DRAINING
\n\t\t\t\tto INACTIVE
. Services in the DRAINING
or\n\t\t\t\tINACTIVE
status can still be viewed with the DescribeServices API operation. However, in the future,\n\t\t\t\t\tINACTIVE
services may be cleaned up and purged from Amazon ECS record\n\t\t\t\tkeeping, and DescribeServices calls on those services return a\n\t\t\t\t\tServiceNotFoundException
error.
If you attempt to create a new service with the same name as an existing service\n\t\t\t\tin either ACTIVE
or DRAINING
status, you receive an\n\t\t\t\terror.
The rolloutState
of a service is only returned for services that use\n\t\t\t\tthe rolling update (ECS
) deployment type that aren't behind a\n\t\t\t\tClassic Load Balancer.
The rollout state of the deployment. When a service deployment is started, it begins\n\t\t\tin an IN_PROGRESS
state. When the service reaches a steady state, the\n\t\t\tdeployment transitions to a COMPLETED
state. If the service fails to reach\n\t\t\ta steady state and circuit breaker is turned on, the deployment transitions to a\n\t\t\t\tFAILED
state. A deployment in FAILED
state doesn't launch\n\t\t\tany new tasks. For more information, see DeploymentCircuitBreaker.
The rolloutState
of a service is only returned for services that use\n\t\t\t\tthe rolling update (ECS
) deployment type that aren't behind a\n\t\t\t\tClassic Load Balancer.
The rollout state of the deployment. When a service deployment is started, it begins\n\t\t\tin an IN_PROGRESS
state. When the service reaches a steady state, the\n\t\t\tdeployment transitions to a COMPLETED
state. If the service fails to reach\n\t\t\ta steady state and circuit breaker is turned on, the deployment transitions to a\n\t\t\t\tFAILED
state. A deployment in FAILED
state doesn't launch\n\t\t\tany new tasks. For more information, see DeploymentCircuitBreaker.
The deployment controller type to use.
\nThere are three deployment controller types available:
\nThe rolling update (ECS
) deployment type involves replacing\n\t\t\t\t\t\tthe current running version of the container with the latest version. The\n\t\t\t\t\t\tnumber of containers Amazon ECS adds or removes from the service during a rolling\n\t\t\t\t\t\tupdate is controlled by adjusting the minimum and maximum number of healthy\n\t\t\t\t\t\ttasks allowed during a service deployment, as specified in the DeploymentConfiguration.
The blue/green (CODE_DEPLOY
) deployment type uses the\n\t\t\t\t\t\tblue/green deployment model powered by CodeDeploy, which allows you to verify a\n\t\t\t\t\t\tnew deployment of a service before sending production traffic to it.
The external (EXTERNAL
) deployment type enables you to use\n\t\t\t\t\t\tany third-party deployment controller for full control over the deployment\n\t\t\t\t\t\tprocess for an Amazon ECS service.
The deployment controller type to use.
\nThere are three deployment controller types available:
\nThe rolling update (ECS
) deployment type involves replacing\n\t\t\t\t\t\tthe current running version of the container with the latest version. The\n\t\t\t\t\t\tnumber of containers Amazon ECS adds or removes from the service during a rolling\n\t\t\t\t\t\tupdate is controlled by adjusting the minimum and maximum number of healthy\n\t\t\t\t\t\ttasks allowed during a service deployment, as specified in the DeploymentConfiguration.
The blue/green (CODE_DEPLOY
) deployment type uses the\n\t\t\t\t\t\tblue/green deployment model powered by CodeDeploy, which allows you to verify a\n\t\t\t\t\t\tnew deployment of a service before sending production traffic to it.
The external (EXTERNAL
) deployment type enables you to use\n\t\t\t\t\t\tany third-party deployment controller for full control over the deployment\n\t\t\t\t\t\tprocess for an Amazon ECS service.
The Docker volume driver to use. The driver value must match the driver name provided\n\t\t\tby Docker because it is used for task placement. If the driver was installed using the\n\t\t\tDocker plugin CLI, use docker plugin ls
to retrieve the driver name from\n\t\t\tyour container instance. If the driver was installed using another method, use Docker\n\t\t\tplugin discovery to retrieve the driver name. This parameter maps to Driver
in the docker create-container command and the\n\t\t\t\txxdriver
option to docker\n\t\t\t\tvolume create.
The Docker volume driver to use. The driver value must match the driver name provided\n\t\t\tby Docker because it is used for task placement. If the driver was installed using the\n\t\t\tDocker plugin CLI, use docker plugin ls
to retrieve the driver name from\n\t\t\tyour container instance. If the driver was installed using another method, use Docker\n\t\t\tplugin discovery to retrieve the driver name. This parameter maps to Driver
in the docker conainer create command and the\n\t\t\t\txxdriver
option to docker\n\t\t\t\tvolume create.
Custom metadata to add to your Docker volume. This parameter maps to\n\t\t\t\tLabels
in the docker create-container command and the xxlabel
option to docker\n\t\t\t\tvolume create.
Custom metadata to add to your Docker volume. This parameter maps to\n\t\t\t\tLabels
in the docker conainer create command and the xxlabel
option to docker\n\t\t\t\tvolume create.
A string array representing the command that the container runs to determine if it is\n\t\t\thealthy. The string array must start with CMD
to run the command arguments\n\t\t\tdirectly, or CMD-SHELL
to run the command with the container's default\n\t\t\tshell.
When you use the Amazon Web Services Management Console JSON panel, the Command Line Interface, or the APIs, enclose the list\n\t\t\tof commands in double quotes and brackets.
\n\n [ \"CMD-SHELL\", \"curl -f http://localhost/ || exit 1\" ]
\n
You don't include the double quotes and brackets when you use the Amazon Web Services Management Console.
\n\n CMD-SHELL, curl -f http://localhost/ || exit 1
\n
An exit code of 0 indicates success, and non-zero exit code indicates failure. For\n\t\t\tmore information, see HealthCheck
in tthe docker create-container command
A string array representing the command that the container runs to determine if it is\n\t\t\thealthy. The string array must start with CMD
to run the command arguments\n\t\t\tdirectly, or CMD-SHELL
to run the command with the container's default\n\t\t\tshell.
When you use the Amazon Web Services Management Console JSON panel, the Command Line Interface, or the APIs, enclose the list\n\t\t\tof commands in double quotes and brackets.
\n\n [ \"CMD-SHELL\", \"curl -f http://localhost/ || exit 1\" ]
\n
You don't include the double quotes and brackets when you use the Amazon Web Services Management Console.
\n\n CMD-SHELL, curl -f http://localhost/ || exit 1
\n
An exit code of 0 indicates success, and non-zero exit code indicates failure. For\n\t\t\tmore information, see HealthCheck
in tthe docker conainer create command
Hostnames and IP address entries that are added to the /etc/hosts
file of\n\t\t\ta container via the extraHosts
parameter of its ContainerDefinition.
Hostnames and IP address entries that are added to the /etc/hosts
file of\n\t\t\ta container via the extraHosts
parameter of its ContainerDefinition.
The Linux capabilities for the container that have been added to the default\n\t\t\tconfiguration provided by Docker. This parameter maps to CapAdd
in the docker create-container command and the\n\t\t\t\t--cap-add
option to docker\n\t\t\t\trun.
Tasks launched on Fargate only support adding the SYS_PTRACE
kernel\n\t\t\t\tcapability.
Valid values: \"ALL\" | \"AUDIT_CONTROL\" | \"AUDIT_WRITE\" | \"BLOCK_SUSPEND\" |\n\t\t\t\t\"CHOWN\" | \"DAC_OVERRIDE\" | \"DAC_READ_SEARCH\" | \"FOWNER\" | \"FSETID\" | \"IPC_LOCK\" |\n\t\t\t\t\"IPC_OWNER\" | \"KILL\" | \"LEASE\" | \"LINUX_IMMUTABLE\" | \"MAC_ADMIN\" | \"MAC_OVERRIDE\" |\n\t\t\t\t\"MKNOD\" | \"NET_ADMIN\" | \"NET_BIND_SERVICE\" | \"NET_BROADCAST\" | \"NET_RAW\" | \"SETFCAP\"\n\t\t\t\t| \"SETGID\" | \"SETPCAP\" | \"SETUID\" | \"SYS_ADMIN\" | \"SYS_BOOT\" | \"SYS_CHROOT\" |\n\t\t\t\t\"SYS_MODULE\" | \"SYS_NICE\" | \"SYS_PACCT\" | \"SYS_PTRACE\" | \"SYS_RAWIO\" |\n\t\t\t\t\"SYS_RESOURCE\" | \"SYS_TIME\" | \"SYS_TTY_CONFIG\" | \"SYSLOG\" |\n\t\t\t\"WAKE_ALARM\"
\n
The Linux capabilities for the container that have been added to the default\n\t\t\tconfiguration provided by Docker. This parameter maps to CapAdd
in the docker conainer create command and the\n\t\t\t\t--cap-add
option to docker\n\t\t\t\trun.
Tasks launched on Fargate only support adding the SYS_PTRACE
kernel\n\t\t\t\tcapability.
Valid values: \"ALL\" | \"AUDIT_CONTROL\" | \"AUDIT_WRITE\" | \"BLOCK_SUSPEND\" |\n\t\t\t\t\"CHOWN\" | \"DAC_OVERRIDE\" | \"DAC_READ_SEARCH\" | \"FOWNER\" | \"FSETID\" | \"IPC_LOCK\" |\n\t\t\t\t\"IPC_OWNER\" | \"KILL\" | \"LEASE\" | \"LINUX_IMMUTABLE\" | \"MAC_ADMIN\" | \"MAC_OVERRIDE\" |\n\t\t\t\t\"MKNOD\" | \"NET_ADMIN\" | \"NET_BIND_SERVICE\" | \"NET_BROADCAST\" | \"NET_RAW\" | \"SETFCAP\"\n\t\t\t\t| \"SETGID\" | \"SETPCAP\" | \"SETUID\" | \"SYS_ADMIN\" | \"SYS_BOOT\" | \"SYS_CHROOT\" |\n\t\t\t\t\"SYS_MODULE\" | \"SYS_NICE\" | \"SYS_PACCT\" | \"SYS_PTRACE\" | \"SYS_RAWIO\" |\n\t\t\t\t\"SYS_RESOURCE\" | \"SYS_TIME\" | \"SYS_TTY_CONFIG\" | \"SYSLOG\" |\n\t\t\t\"WAKE_ALARM\"
\n
The Linux capabilities for the container that have been removed from the default\n\t\t\tconfiguration provided by Docker. This parameter maps to CapDrop
in the docker create-container command and the\n\t\t\t\t--cap-drop
option to docker\n\t\t\t\trun.
Valid values: \"ALL\" | \"AUDIT_CONTROL\" | \"AUDIT_WRITE\" | \"BLOCK_SUSPEND\" |\n\t\t\t\t\"CHOWN\" | \"DAC_OVERRIDE\" | \"DAC_READ_SEARCH\" | \"FOWNER\" | \"FSETID\" | \"IPC_LOCK\" |\n\t\t\t\t\"IPC_OWNER\" | \"KILL\" | \"LEASE\" | \"LINUX_IMMUTABLE\" | \"MAC_ADMIN\" | \"MAC_OVERRIDE\" |\n\t\t\t\t\"MKNOD\" | \"NET_ADMIN\" | \"NET_BIND_SERVICE\" | \"NET_BROADCAST\" | \"NET_RAW\" | \"SETFCAP\"\n\t\t\t\t| \"SETGID\" | \"SETPCAP\" | \"SETUID\" | \"SYS_ADMIN\" | \"SYS_BOOT\" | \"SYS_CHROOT\" |\n\t\t\t\t\"SYS_MODULE\" | \"SYS_NICE\" | \"SYS_PACCT\" | \"SYS_PTRACE\" | \"SYS_RAWIO\" |\n\t\t\t\t\"SYS_RESOURCE\" | \"SYS_TIME\" | \"SYS_TTY_CONFIG\" | \"SYSLOG\" |\n\t\t\t\"WAKE_ALARM\"
\n
The Linux capabilities for the container that have been removed from the default\n\t\t\tconfiguration provided by Docker. This parameter maps to CapDrop
in the docker conainer create command and the\n\t\t\t\t--cap-drop
option to docker\n\t\t\t\trun.
Valid values: \"ALL\" | \"AUDIT_CONTROL\" | \"AUDIT_WRITE\" | \"BLOCK_SUSPEND\" |\n\t\t\t\t\"CHOWN\" | \"DAC_OVERRIDE\" | \"DAC_READ_SEARCH\" | \"FOWNER\" | \"FSETID\" | \"IPC_LOCK\" |\n\t\t\t\t\"IPC_OWNER\" | \"KILL\" | \"LEASE\" | \"LINUX_IMMUTABLE\" | \"MAC_ADMIN\" | \"MAC_OVERRIDE\" |\n\t\t\t\t\"MKNOD\" | \"NET_ADMIN\" | \"NET_BIND_SERVICE\" | \"NET_BROADCAST\" | \"NET_RAW\" | \"SETFCAP\"\n\t\t\t\t| \"SETGID\" | \"SETPCAP\" | \"SETUID\" | \"SYS_ADMIN\" | \"SYS_BOOT\" | \"SYS_CHROOT\" |\n\t\t\t\t\"SYS_MODULE\" | \"SYS_NICE\" | \"SYS_PACCT\" | \"SYS_PTRACE\" | \"SYS_RAWIO\" |\n\t\t\t\t\"SYS_RESOURCE\" | \"SYS_TIME\" | \"SYS_TTY_CONFIG\" | \"SYSLOG\" |\n\t\t\t\"WAKE_ALARM\"
\n
Any host devices to expose to the container. This parameter maps to\n\t\t\tDevices
in tthe docker create-container command and the --device
option to docker run.
If you're using tasks that use the Fargate launch type, the\n\t\t\t\t\tdevices
parameter isn't supported.
Any host devices to expose to the container. This parameter maps to\n\t\t\tDevices
in tthe docker conainer create command and the --device
option to docker run.
If you're using tasks that use the Fargate launch type, the\n\t\t\t\t\tdevices
parameter isn't supported.
Run an init
process inside the container that forwards signals and reaps\n\t\t\tprocesses. This parameter maps to the --init
option to docker run. This parameter requires version 1.25 of the Docker Remote API or greater on your container instance. To check the Docker Remote API version on your container instance, log in to your container instance and run the following command: sudo docker version --format '{{.Server.APIVersion}}'
\n
Run an init
process inside the container that forwards signals and reaps\n\t\t\tprocesses. This parameter maps to the --init
option to docker run.\n\t\t\tThis parameter requires version 1.25 of the Docker Remote API or greater on your container instance. To check the Docker Remote API version on your container instance, log in to your container instance and run the following command: sudo docker version --format '{{.Server.APIVersion}}'
\n
The total amount of swap memory (in MiB) a container can use. This parameter will be\n\t\t\ttranslated to the --memory-swap
option to docker run where the value would be the sum of\n\t\t\tthe container memory plus the maxSwap
value.
If a maxSwap
value of 0
is specified, the container will not\n\t\t\tuse swap. Accepted values are 0
or any positive integer. If the\n\t\t\t\tmaxSwap
parameter is omitted, the container will use the swap\n\t\t\tconfiguration for the container instance it is running on. A maxSwap
value\n\t\t\tmust be set for the swappiness
parameter to be used.
If you're using tasks that use the Fargate launch type, the\n\t\t\t\t\tmaxSwap
parameter isn't supported.
If you're using tasks on Amazon Linux 2023 the swappiness
parameter isn't\n\t\t\t\tsupported.
The total amount of swap memory (in MiB) a container can use. This parameter will be\n\t\t\ttranslated to the --memory-swap
option to docker run where the value would be the sum of\n\t\t\tthe container memory plus the maxSwap
value.
If a maxSwap
value of 0
is specified, the container will not\n\t\t\tuse swap. Accepted values are 0
or any positive integer. If the\n\t\t\t\tmaxSwap
parameter is omitted, the container will use the swap\n\t\t\tconfiguration for the container instance it is running on. A maxSwap
value\n\t\t\tmust be set for the swappiness
parameter to be used.
If you're using tasks that use the Fargate launch type, the\n\t\t\t\t\tmaxSwap
parameter isn't supported.
If you're using tasks on Amazon Linux 2023 the swappiness
parameter isn't\n\t\t\t\tsupported.
Filters the container instances by status. For example, if you specify the\n\t\t\t\tDRAINING
status, the results include only container instances that have\n\t\t\tbeen set to DRAINING
using UpdateContainerInstancesState.\n\t\t\tIf you don't specify this parameter, the default is to include container instances set\n\t\t\tto all states other than INACTIVE
.
Filters the container instances by status. For example, if you specify the\n\t\t\t\tDRAINING
status, the results include only container instances that have\n\t\t\tbeen set to DRAINING
using UpdateContainerInstancesState.\n\t\t\tIf you don't specify this parameter, the default is to include container instances set\n\t\t\tto all states other than INACTIVE
.
The log configuration for the container. This parameter maps to LogConfig
\n\t\t\tin the docker create-container command and the\n\t\t\t\t--log-driver
option to docker\n\t\t\t\t\trun.
By default, containers use the same logging driver that the Docker daemon uses.\n\t\t\tHowever, the container might use a different logging driver than the Docker daemon by\n\t\t\tspecifying a log driver configuration in the container definition.
\nUnderstand the following when specifying a log configuration for your\n\t\t\tcontainers.
\nAmazon ECS currently supports a subset of the logging drivers available to the\n\t\t\t\t\tDocker daemon. Additional log drivers may be available in future releases of the\n\t\t\t\t\tAmazon ECS container agent.
\nFor tasks on Fargate, the supported log drivers are awslogs
,\n\t\t\t\t\t\tsplunk
, and awsfirelens
.
For tasks hosted on Amazon EC2 instances, the supported log drivers are\n\t\t\t\t\t\tawslogs
, fluentd
, gelf
,\n\t\t\t\t\t\tjson-file
, journald
,syslog
,\n\t\t\t\t\t\tsplunk
, and awsfirelens
.
This parameter requires version 1.18 of the Docker Remote API or greater on\n\t\t\t\t\tyour container instance.
\nFor tasks that are hosted on Amazon EC2 instances, the Amazon ECS container agent must\n\t\t\t\t\tregister the available logging drivers with the\n\t\t\t\t\t\tECS_AVAILABLE_LOGGING_DRIVERS
environment variable before\n\t\t\t\t\tcontainers placed on that instance can use these log configuration options. For\n\t\t\t\t\tmore information, see Amazon ECS container agent configuration in the\n\t\t\t\t\tAmazon Elastic Container Service Developer Guide.
For tasks that are on Fargate, because you don't have access to the\n\t\t\t\t\tunderlying infrastructure your tasks are hosted on, any additional software\n\t\t\t\t\tneeded must be installed outside of the task. For example, the Fluentd output\n\t\t\t\t\taggregators or a remote host running Logstash to send Gelf logs to.
\nThe log configuration for the container. This parameter maps to LogConfig
\n\t\t\tin the docker conainer create command and the\n\t\t\t\t--log-driver
option to docker\n\t\t\t\t\trun.
By default, containers use the same logging driver that the Docker daemon uses.\n\t\t\tHowever, the container might use a different logging driver than the Docker daemon by\n\t\t\tspecifying a log driver configuration in the container definition.
\nUnderstand the following when specifying a log configuration for your\n\t\t\tcontainers.
\nAmazon ECS currently supports a subset of the logging drivers available to the\n\t\t\t\t\tDocker daemon. Additional log drivers may be available in future releases of the\n\t\t\t\t\tAmazon ECS container agent.
\nFor tasks on Fargate, the supported log drivers are awslogs
,\n\t\t\t\t\t\tsplunk
, and awsfirelens
.
For tasks hosted on Amazon EC2 instances, the supported log drivers are\n\t\t\t\t\t\tawslogs
, fluentd
, gelf
,\n\t\t\t\t\t\tjson-file
, journald
,syslog
,\n\t\t\t\t\t\tsplunk
, and awsfirelens
.
This parameter requires version 1.18 of the Docker Remote API or greater on\n\t\t\t\t\tyour container instance.
\nFor tasks that are hosted on Amazon EC2 instances, the Amazon ECS container agent must\n\t\t\t\t\tregister the available logging drivers with the\n\t\t\t\t\t\tECS_AVAILABLE_LOGGING_DRIVERS
environment variable before\n\t\t\t\t\tcontainers placed on that instance can use these log configuration options. For\n\t\t\t\t\tmore information, see Amazon ECS container agent configuration in the\n\t\t\t\t\tAmazon Elastic Container Service Developer Guide.
For tasks that are on Fargate, because you don't have access to the\n\t\t\t\t\tunderlying infrastructure your tasks are hosted on, any additional software\n\t\t\t\t\tneeded must be installed outside of the task. For example, the Fluentd output\n\t\t\t\t\taggregators or a remote host running Logstash to send Gelf logs to.
\nThe maximum number of Amazon EC2 instances that Amazon ECS will scale out at one time. The scale\n\t\t\tin process is not affected by this parameter. If this parameter is omitted, the default\n\t\t\tvalue of 10000
is used.
The maximum number of Amazon EC2 instances that Amazon ECS will scale out at one time. If this parameter is omitted, the default\n\t\t\tvalue of 10000
is used.
Details on the network bindings between a container and its host container instance.\n\t\t\tAfter a task reaches the RUNNING
status, manual and automatic host and\n\t\t\tcontainer port assignments are visible in the networkBindings
section of\n\t\t\t\tDescribeTasks API responses.
Details on the network bindings between a container and its host container instance.\n\t\t\tAfter a task reaches the RUNNING
status, manual and automatic host and\n\t\t\tcontainer port assignments are visible in the networkBindings
section of\n\t\t\tDescribeTasks API responses.
Port mappings allow containers to access ports on the host container instance to send\n\t\t\tor receive traffic. Port mappings are specified as part of the container\n\t\t\tdefinition.
\nIf you use containers in a task with the awsvpc
or host
\n\t\t\tnetwork mode, specify the exposed ports using containerPort
. The\n\t\t\t\thostPort
can be left blank or it must be the same value as the\n\t\t\t\tcontainerPort
.
Most fields of this parameter (containerPort
, hostPort
,\n\t\t\tprotocol
) maps to PortBindings
in the docker create-container command and the\n\t\t\t\t--publish
option to docker\n\t\t\t\t\trun
. If the network mode of a task definition is set to\n\t\t\t\thost
, host ports must either be undefined or match the container port\n\t\t\tin the port mapping.
You can't expose the same container port for multiple protocols. If you attempt\n\t\t\t\tthis, an error is returned.
\nAfter a task reaches the RUNNING
status, manual and automatic host and\n\t\t\tcontainer port assignments are visible in the networkBindings
section of\n\t\t\t\tDescribeTasks API responses.
Port mappings allow containers to access ports on the host container instance to send\n\t\t\tor receive traffic. Port mappings are specified as part of the container\n\t\t\tdefinition.
\nIf you use containers in a task with the awsvpc
or host
\n\t\t\tnetwork mode, specify the exposed ports using containerPort
. The\n\t\t\t\thostPort
can be left blank or it must be the same value as the\n\t\t\t\tcontainerPort
.
Most fields of this parameter (containerPort
, hostPort
,\n\t\t\tprotocol
) maps to PortBindings
in the docker conainer create command and the\n\t\t\t\t--publish
option to docker\n\t\t\t\t\trun
. If the network mode of a task definition is set to\n\t\t\t\thost
, host ports must either be undefined or match the container port\n\t\t\tin the port mapping.
You can't expose the same container port for multiple protocols. If you attempt\n\t\t\t\tthis, an error is returned.
\nAfter a task reaches the RUNNING
status, manual and automatic host and\n\t\t\tcontainer port assignments are visible in the networkBindings
section of\n\t\t\tDescribeTasks API responses.
An object representing the protection status details for a task. You can set the\n\t\t\tprotection status with the UpdateTaskProtection API and get the status\n\t\t\tof tasks with the GetTaskProtection API.
" + "smithy.api#documentation": "An object representing the protection status details for a task. You can set the\n\t\t\tprotection status with the UpdateTaskProtection API and get the status\n\t\t\tof tasks with the GetTaskProtection API.
" } }, "com.amazonaws.ecs#ProtectedTasks": { @@ -8871,7 +8871,7 @@ } ], "traits": { - "smithy.api#documentation": "Create or update an attribute on an Amazon ECS resource. If the attribute doesn't exist,\n\t\t\tit's created. If the attribute exists, its value is replaced with the specified value.\n\t\t\tTo delete an attribute, use DeleteAttributes. For more information,\n\t\t\tsee Attributes in the Amazon Elastic Container Service Developer Guide.
" + "smithy.api#documentation": "Create or update an attribute on an Amazon ECS resource. If the attribute doesn't exist,\n\t\t\tit's created. If the attribute exists, its value is replaced with the specified value.\n\t\t\tTo delete an attribute, use DeleteAttributes. For more information,\n\t\t\tsee Attributes in the Amazon Elastic Container Service Developer Guide.
" } }, "com.amazonaws.ecs#PutAttributesRequest": { @@ -8938,7 +8938,7 @@ } ], "traits": { - "smithy.api#documentation": "Modifies the available capacity providers and the default capacity provider strategy\n\t\t\tfor a cluster.
\nYou must specify both the available capacity providers and a default capacity provider\n\t\t\tstrategy for the cluster. If the specified cluster has existing capacity providers\n\t\t\tassociated with it, you must specify all existing capacity providers in addition to any\n\t\t\tnew ones you want to add. Any existing capacity providers that are associated with a\n\t\t\tcluster that are omitted from a PutClusterCapacityProviders API call\n\t\t\twill be disassociated with the cluster. You can only disassociate an existing capacity\n\t\t\tprovider from a cluster if it's not being used by any existing tasks.
\nWhen creating a service or running a task on a cluster, if no capacity provider or\n\t\t\tlaunch type is specified, then the cluster's default capacity provider strategy is used.\n\t\t\tWe recommend that you define a default capacity provider strategy for your cluster.\n\t\t\tHowever, you must specify an empty array ([]
) to bypass defining a default\n\t\t\tstrategy.
Modifies the available capacity providers and the default capacity provider strategy\n\t\t\tfor a cluster.
\nYou must specify both the available capacity providers and a default capacity provider\n\t\t\tstrategy for the cluster. If the specified cluster has existing capacity providers\n\t\t\tassociated with it, you must specify all existing capacity providers in addition to any\n\t\t\tnew ones you want to add. Any existing capacity providers that are associated with a\n\t\t\tcluster that are omitted from a PutClusterCapacityProviders API call\n\t\t\twill be disassociated with the cluster. You can only disassociate an existing capacity\n\t\t\tprovider from a cluster if it's not being used by any existing tasks.
\nWhen creating a service or running a task on a cluster, if no capacity provider or\n\t\t\tlaunch type is specified, then the cluster's default capacity provider strategy is used.\n\t\t\tWe recommend that you define a default capacity provider strategy for your cluster.\n\t\t\tHowever, you must specify an empty array ([]
) to bypass defining a default\n\t\t\tstrategy.
The name of one or more capacity providers to associate with the cluster.
\nIf specifying a capacity provider that uses an Auto Scaling group, the capacity\n\t\t\tprovider must already be created. New capacity providers can be created with the CreateCapacityProvider API operation.
\nTo use a Fargate capacity provider, specify either the FARGATE
or\n\t\t\t\tFARGATE_SPOT
capacity providers. The Fargate capacity providers are\n\t\t\tavailable to all accounts and only need to be associated with a cluster to be\n\t\t\tused.
The name of one or more capacity providers to associate with the cluster.
\nIf specifying a capacity provider that uses an Auto Scaling group, the capacity\n\t\t\tprovider must already be created. New capacity providers can be created with the CreateCapacityProvider API operation.
\nTo use a Fargate capacity provider, specify either the FARGATE
or\n\t\t\t\tFARGATE_SPOT
capacity providers. The Fargate capacity providers are\n\t\t\tavailable to all accounts and only need to be associated with a cluster to be\n\t\t\tused.
The capacity provider strategy to use by default for the cluster.
\nWhen creating a service or running a task on a cluster, if no capacity provider or\n\t\t\tlaunch type is specified then the default capacity provider strategy for the cluster is\n\t\t\tused.
\nA capacity provider strategy consists of one or more capacity providers along with the\n\t\t\t\tbase
and weight
to assign to them. A capacity provider\n\t\t\tmust be associated with the cluster to be used in a capacity provider strategy. The\n\t\t\t\tPutClusterCapacityProviders API is used to associate a capacity\n\t\t\tprovider with a cluster. Only capacity providers with an ACTIVE
or\n\t\t\t\tUPDATING
status can be used.
If specifying a capacity provider that uses an Auto Scaling group, the capacity\n\t\t\tprovider must already be created. New capacity providers can be created with the CreateCapacityProvider API operation.
\nTo use a Fargate capacity provider, specify either the FARGATE
or\n\t\t\t\tFARGATE_SPOT
capacity providers. The Fargate capacity providers are\n\t\t\tavailable to all accounts and only need to be associated with a cluster to be\n\t\t\tused.
The capacity provider strategy to use by default for the cluster.
\nWhen creating a service or running a task on a cluster, if no capacity provider or\n\t\t\tlaunch type is specified then the default capacity provider strategy for the cluster is\n\t\t\tused.
\nA capacity provider strategy consists of one or more capacity providers along with the\n\t\t\t\tbase
and weight
to assign to them. A capacity provider\n\t\t\tmust be associated with the cluster to be used in a capacity provider strategy. The\n\t\t\tPutClusterCapacityProviders API is used to associate a capacity\n\t\t\tprovider with a cluster. Only capacity providers with an ACTIVE
or\n\t\t\t\tUPDATING
status can be used.
If specifying a capacity provider that uses an Auto Scaling group, the capacity\n\t\t\tprovider must already be created. New capacity providers can be created with the CreateCapacityProvider API operation.
\nTo use a Fargate capacity provider, specify either the FARGATE
or\n\t\t\t\tFARGATE_SPOT
capacity providers. The Fargate capacity providers are\n\t\t\tavailable to all accounts and only need to be associated with a cluster to be\n\t\t\tused.
Registers a new task definition from the supplied family
and\n\t\t\t\tcontainerDefinitions
. Optionally, you can add data volumes to your\n\t\t\tcontainers with the volumes
parameter. For more information about task\n\t\t\tdefinition parameters and defaults, see Amazon ECS Task\n\t\t\t\tDefinitions in the Amazon Elastic Container Service Developer Guide.
You can specify a role for your task with the taskRoleArn
parameter. When\n\t\t\tyou specify a role for a task, its containers can then use the latest versions of the\n\t\t\tCLI or SDKs to make API requests to the Amazon Web Services services that are specified in the\n\t\t\tpolicy that's associated with the role. For more information, see IAM\n\t\t\t\tRoles for Tasks in the Amazon Elastic Container Service Developer Guide.
You can specify a Docker networking mode for the containers in your task definition\n\t\t\twith the networkMode
parameter. If you specify the awsvpc
\n\t\t\tnetwork mode, the task is allocated an elastic network interface, and you must specify a\n\t\t\t\tNetworkConfiguration when you create a service or run a task with\n\t\t\tthe task definition. For more information, see Task Networking\n\t\t\tin the Amazon Elastic Container Service Developer Guide.
Registers a new task definition from the supplied family
and\n\t\t\t\tcontainerDefinitions
. Optionally, you can add data volumes to your\n\t\t\tcontainers with the volumes
parameter. For more information about task\n\t\t\tdefinition parameters and defaults, see Amazon ECS Task\n\t\t\t\tDefinitions in the Amazon Elastic Container Service Developer Guide.
You can specify a role for your task with the taskRoleArn
parameter. When\n\t\t\tyou specify a role for a task, its containers can then use the latest versions of the\n\t\t\tCLI or SDKs to make API requests to the Amazon Web Services services that are specified in the\n\t\t\tpolicy that's associated with the role. For more information, see IAM\n\t\t\t\tRoles for Tasks in the Amazon Elastic Container Service Developer Guide.
You can specify a Docker networking mode for the containers in your task definition\n\t\t\twith the networkMode
parameter. If you specify the awsvpc
\n\t\t\tnetwork mode, the task is allocated an elastic network interface, and you must specify a NetworkConfiguration when you create a service or run a task with\n\t\t\tthe task definition. For more information, see Task Networking\n\t\t\tin the Amazon Elastic Container Service Developer Guide.
Starts a new task using the specified task definition.
\nOn March 21, 2024, a change was made to resolve the task definition revision before authorization. When a task definition revision is not specified, authorization will occur using the latest revision of a task definition.
\nYou can allow Amazon ECS to place tasks for you, or you can customize how Amazon ECS places\n\t\t\ttasks using placement constraints and placement strategies. For more information, see\n\t\t\t\tScheduling Tasks in the Amazon Elastic Container Service Developer Guide.
\nAlternatively, you can use StartTask to use your own scheduler or\n\t\t\tplace tasks manually on specific container instances.
\nStarting April 15, 2023, Amazon Web Services will not onboard new customers to Amazon Elastic Inference (EI), and will help current customers migrate their workloads to options that offer better price and performance. After April 15, 2023, new customers will not be able to launch instances with Amazon EI accelerators in Amazon SageMaker, Amazon ECS, or Amazon EC2. However, customers who have used Amazon EI at least once during the past 30-day period are considered current customers and will be able to continue using the service.
\nYou can attach Amazon EBS volumes to Amazon ECS tasks by configuring the volume when creating or\n\t\t\tupdating a service. For more infomation, see Amazon EBS volumes in the Amazon Elastic Container Service Developer Guide.
\nThe Amazon ECS API follows an eventual consistency model. This is because of the\n\t\t\tdistributed nature of the system supporting the API. This means that the result of an\n\t\t\tAPI command you run that affects your Amazon ECS resources might not be immediately visible\n\t\t\tto all subsequent commands you run. Keep this in mind when you carry out an API command\n\t\t\tthat immediately follows a previous API command.
\nTo manage eventual consistency, you can do the following:
\nConfirm the state of the resource before you run a command to modify it. Run\n\t\t\t\t\tthe DescribeTasks command using an exponential backoff algorithm to ensure that\n\t\t\t\t\tyou allow enough time for the previous command to propagate through the system.\n\t\t\t\t\tTo do this, run the DescribeTasks command repeatedly, starting with a couple of\n\t\t\t\t\tseconds of wait time and increasing gradually up to five minutes of wait\n\t\t\t\t\ttime.
\nAdd wait time between subsequent commands, even if the DescribeTasks command\n\t\t\t\t\treturns an accurate response. Apply an exponential backoff algorithm starting\n\t\t\t\t\twith a couple of seconds of wait time, and increase gradually up to about five\n\t\t\t\t\tminutes of wait time.
\nStarts a new task using the specified task definition.
\nOn March 21, 2024, a change was made to resolve the task definition revision before authorization. When a task definition revision is not specified, authorization will occur using the latest revision of a task definition.
\nYou can allow Amazon ECS to place tasks for you, or you can customize how Amazon ECS places\n\t\t\ttasks using placement constraints and placement strategies. For more information, see\n\t\t\t\tScheduling Tasks in the Amazon Elastic Container Service Developer Guide.
\nAlternatively, you can use StartTask
to use your own scheduler or\n\t\t\tplace tasks manually on specific container instances.
Starting April 15, 2023, Amazon Web Services will not onboard new customers to Amazon Elastic Inference (EI), and will help current customers migrate their workloads to options that offer better price and performance. After April 15, 2023, new customers will not be able to launch instances with Amazon EI accelerators in Amazon SageMaker, Amazon ECS, or Amazon EC2. However, customers who have used Amazon EI at least once during the past 30-day period are considered current customers and will be able to continue using the service.
\nYou can attach Amazon EBS volumes to Amazon ECS tasks by configuring the volume when creating or\n\t\t\tupdating a service. For more infomation, see Amazon EBS volumes in the Amazon Elastic Container Service Developer Guide.
\nThe Amazon ECS API follows an eventual consistency model. This is because of the\n\t\t\tdistributed nature of the system supporting the API. This means that the result of an\n\t\t\tAPI command you run that affects your Amazon ECS resources might not be immediately visible\n\t\t\tto all subsequent commands you run. Keep this in mind when you carry out an API command\n\t\t\tthat immediately follows a previous API command.
\nTo manage eventual consistency, you can do the following:
\nConfirm the state of the resource before you run a command to modify it. Run\n\t\t\t\t\tthe DescribeTasks command using an exponential backoff algorithm to ensure that\n\t\t\t\t\tyou allow enough time for the previous command to propagate through the system.\n\t\t\t\t\tTo do this, run the DescribeTasks command repeatedly, starting with a couple of\n\t\t\t\t\tseconds of wait time and increasing gradually up to five minutes of wait\n\t\t\t\t\ttime.
\nAdd wait time between subsequent commands, even if the DescribeTasks command\n\t\t\t\t\treturns an accurate response. Apply an exponential backoff algorithm starting\n\t\t\t\t\twith a couple of seconds of wait time, and increase gradually up to about five\n\t\t\t\t\tminutes of wait time.
\nSpecifies whether to propagate the tags from the task definition to the task. If no\n\t\t\tvalue is specified, the tags aren't propagated. Tags can only be propagated to the task\n\t\t\tduring task creation. To add tags to a task after task creation, use the TagResource API action.
\nAn error will be received if you specify the SERVICE
option when\n\t\t\t\trunning a task.
Specifies whether to propagate the tags from the task definition to the task. If no\n\t\t\tvalue is specified, the tags aren't propagated. Tags can only be propagated to the task\n\t\t\tduring task creation. To add tags to a task after task creation, use theTagResource API action.
\nAn error will be received if you specify the SERVICE
option when\n\t\t\t\trunning a task.
The desired number of instantiations of the task definition to keep running on the\n\t\t\tservice. This value is specified when the service is created with CreateService, and it can be modified with UpdateService.
" + "smithy.api#documentation": "The desired number of instantiations of the task definition to keep running on the\n\t\t\tservice. This value is specified when the service is created with CreateService , and it can be modified with UpdateService.
" } }, "runningCount": { @@ -9884,7 +9884,7 @@ "taskDefinition": { "target": "com.amazonaws.ecs#String", "traits": { - "smithy.api#documentation": "The task definition to use for tasks in the service. This value is specified when the\n\t\t\tservice is created with CreateService, and it can be modified with\n\t\t\t\tUpdateService.
" + "smithy.api#documentation": "The task definition to use for tasks in the service. This value is specified when the\n\t\t\tservice is created with CreateService, and it can be modified with\n\t\t\tUpdateService.
" } }, "deploymentConfiguration": { @@ -10301,7 +10301,7 @@ } }, "traits": { - "smithy.api#documentation": "The specified service isn't active. You can't update a service that's inactive. If you\n\t\t\thave previously deleted a service, you can re-create it with CreateService.
", + "smithy.api#documentation": "The specified service isn't active. You can't update a service that's inactive. If you\n\t\t\thave previously deleted a service, you can re-create it with CreateService.
", "smithy.api#error": "client" } }, @@ -10316,7 +10316,7 @@ } }, "traits": { - "smithy.api#documentation": "The specified service wasn't found. You can view your available services with ListServices. Amazon ECS services are cluster specific and Region\n\t\t\tspecific.
", + "smithy.api#documentation": "The specified service wasn't found. You can view your available services with ListServices. Amazon ECS services are cluster specific and Region\n\t\t\tspecific.
", "smithy.api#error": "client" } }, @@ -10591,7 +10591,7 @@ } ], "traits": { - "smithy.api#documentation": "Starts a new task from the specified task definition on the specified container\n\t\t\tinstance or instances.
\nOn March 21, 2024, a change was made to resolve the task definition revision before authorization. When a task definition revision is not specified, authorization will occur using the latest revision of a task definition.
\nStarting April 15, 2023, Amazon Web Services will not onboard new customers to Amazon Elastic Inference (EI), and will help current customers migrate their workloads to options that offer better price and performance. After April 15, 2023, new customers will not be able to launch instances with Amazon EI accelerators in Amazon SageMaker, Amazon ECS, or Amazon EC2. However, customers who have used Amazon EI at least once during the past 30-day period are considered current customers and will be able to continue using the service.
\nAlternatively, you can use RunTask to place tasks for you. For more\n\t\t\tinformation, see Scheduling Tasks in the Amazon Elastic Container Service Developer Guide.
\nYou can attach Amazon EBS volumes to Amazon ECS tasks by configuring the volume when creating or\n\t\t\tupdating a service. For more infomation, see Amazon EBS volumes in the Amazon Elastic Container Service Developer Guide.
" + "smithy.api#documentation": "Starts a new task from the specified task definition on the specified container\n\t\t\tinstance or instances.
\nOn March 21, 2024, a change was made to resolve the task definition revision before authorization. When a task definition revision is not specified, authorization will occur using the latest revision of a task definition.
\nStarting April 15, 2023, Amazon Web Services will not onboard new customers to Amazon Elastic Inference (EI), and will help current customers migrate their workloads to options that offer better price and performance. After April 15, 2023, new customers will not be able to launch instances with Amazon EI accelerators in Amazon SageMaker, Amazon ECS, or Amazon EC2. However, customers who have used Amazon EI at least once during the past 30-day period are considered current customers and will be able to continue using the service.
\nAlternatively, you can useRunTask
to place tasks for you. For more\n\t\t\tinformation, see Scheduling Tasks in the Amazon Elastic Container Service Developer Guide.
You can attach Amazon EBS volumes to Amazon ECS tasks by configuring the volume when creating or\n\t\t\tupdating a service. For more infomation, see Amazon EBS volumes in the Amazon Elastic Container Service Developer Guide.
" } }, "com.amazonaws.ecs#StartTaskRequest": { @@ -10733,7 +10733,7 @@ } ], "traits": { - "smithy.api#documentation": "Stops a running task. Any tags associated with the task will be deleted.
\nWhen StopTask is called on a task, the equivalent of docker\n\t\t\t\tstop
is issued to the containers running in the task. This results in a\n\t\t\t\tSIGTERM
value and a default 30-second timeout, after which the\n\t\t\t\tSIGKILL
value is sent and the containers are forcibly stopped. If the\n\t\t\tcontainer handles the SIGTERM
value gracefully and exits within 30 seconds\n\t\t\tfrom receiving it, no SIGKILL
value is sent.
For Windows containers, POSIX signals do not work and runtime stops the container by\n\t\t\tsending a CTRL_SHUTDOWN_EVENT
. For more information, see Unable to react to graceful shutdown\n\t\t\t\tof (Windows) container #25982 on GitHub.
The default 30-second timeout can be configured on the Amazon ECS container agent with\n\t\t\t\tthe ECS_CONTAINER_STOP_TIMEOUT
variable. For more information, see\n\t\t\t\t\tAmazon ECS Container Agent Configuration in the\n\t\t\t\tAmazon Elastic Container Service Developer Guide.
Stops a running task. Any tags associated with the task will be deleted.
\nWhen you call StopTask
on a task, the equivalent of docker\n\t\t\t\tstop
is issued to the containers running in the task. This results in a\n\t\t\t\tSIGTERM
value and a default 30-second timeout, after which the\n\t\t\t\tSIGKILL
value is sent and the containers are forcibly stopped. If the\n\t\t\tcontainer handles the SIGTERM
value gracefully and exits within 30 seconds\n\t\t\tfrom receiving it, no SIGKILL
value is sent.
For Windows containers, POSIX signals do not work and runtime stops the container by\n\t\t\tsending a CTRL_SHUTDOWN_EVENT
. For more information, see Unable to react to graceful shutdown\n\t\t\t\tof (Windows) container #25982 on GitHub.
The default 30-second timeout can be configured on the Amazon ECS container agent with\n\t\t\t\tthe ECS_CONTAINER_STOP_TIMEOUT
variable. For more information, see\n\t\t\t\t\tAmazon ECS Container Agent Configuration in the\n\t\t\t\tAmazon Elastic Container Service Developer Guide.
An optional message specified when a task is stopped. For example, if you're using a\n\t\t\tcustom scheduler, you can use this parameter to specify the reason for stopping the task\n\t\t\there, and the message appears in subsequent DescribeTasks API\n\t\t\toperations on this task.
" + "smithy.api#documentation": "An optional message specified when a task is stopped. For example, if you're using a\n\t\t\tcustom scheduler, you can use this parameter to specify the reason for stopping the task\n\t\t\there, and the message appears in subsequent DescribeTasks> API\n\t\t\toperations on this task.
" } } }, @@ -11074,7 +11074,7 @@ } }, "traits": { - "smithy.api#documentation": "A list of namespaced kernel parameters to set in the container. This parameter maps to\n\t\t\tSysctls
in tthe docker create-container command and the --sysctl
option to docker run. For example, you can configure\n\t\t\t\tnet.ipv4.tcp_keepalive_time
setting to maintain longer lived\n\t\t\tconnections.
We don't recommend that you specify network-related systemControls
\n\t\t\tparameters for multiple containers in a single task that also uses either the\n\t\t\t\tawsvpc
or host
network mode. Doing this has the following\n\t\t\tdisadvantages:
For tasks that use the awsvpc
network mode including Fargate,\n\t\t\t\t\tif you set systemControls
for any container, it applies to all\n\t\t\t\t\tcontainers in the task. If you set different systemControls
for\n\t\t\t\t\tmultiple containers in a single task, the container that's started last\n\t\t\t\t\tdetermines which systemControls
take effect.
For tasks that use the host
network mode, the network namespace\n\t\t\t\t\t\tsystemControls
aren't supported.
If you're setting an IPC resource namespace to use for the containers in the task, the\n\t\t\tfollowing conditions apply to your system controls. For more information, see IPC mode.
\nFor tasks that use the host
IPC mode, IPC namespace\n\t\t\t\t\t\tsystemControls
aren't supported.
For tasks that use the task
IPC mode, IPC namespace\n\t\t\t\t\t\tsystemControls
values apply to all containers within a\n\t\t\t\t\ttask.
This parameter is not supported for Windows containers.
\nThis parameter is only supported for tasks that are hosted on\n Fargate if the tasks are using platform version 1.4.0
or later\n (Linux). This isn't supported for Windows containers on\n Fargate.
A list of namespaced kernel parameters to set in the container. This parameter maps to\n\t\t\tSysctls
in tthe docker conainer create command and the --sysctl
option to docker run. For example, you can configure\n\t\t\t\tnet.ipv4.tcp_keepalive_time
setting to maintain longer lived\n\t\t\tconnections.
We don't recommend that you specify network-related systemControls
\n\t\t\tparameters for multiple containers in a single task that also uses either the\n\t\t\t\tawsvpc
or host
network mode. Doing this has the following\n\t\t\tdisadvantages:
For tasks that use the awsvpc
network mode including Fargate,\n\t\t\t\t\tif you set systemControls
for any container, it applies to all\n\t\t\t\t\tcontainers in the task. If you set different systemControls
for\n\t\t\t\t\tmultiple containers in a single task, the container that's started last\n\t\t\t\t\tdetermines which systemControls
take effect.
For tasks that use the host
network mode, the network namespace\n\t\t\t\t\t\tsystemControls
aren't supported.
If you're setting an IPC resource namespace to use for the containers in the task, the\n\t\t\tfollowing conditions apply to your system controls. For more information, see IPC mode.
\nFor tasks that use the host
IPC mode, IPC namespace\n\t\t\t\t\t\tsystemControls
aren't supported.
For tasks that use the task
IPC mode, IPC namespace\n\t\t\t\t\t\tsystemControls
values apply to all containers within a\n\t\t\t\t\ttask.
This parameter is not supported for Windows containers.
\nThis parameter is only supported for tasks that are hosted on\n Fargate if the tasks are using platform version 1.4.0
or later\n (Linux). This isn't supported for Windows containers on\n Fargate.
The container instance attributes required by your task. When an Amazon EC2 instance is\n\t\t\tregistered to your cluster, the Amazon ECS container agent assigns some standard attributes\n\t\t\tto the instance. You can apply custom attributes. These are specified as key-value pairs\n\t\t\tusing the Amazon ECS console or the PutAttributes API. These attributes are\n\t\t\tused when determining task placement for tasks hosted on Amazon EC2 instances. For more\n\t\t\tinformation, see Attributes in the Amazon Elastic Container Service Developer Guide.
\nThis parameter isn't supported for tasks run on Fargate.
\nThe container instance attributes required by your task. When an Amazon EC2 instance is\n\t\t\tregistered to your cluster, the Amazon ECS container agent assigns some standard attributes\n\t\t\tto the instance. You can apply custom attributes. These are specified as key-value pairs\n\t\t\tusing the Amazon ECS console or the PutAttributes API. These attributes are\n\t\t\tused when determining task placement for tasks hosted on Amazon EC2 instances. For more\n\t\t\tinformation, see Attributes in the Amazon Elastic Container Service Developer Guide.
\nThis parameter isn't supported for tasks run on Fargate.
\nThe specified task set wasn't found. You can view your available task sets with DescribeTaskSets. Task sets are specific to each cluster, service and\n\t\t\tRegion.
", + "smithy.api#documentation": "The specified task set wasn't found. You can view your available task sets with DescribeTaskSets. Task sets are specific to each cluster, service and\n\t\t\tRegion.
", "smithy.api#error": "client" } }, @@ -12686,7 +12686,7 @@ "settings": { "target": "com.amazonaws.ecs#ClusterSettings", "traits": { - "smithy.api#documentation": "The setting to use by default for a cluster. This parameter is used to turn on CloudWatch\n\t\t\tContainer Insights for a cluster. If this value is specified, it overrides the\n\t\t\t\tcontainerInsights
value set with PutAccountSetting or\n\t\t\t\tPutAccountSettingDefault.
Currently, if you delete an existing cluster that does not have Container Insights\n\t\t\t\tturned on, and then create a new cluster with the same name with Container Insights\n\t\t\t\ttuned on, Container Insights will not actually be turned on. If you want to preserve\n\t\t\t\tthe same name for your existing cluster and turn on Container Insights, you must\n\t\t\t\twait 7 days before you can re-create it.
\nThe setting to use by default for a cluster. This parameter is used to turn on CloudWatch\n\t\t\tContainer Insights for a cluster. If this value is specified, it overrides the\n\t\t\tcontainerInsights
value set with PutAccountSetting or\n\t\t\tPutAccountSettingDefault.
Currently, if you delete an existing cluster that does not have Container Insights\n\t\t\t\tturned on, and then create a new cluster with the same name with Container Insights\n\t\t\t\ttuned on, Container Insights will not actually be turned on. If you want to preserve\n\t\t\t\tthe same name for your existing cluster and turn on Container Insights, you must\n\t\t\t\twait 7 days before you can re-create it.
\nModifies the status of an Amazon ECS container instance.
\nOnce a container instance has reached an ACTIVE
state, you can change the\n\t\t\tstatus of a container instance to DRAINING
to manually remove an instance\n\t\t\tfrom a cluster, for example to perform system updates, update the Docker daemon, or\n\t\t\tscale down the cluster size.
A container instance can't be changed to DRAINING
until it has\n\t\t\t\treached an ACTIVE
status. If the instance is in any other status, an\n\t\t\t\terror will be received.
When you set a container instance to DRAINING
, Amazon ECS prevents new tasks\n\t\t\tfrom being scheduled for placement on the container instance and replacement service\n\t\t\ttasks are started on other container instances in the cluster if the resources are\n\t\t\tavailable. Service tasks on the container instance that are in the PENDING
\n\t\t\tstate are stopped immediately.
Service tasks on the container instance that are in the RUNNING
state are\n\t\t\tstopped and replaced according to the service's deployment configuration parameters,\n\t\t\t\tminimumHealthyPercent
and maximumPercent
. You can change\n\t\t\tthe deployment configuration of your service using UpdateService.
If minimumHealthyPercent
is below 100%, the scheduler can ignore\n\t\t\t\t\t\tdesiredCount
temporarily during task replacement. For example,\n\t\t\t\t\t\tdesiredCount
is four tasks, a minimum of 50% allows the\n\t\t\t\t\tscheduler to stop two existing tasks before starting two new tasks. If the\n\t\t\t\t\tminimum is 100%, the service scheduler can't remove existing tasks until the\n\t\t\t\t\treplacement tasks are considered healthy. Tasks for services that do not use a\n\t\t\t\t\tload balancer are considered healthy if they're in the RUNNING
\n\t\t\t\t\tstate. Tasks for services that use a load balancer are considered healthy if\n\t\t\t\t\tthey're in the RUNNING
state and are reported as healthy by the\n\t\t\t\t\tload balancer.
The maximumPercent
parameter represents an upper limit on the\n\t\t\t\t\tnumber of running tasks during task replacement. You can use this to define the\n\t\t\t\t\treplacement batch size. For example, if desiredCount
is four tasks,\n\t\t\t\t\ta maximum of 200% starts four new tasks before stopping the four tasks to be\n\t\t\t\t\tdrained, provided that the cluster resources required to do this are available.\n\t\t\t\t\tIf the maximum is 100%, then replacement tasks can't start until the draining\n\t\t\t\t\ttasks have stopped.
Any PENDING
or RUNNING
tasks that do not belong to a service\n\t\t\taren't affected. You must wait for them to finish or stop them manually.
A container instance has completed draining when it has no more RUNNING
\n\t\t\ttasks. You can verify this using ListTasks.
When a container instance has been drained, you can set a container instance to\n\t\t\t\tACTIVE
status and once it has reached that status the Amazon ECS scheduler\n\t\t\tcan begin scheduling tasks on the instance again.
Modifies the status of an Amazon ECS container instance.
\nOnce a container instance has reached an ACTIVE
state, you can change the\n\t\t\tstatus of a container instance to DRAINING
to manually remove an instance\n\t\t\tfrom a cluster, for example to perform system updates, update the Docker daemon, or\n\t\t\tscale down the cluster size.
A container instance can't be changed to DRAINING
until it has\n\t\t\t\treached an ACTIVE
status. If the instance is in any other status, an\n\t\t\t\terror will be received.
When you set a container instance to DRAINING
, Amazon ECS prevents new tasks\n\t\t\tfrom being scheduled for placement on the container instance and replacement service\n\t\t\ttasks are started on other container instances in the cluster if the resources are\n\t\t\tavailable. Service tasks on the container instance that are in the PENDING
\n\t\t\tstate are stopped immediately.
Service tasks on the container instance that are in the RUNNING
state are\n\t\t\tstopped and replaced according to the service's deployment configuration parameters,\n\t\t\t\tminimumHealthyPercent
and maximumPercent
. You can change\n\t\t\tthe deployment configuration of your service using UpdateService.
If minimumHealthyPercent
is below 100%, the scheduler can ignore\n\t\t\t\t\t\tdesiredCount
temporarily during task replacement. For example,\n\t\t\t\t\t\tdesiredCount
is four tasks, a minimum of 50% allows the\n\t\t\t\t\tscheduler to stop two existing tasks before starting two new tasks. If the\n\t\t\t\t\tminimum is 100%, the service scheduler can't remove existing tasks until the\n\t\t\t\t\treplacement tasks are considered healthy. Tasks for services that do not use a\n\t\t\t\t\tload balancer are considered healthy if they're in the RUNNING
\n\t\t\t\t\tstate. Tasks for services that use a load balancer are considered healthy if\n\t\t\t\t\tthey're in the RUNNING
state and are reported as healthy by the\n\t\t\t\t\tload balancer.
The maximumPercent
parameter represents an upper limit on the\n\t\t\t\t\tnumber of running tasks during task replacement. You can use this to define the\n\t\t\t\t\treplacement batch size. For example, if desiredCount
is four tasks,\n\t\t\t\t\ta maximum of 200% starts four new tasks before stopping the four tasks to be\n\t\t\t\t\tdrained, provided that the cluster resources required to do this are available.\n\t\t\t\t\tIf the maximum is 100%, then replacement tasks can't start until the draining\n\t\t\t\t\ttasks have stopped.
Any PENDING
or RUNNING
tasks that do not belong to a service\n\t\t\taren't affected. You must wait for them to finish or stop them manually.
A container instance has completed draining when it has no more RUNNING
\n\t\t\ttasks. You can verify this using ListTasks.
When a container instance has been drained, you can set a container instance to\n\t\t\t\tACTIVE
status and once it has reached that status the Amazon ECS scheduler\n\t\t\tcan begin scheduling tasks on the instance again.
Modifies the parameters of a service.
\nOn March 21, 2024, a change was made to resolve the task definition revision before authorization. When a task definition revision is not specified, authorization will occur using the latest revision of a task definition.
\nFor services using the rolling update (ECS
) you can update the desired\n\t\t\tcount, deployment configuration, network configuration, load balancers, service\n\t\t\tregistries, enable ECS managed tags option, propagate tags option, task placement\n\t\t\tconstraints and strategies, and task definition. When you update any of these\n\t\t\tparameters, Amazon ECS starts new tasks with the new configuration.
You can attach Amazon EBS volumes to Amazon ECS tasks by configuring the volume when starting or\n\t\t\trunning a task, or when creating or updating a service. For more infomation, see Amazon EBS volumes in the Amazon Elastic Container Service Developer Guide. You can update\n\t\t\tyour volume configurations and trigger a new deployment.\n\t\t\t\tvolumeConfigurations
is only supported for REPLICA service and not\n\t\t\tDAEMON service. If you leave volumeConfigurations
\n null
, it doesn't trigger a new deployment. For more infomation on volumes,\n\t\t\tsee Amazon EBS volumes in the Amazon Elastic Container Service Developer Guide.
For services using the blue/green (CODE_DEPLOY
) deployment controller,\n\t\t\tonly the desired count, deployment configuration, health check grace period, task\n\t\t\tplacement constraints and strategies, enable ECS managed tags option, and propagate tags\n\t\t\tcan be updated using this API. If the network configuration, platform version, task\n\t\t\tdefinition, or load balancer need to be updated, create a new CodeDeploy deployment. For more\n\t\t\tinformation, see CreateDeployment in the CodeDeploy API Reference.
For services using an external deployment controller, you can update only the desired\n\t\t\tcount, task placement constraints and strategies, health check grace period, enable ECS\n\t\t\tmanaged tags option, and propagate tags option, using this API. If the launch type, load\n\t\t\tbalancer, network configuration, platform version, or task definition need to be\n\t\t\tupdated, create a new task set For more information, see CreateTaskSet.
\nYou can add to or subtract from the number of instantiations of a task definition in a\n\t\t\tservice by specifying the cluster that the service is running in and a new\n\t\t\t\tdesiredCount
parameter.
You can attach Amazon EBS volumes to Amazon ECS tasks by configuring the volume when starting or\n\t\t\trunning a task, or when creating or updating a service. For more infomation, see Amazon EBS volumes in the Amazon Elastic Container Service Developer Guide.
\nIf you have updated the container image of your application, you can create a new task\n\t\t\tdefinition with that image and deploy it to your service. The service scheduler uses the\n\t\t\tminimum healthy percent and maximum percent parameters (in the service's deployment\n\t\t\tconfiguration) to determine the deployment strategy.
\nIf your updated Docker image uses the same tag as what is in the existing task\n\t\t\t\tdefinition for your service (for example, my_image:latest
), you don't\n\t\t\t\tneed to create a new revision of your task definition. You can update the service\n\t\t\t\tusing the forceNewDeployment
option. The new tasks launched by the\n\t\t\t\tdeployment pull the current image/tag combination from your repository when they\n\t\t\t\tstart.
You can also update the deployment configuration of a service. When a deployment is\n\t\t\ttriggered by updating the task definition of a service, the service scheduler uses the\n\t\t\tdeployment configuration parameters, minimumHealthyPercent
and\n\t\t\t\tmaximumPercent
, to determine the deployment strategy.
If minimumHealthyPercent
is below 100%, the scheduler can ignore\n\t\t\t\t\t\tdesiredCount
temporarily during a deployment. For example, if\n\t\t\t\t\t\tdesiredCount
is four tasks, a minimum of 50% allows the\n\t\t\t\t\tscheduler to stop two existing tasks before starting two new tasks. Tasks for\n\t\t\t\t\tservices that don't use a load balancer are considered healthy if they're in the\n\t\t\t\t\t\tRUNNING
state. Tasks for services that use a load balancer are\n\t\t\t\t\tconsidered healthy if they're in the RUNNING
state and are reported\n\t\t\t\t\tas healthy by the load balancer.
The maximumPercent
parameter represents an upper limit on the\n\t\t\t\t\tnumber of running tasks during a deployment. You can use it to define the\n\t\t\t\t\tdeployment batch size. For example, if desiredCount
is four tasks,\n\t\t\t\t\ta maximum of 200% starts four new tasks before stopping the four older tasks\n\t\t\t\t\t(provided that the cluster resources required to do this are available).
When UpdateService stops a task during a deployment, the equivalent\n\t\t\tof docker stop
is issued to the containers running in the task. This\n\t\t\tresults in a SIGTERM
and a 30-second timeout. After this,\n\t\t\t\tSIGKILL
is sent and the containers are forcibly stopped. If the\n\t\t\tcontainer handles the SIGTERM
gracefully and exits within 30 seconds from\n\t\t\treceiving it, no SIGKILL
is sent.
When the service scheduler launches new tasks, it determines task placement in your\n\t\t\tcluster with the following logic.
\nDetermine which of the container instances in your cluster can support your\n\t\t\t\t\tservice's task definition. For example, they have the required CPU, memory,\n\t\t\t\t\tports, and container instance attributes.
\nBy default, the service scheduler attempts to balance tasks across\n\t\t\t\t\tAvailability Zones in this manner even though you can choose a different\n\t\t\t\t\tplacement strategy.
\nSort the valid container instances by the fewest number of running\n\t\t\t\t\t\t\ttasks for this service in the same Availability Zone as the instance.\n\t\t\t\t\t\t\tFor example, if zone A has one running service task and zones B and C\n\t\t\t\t\t\t\teach have zero, valid container instances in either zone B or C are\n\t\t\t\t\t\t\tconsidered optimal for placement.
\nPlace the new service task on a valid container instance in an optimal\n\t\t\t\t\t\t\tAvailability Zone (based on the previous steps), favoring container\n\t\t\t\t\t\t\tinstances with the fewest number of running tasks for this\n\t\t\t\t\t\t\tservice.
\nWhen the service scheduler stops running tasks, it attempts to maintain balance across\n\t\t\tthe Availability Zones in your cluster using the following logic:
\nSort the container instances by the largest number of running tasks for this\n\t\t\t\t\tservice in the same Availability Zone as the instance. For example, if zone A\n\t\t\t\t\thas one running service task and zones B and C each have two, container\n\t\t\t\t\tinstances in either zone B or C are considered optimal for termination.
\nStop the task on a container instance in an optimal Availability Zone (based\n\t\t\t\t\ton the previous steps), favoring container instances with the largest number of\n\t\t\t\t\trunning tasks for this service.
\nYou must have a service-linked role when you update any of the following service\n\t\t\t\tproperties:
\n\n loadBalancers
,
\n serviceRegistries
\n
For more information about the role see the CreateService
request\n\t\t\t\tparameter \n role
\n .
Modifies the parameters of a service.
\nOn March 21, 2024, a change was made to resolve the task definition revision before authorization. When a task definition revision is not specified, authorization will occur using the latest revision of a task definition.
\nFor services using the rolling update (ECS
) you can update the desired\n\t\t\tcount, deployment configuration, network configuration, load balancers, service\n\t\t\tregistries, enable ECS managed tags option, propagate tags option, task placement\n\t\t\tconstraints and strategies, and task definition. When you update any of these\n\t\t\tparameters, Amazon ECS starts new tasks with the new configuration.
You can attach Amazon EBS volumes to Amazon ECS tasks by configuring the volume when starting or\n\t\t\trunning a task, or when creating or updating a service. For more infomation, see Amazon EBS volumes in the Amazon Elastic Container Service Developer Guide. You can update\n\t\t\tyour volume configurations and trigger a new deployment.\n\t\t\t\tvolumeConfigurations
is only supported for REPLICA service and not\n\t\t\tDAEMON service. If you leave volumeConfigurations
\n null
, it doesn't trigger a new deployment. For more infomation on volumes,\n\t\t\tsee Amazon EBS volumes in the Amazon Elastic Container Service Developer Guide.
For services using the blue/green (CODE_DEPLOY
) deployment controller,\n\t\t\tonly the desired count, deployment configuration, health check grace period, task\n\t\t\tplacement constraints and strategies, enable ECS managed tags option, and propagate tags\n\t\t\tcan be updated using this API. If the network configuration, platform version, task\n\t\t\tdefinition, or load balancer need to be updated, create a new CodeDeploy deployment. For more\n\t\t\tinformation, see CreateDeployment in the CodeDeploy API Reference.
For services using an external deployment controller, you can update only the desired\n\t\t\tcount, task placement constraints and strategies, health check grace period, enable ECS\n\t\t\tmanaged tags option, and propagate tags option, using this API. If the launch type, load\n\t\t\tbalancer, network configuration, platform version, or task definition need to be\n\t\t\tupdated, create a new task set For more information, see CreateTaskSet.
\nYou can add to or subtract from the number of instantiations of a task definition in a\n\t\t\tservice by specifying the cluster that the service is running in and a new\n\t\t\t\tdesiredCount
parameter.
You can attach Amazon EBS volumes to Amazon ECS tasks by configuring the volume when starting or\n\t\t\trunning a task, or when creating or updating a service. For more infomation, see Amazon EBS volumes in the Amazon Elastic Container Service Developer Guide.
\nIf you have updated the container image of your application, you can create a new task\n\t\t\tdefinition with that image and deploy it to your service. The service scheduler uses the\n\t\t\tminimum healthy percent and maximum percent parameters (in the service's deployment\n\t\t\tconfiguration) to determine the deployment strategy.
\nIf your updated Docker image uses the same tag as what is in the existing task\n\t\t\t\tdefinition for your service (for example, my_image:latest
), you don't\n\t\t\t\tneed to create a new revision of your task definition. You can update the service\n\t\t\t\tusing the forceNewDeployment
option. The new tasks launched by the\n\t\t\t\tdeployment pull the current image/tag combination from your repository when they\n\t\t\t\tstart.
You can also update the deployment configuration of a service. When a deployment is\n\t\t\ttriggered by updating the task definition of a service, the service scheduler uses the\n\t\t\tdeployment configuration parameters, minimumHealthyPercent
and\n\t\t\t\tmaximumPercent
, to determine the deployment strategy.
If minimumHealthyPercent
is below 100%, the scheduler can ignore\n\t\t\t\t\t\tdesiredCount
temporarily during a deployment. For example, if\n\t\t\t\t\t\tdesiredCount
is four tasks, a minimum of 50% allows the\n\t\t\t\t\tscheduler to stop two existing tasks before starting two new tasks. Tasks for\n\t\t\t\t\tservices that don't use a load balancer are considered healthy if they're in the\n\t\t\t\t\t\tRUNNING
state. Tasks for services that use a load balancer are\n\t\t\t\t\tconsidered healthy if they're in the RUNNING
state and are reported\n\t\t\t\t\tas healthy by the load balancer.
The maximumPercent
parameter represents an upper limit on the\n\t\t\t\t\tnumber of running tasks during a deployment. You can use it to define the\n\t\t\t\t\tdeployment batch size. For example, if desiredCount
is four tasks,\n\t\t\t\t\ta maximum of 200% starts four new tasks before stopping the four older tasks\n\t\t\t\t\t(provided that the cluster resources required to do this are available).
When UpdateService stops a task during a deployment, the equivalent\n\t\t\tof docker stop
is issued to the containers running in the task. This\n\t\t\tresults in a SIGTERM
and a 30-second timeout. After this,\n\t\t\t\tSIGKILL
is sent and the containers are forcibly stopped. If the\n\t\t\tcontainer handles the SIGTERM
gracefully and exits within 30 seconds from\n\t\t\treceiving it, no SIGKILL
is sent.
When the service scheduler launches new tasks, it determines task placement in your\n\t\t\tcluster with the following logic.
\nDetermine which of the container instances in your cluster can support your\n\t\t\t\t\tservice's task definition. For example, they have the required CPU, memory,\n\t\t\t\t\tports, and container instance attributes.
\nBy default, the service scheduler attempts to balance tasks across\n\t\t\t\t\tAvailability Zones in this manner even though you can choose a different\n\t\t\t\t\tplacement strategy.
\nSort the valid container instances by the fewest number of running\n\t\t\t\t\t\t\ttasks for this service in the same Availability Zone as the instance.\n\t\t\t\t\t\t\tFor example, if zone A has one running service task and zones B and C\n\t\t\t\t\t\t\teach have zero, valid container instances in either zone B or C are\n\t\t\t\t\t\t\tconsidered optimal for placement.
\nPlace the new service task on a valid container instance in an optimal\n\t\t\t\t\t\t\tAvailability Zone (based on the previous steps), favoring container\n\t\t\t\t\t\t\tinstances with the fewest number of running tasks for this\n\t\t\t\t\t\t\tservice.
\nWhen the service scheduler stops running tasks, it attempts to maintain balance across\n\t\t\tthe Availability Zones in your cluster using the following logic:
\nSort the container instances by the largest number of running tasks for this\n\t\t\t\t\tservice in the same Availability Zone as the instance. For example, if zone A\n\t\t\t\t\thas one running service task and zones B and C each have two, container\n\t\t\t\t\tinstances in either zone B or C are considered optimal for termination.
\nStop the task on a container instance in an optimal Availability Zone (based\n\t\t\t\t\ton the previous steps), favoring container instances with the largest number of\n\t\t\t\t\trunning tasks for this service.
\nYou must have a service-linked role when you update any of the following service\n\t\t\t\tproperties:
\n\n loadBalancers
,
\n serviceRegistries
\n
For more information about the role see the CreateService
request\n\t\t\t\tparameter \n role
\n .
The capacity provider strategy to update the service to use.
\nif the service uses the default capacity provider strategy for the cluster, the\n\t\t\tservice can be updated to use one or more capacity providers as opposed to the default\n\t\t\tcapacity provider strategy. However, when a service is using a capacity provider\n\t\t\tstrategy that's not the default capacity provider strategy, the service can't be updated\n\t\t\tto use the cluster's default capacity provider strategy.
\nA capacity provider strategy consists of one or more capacity providers along with the\n\t\t\t\tbase
and weight
to assign to them. A capacity provider\n\t\t\tmust be associated with the cluster to be used in a capacity provider strategy. The\n\t\t\t\tPutClusterCapacityProviders API is used to associate a capacity\n\t\t\tprovider with a cluster. Only capacity providers with an ACTIVE
or\n\t\t\t\tUPDATING
status can be used.
If specifying a capacity provider that uses an Auto Scaling group, the capacity\n\t\t\tprovider must already be created. New capacity providers can be created with the CreateCapacityProvider API operation.
\nTo use a Fargate capacity provider, specify either the FARGATE
or\n\t\t\t\tFARGATE_SPOT
capacity providers. The Fargate capacity providers are\n\t\t\tavailable to all accounts and only need to be associated with a cluster to be\n\t\t\tused.
The PutClusterCapacityProviders API operation is used to update the\n\t\t\tlist of available capacity providers for a cluster after the cluster is created.
\n " + "smithy.api#documentation": "The capacity provider strategy to update the service to use.
\nif the service uses the default capacity provider strategy for the cluster, the\n\t\t\tservice can be updated to use one or more capacity providers as opposed to the default\n\t\t\tcapacity provider strategy. However, when a service is using a capacity provider\n\t\t\tstrategy that's not the default capacity provider strategy, the service can't be updated\n\t\t\tto use the cluster's default capacity provider strategy.
\nA capacity provider strategy consists of one or more capacity providers along with the\n\t\t\t\tbase
and weight
to assign to them. A capacity provider\n\t\t\tmust be associated with the cluster to be used in a capacity provider strategy. The\n\t\t\tPutClusterCapacityProviders API is used to associate a capacity\n\t\t\tprovider with a cluster. Only capacity providers with an ACTIVE
or\n\t\t\t\tUPDATING
status can be used.
If specifying a capacity provider that uses an Auto Scaling group, the capacity\n\t\t\tprovider must already be created. New capacity providers can be created with the CreateClusterCapacityProvider API operation.
\nTo use a Fargate capacity provider, specify either the FARGATE
or\n\t\t\t\tFARGATE_SPOT
capacity providers. The Fargate capacity providers are\n\t\t\tavailable to all accounts and only need to be associated with a cluster to be\n\t\t\tused.
The PutClusterCapacityProvidersAPI operation is used to update the\n\t\t\tlist of available capacity providers for a cluster after the cluster is created.
\n " } }, "deploymentConfiguration": {