diff --git a/packages/@aws-cdk/cfnspec/spec-source/cfn-docs/cfn-docs.json b/packages/@aws-cdk/cfnspec/spec-source/cfn-docs/cfn-docs.json index 4eeef575142a5..38f776623f7b8 100644 --- a/packages/@aws-cdk/cfnspec/spec-source/cfn-docs/cfn-docs.json +++ b/packages/@aws-cdk/cfnspec/spec-source/cfn-docs/cfn-docs.json @@ -1523,9 +1523,9 @@ }, "description": "The `AWS::ApiGateway::UsagePlanKey` resource associates an API key with a usage plan. This association determines which users the usage plan is applied to.", "properties": { - "KeyId": "The Id of the UsagePlanKey resource to be deleted.", + "KeyId": "The Id of the UsagePlanKey resource.", "KeyType": "The type of a UsagePlanKey resource for a plan customer.", - "UsagePlanId": "The Id of the UsagePlan resource representing the usage plan containing the to-be-deleted UsagePlanKey resource representing a plan customer." + "UsagePlanId": "The Id of the UsagePlan resource representing the usage plan containing the UsagePlanKey resource representing a plan customer." } }, "AWS::ApiGateway::VpcLink": { @@ -4548,7 +4548,7 @@ "DisplayName": "The display name of the app block.", "Name": "The name of the app block.\n\n*Pattern* : `^[a-zA-Z0-9][a-zA-Z0-9_.-]{0,100}$`", "PackagingType": "The packaging type of the app block.", - "PostSetupScriptDetails": "The post setup script details of the app block.\n\nThis only applies to app blocks with PackagingType `APPSTREAM2` .", + "PostSetupScriptDetails": "The post setup script details of the app block.", "SetupScriptDetails": "The setup script details of the app block.", "SourceS3Location": "The source S3 location of the app block.", "Tags": "The tags of the app block." @@ -4574,22 +4574,22 @@ }, "AWS::AppStream::AppBlockBuilder": { "attributes": { - "Arn": "", - "CreatedTime": "", - "Ref": "" + "Arn": "The ARN of the app block builder.", + "CreatedTime": "The time when the app block builder was created.", + "Ref": "When you pass the logical ID of this resource to the intrinsic `Ref` function, `Ref` returns the `Name` of the app block builder, such as `abcdefAppBlockBuilder` .\n\nFor more information about using the `Ref` function, see [Ref](https://docs.aws.amazon.com/AWSCloudFormation/latest/UserGuide/intrinsic-function-reference-ref.html) ." }, "description": "Creates an app block builder.", "properties": { - "AccessEndpoints": "", - "AppBlockArns": "", + "AccessEndpoints": "The access endpoints of the app block builder.", + "AppBlockArns": "The ARN of the app block.\n\n*Maximum* : `1`", "Description": "The description of the app block builder.", "DisplayName": "The display name of the app block builder.", "EnableDefaultInternetAccess": "Indicates whether default internet access is enabled for the app block builder.", "IamRoleArn": "The ARN of the IAM role that is applied to the app block builder.", "InstanceType": "The instance type of the app block builder.", "Name": "The name of the app block builder.", - "Platform": "The platform of the app block builder.\n\n`WINDOWS_SERVER_2019` is the only valid value.", - "Tags": "", + "Platform": "The platform of the app block builder.\n\n*Allowed values* : `WINDOWS_SERVER_2019`", + "Tags": "The tags of the app block builder.", "VpcConfig": "The VPC configuration for the app block builder." } }, @@ -10863,40 +10863,40 @@ }, "AWS::Comprehend::DocumentClassifier": { "attributes": { - "Arn": "", - "Ref": "" + "Arn": "The Amazon Resource Name (ARN) of the document classifier.", + "Ref": "`Ref` returns the Amazon Resource Name (ARN) of the document classifier." }, - "description": "", + "description": "This resource creates and trains a document classifier to categorize documents. You provide a set of training documents that are labeled with the categories that you want to identify. After the classifier is trained you can use it to categorize a set of labeled documents into the categories. For more information, see [Document Classification](https://docs.aws.amazon.com/comprehend/latest/dg/how-document-classification.html) in the Comprehend Developer Guide.", "properties": { - "DataAccessRoleArn": "", - "DocumentClassifierName": "", - "InputDataConfig": "", - "LanguageCode": "", - "Mode": "", - "ModelKmsKeyId": "", - "ModelPolicy": "", - "OutputDataConfig": "", - "Tags": "", - "VersionName": "", - "VolumeKmsKeyId": "", - "VpcConfig": "" + "DataAccessRoleArn": "The Amazon Resource Name (ARN) of the IAM role that grants Amazon Comprehend read access to your input data.", + "DocumentClassifierName": "The name of the document classifier.", + "InputDataConfig": "Specifies the format and location of the input data for the job.", + "LanguageCode": "The language of the input documents. You can specify any of the languages supported by Amazon Comprehend. All documents must be in the same language.", + "Mode": "Indicates the mode in which the classifier will be trained. The classifier can be trained in multi-class mode, which identifies one and only one class for each document, or multi-label mode, which identifies one or more labels for each document. In multi-label mode, multiple labels for an individual document are separated by a delimiter. The default delimiter between labels is a pipe (|).", + "ModelKmsKeyId": "ID for the AWS KMS key that Amazon Comprehend uses to encrypt trained custom models. The ModelKmsKeyId can be either of the following formats:\n\n- KMS Key ID: `\"1234abcd-12ab-34cd-56ef-1234567890ab\"`\n- Amazon Resource Name (ARN) of a KMS Key: `\"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"`", + "ModelPolicy": "The resource-based policy to attach to your custom document classifier model. You can use this policy to allow another AWS account to import your custom model.\n\nProvide your policy as a JSON body that you enter as a UTF-8 encoded string without line breaks. To provide valid JSON, enclose the attribute names and values in double quotes. If the JSON body is also enclosed in double quotes, then you must escape the double quotes that are inside the policy:\n\n`\"{\\\"attribute\\\": \\\"value\\\", \\\"attribute\\\": [\\\"value\\\"]}\"`\n\nTo avoid escaping quotes, you can use single quotes to enclose the policy and double quotes to enclose the JSON names and values:\n\n`'{\"attribute\": \"value\", \"attribute\": [\"value\"]}'`", + "OutputDataConfig": "Provides output results configuration parameters for custom classifier jobs.", + "Tags": "Tags to associate with the document classifier. A tag is a key-value pair that adds as a metadata to a resource used by Amazon Comprehend. For example, a tag with \"Sales\" as the key might be added to a resource to indicate its use by the sales department.", + "VersionName": "The version name given to the newly created classifier. Version names can have a maximum of 256 characters. Alphanumeric characters, hyphens (-) and underscores (_) are allowed. The version name must be unique among all models with the same classifier name in the AWS account / AWS Region .", + "VolumeKmsKeyId": "ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats:\n\n- KMS Key ID: `\"1234abcd-12ab-34cd-56ef-1234567890ab\"`\n- Amazon Resource Name (ARN) of a KMS Key: `\"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"`", + "VpcConfig": "Configuration parameters for a private Virtual Private Cloud (VPC) containing the resources you are using for your custom classifier. For more information, see [Amazon VPC](https://docs.aws.amazon.com/vpc/latest/userguide/what-is-amazon-vpc.html) ." } }, "AWS::Comprehend::DocumentClassifier.AugmentedManifestsListItem": { "attributes": {}, - "description": "", + "description": "An augmented manifest file that provides training data for your custom model. An augmented manifest file is a labeled dataset that is produced by Amazon SageMaker Ground Truth.", "properties": { - "AttributeNames": "", - "S3Uri": "", - "Split": "" + "AttributeNames": "The JSON attribute that contains the annotations for your training documents. The number of attribute names that you specify depends on whether your augmented manifest file is the output of a single labeling job or a chained labeling job.\n\nIf your file is the output of a single labeling job, specify the LabelAttributeName key that was used when the job was created in Ground Truth.\n\nIf your file is the output of a chained labeling job, specify the LabelAttributeName key for one or more jobs in the chain. Each LabelAttributeName key provides the annotations from an individual job.", + "S3Uri": "The Amazon S3 location of the augmented manifest file.", + "Split": "The purpose of the data you've provided in the augmented manifest. You can either train or test this data. If you don't specify, the default is train.\n\nTRAIN - all of the documents in the manifest will be used for training. If no test documents are provided, Amazon Comprehend will automatically reserve a portion of the training documents for testing.\n\nTEST - all of the documents in the manifest will be used for testing." } }, "AWS::Comprehend::DocumentClassifier.DocumentClassifierDocuments": { "attributes": {}, - "description": "", + "description": "The location of the training documents. This parameter is required in a request to create a native document model.", "properties": { - "S3Uri": "", - "TestS3Uri": "" + "S3Uri": "The S3 URI location of the training documents specified in the S3Uri CSV file.", + "TestS3Uri": "The S3 URI location of the test documents included in the TestS3Uri CSV file. This field is not required if you do not specify a test CSV file." } }, "AWS::Comprehend::DocumentClassifier.DocumentClassifierInputDataConfig": { @@ -10906,8 +10906,8 @@ "AugmentedManifests": "A list of augmented manifest files that provide training data for your custom model. An augmented manifest file is a labeled dataset that is produced by Amazon SageMaker Ground Truth.\n\nThis parameter is required if you set `DataFormat` to `AUGMENTED_MANIFEST` .", "DataFormat": "The format of your training data:\n\n- `COMPREHEND_CSV` : A two-column CSV file, where labels are provided in the first column, and documents are provided in the second. If you use this value, you must provide the `S3Uri` parameter in your request.\n- `AUGMENTED_MANIFEST` : A labeled dataset that is produced by Amazon SageMaker Ground Truth. This file is in JSON lines format. Each line is a complete JSON object that contains a training document and its associated labels.\n\nIf you use this value, you must provide the `AugmentedManifests` parameter in your request.\n\nIf you don't specify a value, Amazon Comprehend uses `COMPREHEND_CSV` as the default.", "DocumentReaderConfig": "", - "DocumentType": "", - "Documents": "", + "DocumentType": "The type of input documents for training the model. Provide plain-text documents to create a plain-text model, and provide semi-structured documents to create a native document model.", + "Documents": "The S3 location of the training documents. This parameter is required in a request to create a native document model.", "LabelDelimiter": "Indicates the delimiter used to separate each label for training a multi-label classifier. The default delimiter between labels is a pipe (|). You can use a different character as a delimiter (if it's an allowed character) by specifying it under Delimiter for labels. If the training documents use a delimiter other than the default or the delimiter you specify, the labels on that line will be combined to make a single unique label, such as LABELLABELLABEL.", "S3Uri": "The Amazon S3 URI for the input data. The S3 bucket must be in the same Region as the API endpoint that you are calling. The URI can point to a single input file or it can provide the prefix for a collection of input files.\n\nFor example, if you use the URI `S3://bucketName/prefix` , if the prefix is a single file, Amazon Comprehend uses that file as input. If more than one file begins with the prefix, Amazon Comprehend uses all of them as input.\n\nThis parameter is required if you set `DataFormat` to `COMPREHEND_CSV` .", "TestS3Uri": "This specifies the Amazon S3 location where the test annotations for an entity recognizer are located. The URI must be in the same AWS Region as the API endpoint that you are calling." @@ -10915,19 +10915,19 @@ }, "AWS::Comprehend::DocumentClassifier.DocumentClassifierOutputDataConfig": { "attributes": {}, - "description": "", + "description": "Provide the location for output data from a custom classifier job. This field is mandatory if you are training a native document model.", "properties": { - "KmsKeyId": "", - "S3Uri": "" + "KmsKeyId": "ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt the output results from an analysis job. The KmsKeyId can be one of the following formats:\n\n- KMS Key ID: `\"1234abcd-12ab-34cd-56ef-1234567890ab\"`\n- Amazon Resource Name (ARN) of a KMS Key: `\"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"`\n- KMS Key Alias: `\"alias/ExampleAlias\"`\n- ARN of a KMS Key Alias: `\"arn:aws:kms:us-west-2:111122223333:alias/ExampleAlias\"`", + "S3Uri": "When you use the `OutputDataConfig` object while creating a custom classifier, you specify the Amazon S3 location where you want to write the confusion matrix and other output files. The URI must be in the same Region as the API endpoint that you are calling. The location is used as the prefix for the actual location of this output file.\n\nWhen the custom classifier job is finished, the service creates the output file in a directory specific to the job. The `S3Uri` field contains the location of the output file, called `output.tar.gz` . It is a compressed archive that contains the confusion matrix." } }, "AWS::Comprehend::DocumentClassifier.DocumentReaderConfig": { "attributes": {}, - "description": "", + "description": "Provides configuration parameters to override the default actions for extracting text from PDF documents and image files.\n\nBy default, Amazon Comprehend performs the following actions to extract text from files, based on the input file type:\n\n- *Word files* - Amazon Comprehend parser extracts the text.\n- *Digital PDF files* - Amazon Comprehend parser extracts the text.\n- *Image files and scanned PDF files* - Amazon Comprehend uses the Amazon Textract `DetectDocumentText` API to extract the text.\n\n`DocumentReaderConfig` does not apply to plain text files or Word files.\n\nFor image files and PDF documents, you can override these default actions using the fields listed below. For more information, see [Setting text extraction options](https://docs.aws.amazon.com/comprehend/latest/dg/idp-set-textract-options.html) in the Comprehend Developer Guide.", "properties": { - "DocumentReadAction": "", - "DocumentReadMode": "", - "FeatureTypes": "" + "DocumentReadAction": "This field defines the Amazon Textract API operation that Amazon Comprehend uses to extract text from PDF files and image files. Enter one of the following values:\n\n- `TEXTRACT_DETECT_DOCUMENT_TEXT` - The Amazon Comprehend service uses the `DetectDocumentText` API operation.\n- `TEXTRACT_ANALYZE_DOCUMENT` - The Amazon Comprehend service uses the `AnalyzeDocument` API operation.", + "DocumentReadMode": "Determines the text extraction actions for PDF files. Enter one of the following values:\n\n- `SERVICE_DEFAULT` - use the Amazon Comprehend service defaults for PDF files.\n- `FORCE_DOCUMENT_READ_ACTION` - Amazon Comprehend uses the Textract API specified by DocumentReadAction for all PDF files, including digital PDF files.", + "FeatureTypes": "Specifies the type of Amazon Textract features to apply. If you chose `TEXTRACT_ANALYZE_DOCUMENT` as the read action, you must specify one or both of the following values:\n\n- `TABLES` - Returns information about any tables that are detected in the input document.\n- `FORMS` - Returns information and the data from any forms that are detected in the input document." } }, "AWS::Comprehend::DocumentClassifier.VpcConfig": { @@ -17312,7 +17312,7 @@ }, "description": "Attaches a VPC to a transit gateway.\n\nIf you attach a VPC with a CIDR range that overlaps the CIDR range of a VPC that is already attached, the new VPC CIDR range is not propagated to the default propagation route table.\n\nTo send VPC traffic to an attached transit gateway, add a route to the VPC route table using [AWS::EC2::Route](https://docs.aws.amazon.com/AWSCloudFormation/latest/UserGuide/aws-resource-ec2-route.html) .\n\nTo update tags for a VPC attachment after creation without replacing the attachment, use [AWS::EC2::TransitGatewayVpcAttachment](https://docs.aws.amazon.com/AWSCloudFormation/latest/UserGuide/aws-resource-ec2-transitgatewayvpcattachment.html) instead.", "properties": { - "Options": "The VPC attachment options, in JSON or YAML.\n\n- `ApplianceModeSupport` - Set to `enable` or `disable` . The default is `disable` .\n- `DnsSupport` - Set to `enable` or `disable` . The default is `enable` .\n- `Ipv6Support` - Set to `enable` or `disable` . The default is `disable` .", + "Options": "The VPC attachment options.", "SubnetIds": "The IDs of one or more subnets. You can specify only one subnet per Availability Zone. You must specify at least one subnet, but we recommend that you specify two subnets for better availability. The transit gateway uses one IP address from each specified subnet.", "Tags": "The tags for the attachment.", "TransitGatewayId": "The ID of the transit gateway.", @@ -17323,9 +17323,9 @@ "attributes": {}, "description": "Describes the VPC attachment options.", "properties": { - "ApplianceModeSupport": "Indicates whether appliance mode support is enabled.", - "DnsSupport": "Indicates whether DNS support is enabled.", - "Ipv6Support": "Indicates whether IPv6 support is disabled." + "ApplianceModeSupport": "Enable or disable appliance mode support. The default is `disable` .", + "DnsSupport": "Enable or disable DNS support. The default is `disable` .", + "Ipv6Support": "Enable or disable IPv6 support. The default is `disable` ." } }, "AWS::EC2::TransitGatewayConnect": { @@ -17503,7 +17503,7 @@ "description": "Specifies a VPC attachment.", "properties": { "AddSubnetIds": "The IDs of one or more subnets to add. You can specify at most one subnet per Availability Zone.", - "Options": "The VPC attachment options, in JSON or YAML.\n\n- `ApplianceModeSupport` - Set to `enable` or `disable` . The default is `disable` .\n- `DnsSupport` - Set to `enable` or `disable` . The default is `enable` .\n- `Ipv6Support` - Set to `enable` or `disable` . The default is `disable` .", + "Options": "The VPC attachment options.", "RemoveSubnetIds": "The IDs of one or more subnets to remove.", "SubnetIds": "The IDs of the subnets.", "Tags": "The tags for the VPC attachment.", @@ -17515,9 +17515,9 @@ "attributes": {}, "description": "Describes the VPC attachment options.", "properties": { - "ApplianceModeSupport": "Indicates whether appliance mode support is enabled.", - "DnsSupport": "Indicates whether DNS support is enabled.", - "Ipv6Support": "Indicates whether IPv6 support is disabled." + "ApplianceModeSupport": "Enable or disable appliance mode support. The default is `disable` .", + "DnsSupport": "Enable or disable DNS support. The default is `disable` .", + "Ipv6Support": "Enable or disable IPv6 support. The default is `disable` ." } }, "AWS::EC2::VPC": { @@ -50540,9 +50540,9 @@ }, "AWS::QuickSight::DataSet.DataSetRefreshProperties": { "attributes": {}, - "description": "", + "description": "The refresh properties of a dataset.", "properties": { - "RefreshConfiguration": "" + "RefreshConfiguration": "The refresh configuration for a dataset." } }, "AWS::QuickSight::DataSet.DataSetUsageConfiguration": { @@ -50578,7 +50578,7 @@ "attributes": {}, "description": "", "properties": { - "StaticValues": "" + "StaticValues": "A list of static default values for a given date time parameter. The valid format for this property is `yyyy-MM-dd\u2019T\u2019HH:mm:ss\u2019Z\u2019` ." } }, "AWS::QuickSight::DataSet.DecimalDatasetParameter": { @@ -50624,9 +50624,9 @@ }, "AWS::QuickSight::DataSet.IncrementalRefresh": { "attributes": {}, - "description": "", + "description": "The incremental refresh configuration for a dataset.", "properties": { - "LookbackWindow": "" + "LookbackWindow": "The lookback window setup for an incremental refresh configuration." } }, "AWS::QuickSight::DataSet.IngestionWaitPolicy": { @@ -50701,18 +50701,18 @@ }, "AWS::QuickSight::DataSet.LookbackWindow": { "attributes": {}, - "description": "", + "description": "The lookback window setup of an incremental refresh configuration.", "properties": { - "ColumnName": "", - "Size": "", - "SizeUnit": "" + "ColumnName": "The name of the lookback window column.", + "Size": "The lookback window column size.", + "SizeUnit": "The size unit that is used for the lookback window column. Valid values for this structure are `HOUR` , `DAY` , and `WEEK` ." } }, "AWS::QuickSight::DataSet.NewDefaultValues": { "attributes": {}, "description": "", "properties": { - "DateTimeStaticValues": "", + "DateTimeStaticValues": "A list of static default values for a given date time parameter. The valid format for this property is `yyyy-MM-dd\u2019T\u2019HH:mm:ss\u2019Z\u2019` .", "DecimalStaticValues": "", "IntegerStaticValues": "", "StringStaticValues": "" @@ -50729,10 +50729,10 @@ }, "AWS::QuickSight::DataSet.OverrideDatasetParameterOperation": { "attributes": {}, - "description": "", + "description": "A transform operation that overrides the dataset parameter values that are defined in another dataset.", "properties": { - "NewDefaultValues": "", - "NewParameterName": "", + "NewDefaultValues": "The new default values for the parameter.", + "NewParameterName": "The new name for the parameter.", "ParameterName": "" } }, @@ -50754,9 +50754,9 @@ }, "AWS::QuickSight::DataSet.RefreshConfiguration": { "attributes": {}, - "description": "", + "description": "The refresh configuration of a dataset.", "properties": { - "IncrementalRefresh": "" + "IncrementalRefresh": "The incremental refresh for the dataset." } }, "AWS::QuickSight::DataSet.RelationalTable": { @@ -55983,7 +55983,7 @@ "EnableHttpEndpoint": "A value that indicates whether to enable the HTTP endpoint for an Aurora Serverless DB cluster. By default, the HTTP endpoint is disabled.\n\nWhen enabled, the HTTP endpoint provides a connectionless web service API for running SQL queries on the Aurora Serverless DB cluster. You can also query your database from inside the RDS console with the query editor.\n\nFor more information, see [Using the Data API for Aurora Serverless](https://docs.aws.amazon.com/AmazonRDS/latest/AuroraUserGuide/data-api.html) in the *Amazon Aurora User Guide* .\n\nValid for: Aurora DB clusters only", "EnableIAMDatabaseAuthentication": "A value that indicates whether to enable mapping of AWS Identity and Access Management (IAM) accounts to database accounts. By default, mapping is disabled.\n\nFor more information, see [IAM Database Authentication](https://docs.aws.amazon.com/AmazonRDS/latest/AuroraUserGuide/UsingWithRDS.IAMDBAuth.html) in the *Amazon Aurora User Guide.*\n\nValid for: Aurora DB clusters only", "Engine": "The name of the database engine to be used for this DB cluster.\n\nValid Values:\n\n- `aurora-mysql`\n- `aurora-postgresql`\n- `mysql`\n- `postgres`\n\nValid for: Aurora DB clusters and Multi-AZ DB clusters", - "EngineMode": "The DB engine mode of the DB cluster, either `provisioned` or `serverless` .\n\nThe `serverless` engine mode only supports Aurora Serverless v1. Currently, AWS CloudFormation doesn't support Aurora Serverless v2.\n\nLimitations and requirements apply to some DB engine modes. For more information, see the following sections in the *Amazon Aurora User Guide* :\n\n- [Limitations of Aurora Serverless v1](https://docs.aws.amazon.com/AmazonRDS/latest/AuroraUserGuide/aurora-serverless.html#aurora-serverless.limitations)\n- [Requirements for Aurora Serverless v2](https://docs.aws.amazon.com/AmazonRDS/latest/AuroraUserGuide/aurora-serverless-v2.requirements.html)\n- [Limitations of parallel query](https://docs.aws.amazon.com/AmazonRDS/latest/AuroraUserGuide/aurora-mysql-parallel-query.html#aurora-mysql-parallel-query-limitations)\n- [Limitations of Aurora global databases](https://docs.aws.amazon.com/AmazonRDS/latest/AuroraUserGuide/aurora-global-database.html#aurora-global-database.limitations)\n\nValid for: Aurora DB clusters only", + "EngineMode": "The DB engine mode of the DB cluster, either `provisioned` or `serverless` .\n\nThe `serverless` engine mode only supports Aurora Serverless v1.\n\nLimitations and requirements apply to some DB engine modes. For more information, see the following sections in the *Amazon Aurora User Guide* :\n\n- [Limitations of Aurora Serverless v1](https://docs.aws.amazon.com/AmazonRDS/latest/AuroraUserGuide/aurora-serverless.html#aurora-serverless.limitations)\n- [Requirements for Aurora Serverless v2](https://docs.aws.amazon.com/AmazonRDS/latest/AuroraUserGuide/aurora-serverless-v2.requirements.html)\n- [Limitations of parallel query](https://docs.aws.amazon.com/AmazonRDS/latest/AuroraUserGuide/aurora-mysql-parallel-query.html#aurora-mysql-parallel-query-limitations)\n- [Limitations of Aurora global databases](https://docs.aws.amazon.com/AmazonRDS/latest/AuroraUserGuide/aurora-global-database.html#aurora-global-database.limitations)\n\nValid for: Aurora DB clusters only", "EngineVersion": "The version number of the database engine to use.\n\nTo list all of the available engine versions for Aurora MySQL version 2 (5.7-compatible) and version 3 (8.0-compatible), use the following command:\n\n`aws rds describe-db-engine-versions --engine aurora-mysql --query \"DBEngineVersions[].EngineVersion\"`\n\nYou can supply either `5.7` or `8.0` to use the default engine version for Aurora MySQL version 2 or version 3, respectively.\n\nTo list all of the available engine versions for Aurora PostgreSQL, use the following command:\n\n`aws rds describe-db-engine-versions --engine aurora-postgresql --query \"DBEngineVersions[].EngineVersion\"`\n\nTo list all of the available engine versions for RDS for MySQL, use the following command:\n\n`aws rds describe-db-engine-versions --engine mysql --query \"DBEngineVersions[].EngineVersion\"`\n\nTo list all of the available engine versions for RDS for PostgreSQL, use the following command:\n\n`aws rds describe-db-engine-versions --engine postgres --query \"DBEngineVersions[].EngineVersion\"`\n\n*Aurora MySQL*\n\nFor information, see [Database engine updates for Amazon Aurora MySQL](https://docs.aws.amazon.com/AmazonRDS/latest/AuroraUserGuide/AuroraMySQL.Updates.html) in the *Amazon Aurora User Guide* .\n\n*Aurora PostgreSQL*\n\nFor information, see [Amazon Aurora PostgreSQL releases and engine versions](https://docs.aws.amazon.com/AmazonRDS/latest/AuroraUserGuide/AuroraPostgreSQL.Updates.20180305.html) in the *Amazon Aurora User Guide* .\n\n*MySQL*\n\nFor information, see [Amazon RDS for MySQL](https://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/CHAP_MySQL.html#MySQL.Concepts.VersionMgmt) in the *Amazon RDS User Guide* .\n\n*PostgreSQL*\n\nFor information, see [Amazon RDS for PostgreSQL](https://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/CHAP_PostgreSQL.html#PostgreSQL.Concepts) in the *Amazon RDS User Guide* .\n\nValid for: Aurora DB clusters and Multi-AZ DB clusters", "GlobalClusterIdentifier": "If you are configuring an Aurora global database cluster and want your Aurora DB cluster to be a secondary member in the global database cluster, specify the global cluster ID of the global database cluster. To define the primary database cluster of the global cluster, use the [AWS::RDS::GlobalCluster](https://docs.aws.amazon.com/AWSCloudFormation/latest/UserGuide/aws-resource-rds-globalcluster.html) resource.\n\nIf you aren't configuring a global database cluster, don't specify this property.\n\n> To remove the DB cluster from a global database cluster, specify an empty value for the `GlobalClusterIdentifier` property. \n\nFor information about Aurora global databases, see [Working with Amazon Aurora Global Databases](https://docs.aws.amazon.com/AmazonRDS/latest/AuroraUserGuide/aurora-global-database.html) in the *Amazon Aurora User Guide* .\n\nValid for: Aurora DB clusters only", "Iops": "The amount of Provisioned IOPS (input/output operations per second) to be initially allocated for each DB instance in the Multi-AZ DB cluster.\n\nFor information about valid IOPS values, see [Provisioned IOPS storage](https://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/CHAP_Storage.html#USER_PIOPS) in the *Amazon RDS User Guide* .\n\nThis setting is required to create a Multi-AZ DB cluster.\n\nValid for Cluster Type: Multi-AZ DB clusters only\n\nConstraints:\n\n- Must be a multiple between .5 and 50 of the storage amount for the DB cluster.",