• Stars
    star
    135
  • Rank 259,781 (Top 6 %)
  • Language HCL
  • License
    Apache License 2.0
  • Created almost 7 years ago
  • Updated about 2 months ago

Reviews

There are no reviews yet. Be the first to send feedback to the community and the maintainers!

Repository Details

Terraform module to provision an RDS Aurora cluster for MySQL or Postgres

terraform-aws-rds-cluster Latest Release Slack Community

README Header

Cloud Posse

Terraform module to provision an RDS Aurora cluster for MySQL or Postgres.

Supports Amazon Aurora Serverless.


This project is part of our comprehensive "SweetOps" approach towards DevOps.

Terraform Open Source Modules

It's 100% Open Source and licensed under the APACHE2.

We literally have hundreds of terraform modules that are Open Source and well-maintained. Check them out!

Security & Compliance

Security scanning is graciously provided by Bridgecrew. Bridgecrew is the leading fully hosted, cloud-native solution providing continuous Terraform security and compliance.

Benchmark Description
Infrastructure Security Infrastructure Security Compliance
CIS KUBERNETES Center for Internet Security, KUBERNETES Compliance
CIS AWS Center for Internet Security, AWS Compliance
CIS AZURE Center for Internet Security, AZURE Compliance
PCI-DSS Payment Card Industry Data Security Standards Compliance
NIST-800-53 National Institute of Standards and Technology Compliance
ISO27001 Information Security Management System, ISO/IEC 27001 Compliance
SOC2 Service Organization Control 2 Compliance
CIS GCP Center for Internet Security, GCP Compliance
HIPAA Health Insurance Portability and Accountability Compliance

Usage

IMPORTANT: We do not pin modules to versions in our examples because of the difficulty of keeping the versions in the documentation in sync with the latest released versions. We highly recommend that in your code you pin the version to the exact version you are using so that your infrastructure remains stable, and update versions in a systematic way so that they do not catch you by surprise.

Also, because of a bug in the Terraform registry (hashicorp/terraform#21417), the registry shows many of our inputs as required when in fact they are optional. The table below correctly indicates which inputs are required.

For a complete example, see examples/complete.

For automated tests of the complete example using bats and Terratest (which tests and deploys the example on AWS), see test.

Basic example

module "rds_cluster_aurora_postgres" {
  source = "cloudposse/rds-cluster/aws"
  # Cloud Posse recommends pinning every module to a specific version
  # version     = "x.x.x"

  name            = "postgres"
  engine          = "aurora-postgresql"
  cluster_family  = "aurora-postgresql9.6"
  # 1 writer, 1 reader
  cluster_size    = 2
  # 1 writer, 3 reader
  # cluster_size    = 4
  # 1 writer, 5 reader
  # cluster_size    = 6
  namespace       = "eg"
  stage           = "dev"
  admin_user      = "admin1"
  admin_password  = "Test123456789"
  db_name         = "dbname"
  db_port         = 5432
  instance_type   = "db.r4.large"
  vpc_id          = "vpc-xxxxxxxx"
  security_groups = ["sg-xxxxxxxx"]
  subnets         = ["subnet-xxxxxxxx", "subnet-xxxxxxxx"]
  zone_id         = "Zxxxxxxxx"
}

Serverless Aurora MySQL 5.6

module "rds_cluster_aurora_mysql_serverless" {
  source = "cloudposse/rds-cluster/aws"
  # Cloud Posse recommends pinning every module to a specific version
  # version     = "x.x.x"
  namespace            = "eg"
  stage                = "dev"
  name                 = "db"
  engine               = "aurora"
  engine_mode          = "serverless"
  cluster_family       = "aurora5.6"
  cluster_size         = 0
  admin_user           = "admin1"
  admin_password       = "Test123456789"
  db_name              = "dbname"
  db_port              = 3306
  instance_type        = "db.t2.small"
  vpc_id               = "vpc-xxxxxxxx"
  security_groups      = ["sg-xxxxxxxx"]
  subnets              = ["subnet-xxxxxxxx", "subnet-xxxxxxxx"]
  zone_id              = "Zxxxxxxxx"
  enable_http_endpoint = true

  scaling_configuration = [
    {
      auto_pause               = true
      max_capacity             = 256
      min_capacity             = 2
      seconds_until_auto_pause = 300
    }
  ]
}

Serverless Aurora 2.07.1 MySQL 5.7

module "rds_cluster_aurora_mysql_serverless" {
  source = "cloudposse/rds-cluster/aws"
  # Cloud Posse recommends pinning every module to a specific version
  # version     = "x.x.x"
  namespace            = "eg"
  stage                = "dev"
  name                 = "db"
  engine               = "aurora-mysql"
  engine_mode          = "serverless"
  engine_version       = "5.7.mysql_aurora.2.07.1"
  cluster_family       = "aurora-mysql5.7"
  cluster_size         = 0
  admin_user           = "admin1"
  admin_password       = "Test123456789"
  db_name              = "dbname"
  db_port              = 3306
  vpc_id               = "vpc-xxxxxxxx"
  security_groups      = ["sg-xxxxxxxx"]
  subnets              = ["subnet-xxxxxxxx", "subnet-xxxxxxxx"]
  zone_id              = "Zxxxxxxxx"
  enable_http_endpoint = true

  scaling_configuration = [
    {
      auto_pause               = true
      max_capacity             = 16
      min_capacity             = 1
      seconds_until_auto_pause = 300
      timeout_action           = "ForceApplyCapacityChange"
    }
  ]
}

With cluster parameters

module "rds_cluster_aurora_mysql" {
  source = "cloudposse/rds-cluster/aws"
  # Cloud Posse recommends pinning every module to a specific version
  # version     = "x.x.x"
  engine          = "aurora"
  cluster_family  = "aurora-mysql5.7"
  cluster_size    = 2
  namespace       = "eg"
  stage           = "dev"
  name            = "db"
  admin_user      = "admin1"
  admin_password  = "Test123456789"
  db_name         = "dbname"
  instance_type   = "db.t2.small"
  vpc_id          = "vpc-xxxxxxx"
  security_groups = ["sg-xxxxxxxx"]
  subnets         = ["subnet-xxxxxxxx", "subnet-xxxxxxxx"]
  zone_id         = "Zxxxxxxxx"

  cluster_parameters = [
    {
      name  = "character_set_client"
      value = "utf8"
    },
    {
      name  = "character_set_connection"
      value = "utf8"
    },
    {
      name  = "character_set_database"
      value = "utf8"
    },
    {
      name  = "character_set_results"
      value = "utf8"
    },
    {
      name  = "character_set_server"
      value = "utf8"
    },
    {
      name  = "collation_connection"
      value = "utf8_bin"
    },
    {
      name  = "collation_server"
      value = "utf8_bin"
    },
    {
      name         = "lower_case_table_names"
      value        = "1"
      apply_method = "pending-reboot"
    },
    {
      name         = "skip-character-set-client-handshake"
      value        = "1"
      apply_method = "pending-reboot"
    }
  ]
}

With enhanced monitoring

# create IAM role for monitoring
resource "aws_iam_role" "enhanced_monitoring" {
  name               = "rds-cluster-example-1"
  assume_role_policy = data.aws_iam_policy_document.enhanced_monitoring.json
}

# Attach Amazon's managed policy for RDS enhanced monitoring
resource "aws_iam_role_policy_attachment" "enhanced_monitoring" {
  role       = aws_iam_role.enhanced_monitoring.name
  policy_arn = "arn:aws:iam::aws:policy/service-role/AmazonRDSEnhancedMonitoringRole"
}

# allow rds to assume this role
data "aws_iam_policy_document" "enhanced_monitoring" {
  statement {
      actions = [
      "sts:AssumeRole",
    ]

    effect = "Allow"

    principals {
      type        = "Service"
      identifiers = ["monitoring.rds.amazonaws.com"]
    }
  }
}

module "rds_cluster_aurora_postgres" {
  source = "cloudposse/rds-cluster/aws"
  # Cloud Posse recommends pinning every module to a specific version
  # version     = "x.x.x"
  engine          = "aurora-postgresql"
  cluster_family  = "aurora-postgresql9.6"
  cluster_size    = 2
  namespace       = "eg"
  stage           = "dev"
  name            = "db"
  admin_user      = "admin1"
  admin_password  = "Test123456789"
  db_name         = "dbname"
  db_port         = 5432
  instance_type   = "db.r4.large"
  vpc_id          = "vpc-xxxxxxx"
  security_groups = ["sg-xxxxxxxx"]
  subnets         = ["subnet-xxxxxxxx", "subnet-xxxxxxxx"]
  zone_id         = "Zxxxxxxxx"

  # enable monitoring every 30 seconds
  rds_monitoring_interval = 30

  # reference iam role created above
  rds_monitoring_role_arn = aws_iam_role.enhanced_monitoring.arn
}

Examples

Review the complete example to see how to use this module.

Makefile Targets

Available targets:

  help                                Help screen
  help/all                            Display help for all targets
  help/short                          This help short screen
  lint                                Lint terraform code

Requirements

Name Version
terraform >= 1.0.0
aws >= 4.23.0
null >= 2.0

Providers

Name Version
aws >= 4.23.0

Modules

Name Source Version
dns_master cloudposse/route53-cluster-hostname/aws 0.12.2
dns_replicas cloudposse/route53-cluster-hostname/aws 0.12.2
enhanced_monitoring_label cloudposse/label/null 0.25.0
this cloudposse/label/null 0.25.0

Resources

Name Type
aws_appautoscaling_policy.replicas resource
aws_appautoscaling_target.replicas resource
aws_db_parameter_group.default resource
aws_db_subnet_group.default resource
aws_iam_role.enhanced_monitoring resource
aws_iam_role_policy_attachment.enhanced_monitoring resource
aws_rds_cluster.primary resource
aws_rds_cluster.secondary resource
aws_rds_cluster_activity_stream.primary resource
aws_rds_cluster_instance.default resource
aws_rds_cluster_parameter_group.default resource
aws_security_group.default resource
aws_security_group_rule.egress resource
aws_security_group_rule.ingress_cidr_blocks resource
aws_security_group_rule.ingress_security_groups resource
aws_security_group_rule.traffic_inside_security_group resource
aws_iam_policy_document.enhanced_monitoring data source
aws_partition.current data source

Inputs

Name Description Type Default Required
activity_stream_enabled Whether to enable Activity Streams bool false no
activity_stream_kms_key_id The ARN for the KMS key to encrypt Activity Stream Data data. When specifying activity_stream_kms_key_id, activity_stream_enabled needs to be set to true string "" no
activity_stream_mode The mode for the Activity Streams. async and sync are supported. Defaults to async string "async" no
additional_tag_map Additional key-value pairs to add to each map in tags_as_list_of_maps. Not added to tags or id.
This is for some rare cases where resources want additional configuration of tags
and therefore take a list of maps with tag key, value, and additional configuration.
map(string) {} no
admin_password Password for the master DB user. Ignored if snapshot_identifier or replication_source_identifier is provided string "" no
admin_user Username for the master DB user. Ignored if snapshot_identifier or replication_source_identifier is provided string "admin" no
allocated_storage The allocated storage in GBs number null no
allow_major_version_upgrade Enable to allow major engine version upgrades when changing engine versions. Defaults to false. bool false no
allowed_cidr_blocks List of CIDR blocks allowed to access the cluster list(string) [] no
apply_immediately Specifies whether any cluster modifications are applied immediately, or during the next maintenance window bool true no
attributes ID element. Additional attributes (e.g. workers or cluster) to add to id,
in the order they appear in the list. New attributes are appended to the
end of the list. The elements of the list are joined by the delimiter
and treated as a single ID element.
list(string) [] no
auto_minor_version_upgrade Indicates that minor engine upgrades will be applied automatically to the DB instance during the maintenance window bool true no
autoscaling_enabled Whether to enable cluster autoscaling bool false no
autoscaling_max_capacity Maximum number of instances to be maintained by the autoscaler number 5 no
autoscaling_min_capacity Minimum number of instances to be maintained by the autoscaler number 1 no
autoscaling_policy_type Autoscaling policy type. TargetTrackingScaling and StepScaling are supported string "TargetTrackingScaling" no
autoscaling_scale_in_cooldown The amount of time, in seconds, after a scaling activity completes and before the next scaling down activity can start. Default is 300s number 300 no
autoscaling_scale_out_cooldown The amount of time, in seconds, after a scaling activity completes and before the next scaling up activity can start. Default is 300s number 300 no
autoscaling_target_metrics The metrics type to use. If this value isn't provided the default is CPU utilization string "RDSReaderAverageCPUUtilization" no
autoscaling_target_value The target value to scale with respect to target metrics number 75 no
backtrack_window The target backtrack window, in seconds. Only available for aurora engine currently. Must be between 0 and 259200 (72 hours) number 0 no
backup_window Daily time range during which the backups happen string "07:00-09:00" no
ca_cert_identifier The identifier of the CA certificate for the DB instance string null no
cluster_dns_name Name of the cluster CNAME record to create in the parent DNS zone specified by zone_id. If left empty, the name will be auto-asigned using the format master.var.name string "" no
cluster_family The family of the DB cluster parameter group string "aurora5.6" no
cluster_identifier The RDS Cluster Identifier. Will use generated label ID if not supplied string "" no
cluster_parameters List of DB cluster parameters to apply
list(object({
apply_method = string
name = string
value = string
}))
[] no
cluster_size Number of DB instances to create in the cluster number 2 no
cluster_type Either regional or global.
If regional will be created as a normal, standalone DB.
If global, will be made part of a Global cluster (requires global_cluster_identifier).
string "regional" no
context Single object for setting entire context at once.
See description of individual variables for details.
Leave string and numeric variables as null to use default value.
Individual variable settings (non-null) override settings in context object,
except for attributes, tags, and additional_tag_map, which are merged.
any
{
"additional_tag_map": {},
"attributes": [],
"delimiter": null,
"descriptor_formats": {},
"enabled": true,
"environment": null,
"id_length_limit": null,
"label_key_case": null,
"label_order": [],
"label_value_case": null,
"labels_as_tags": [
"unset"
],
"name": null,
"namespace": null,
"regex_replace_chars": null,
"stage": null,
"tags": {},
"tenant": null
}
no
copy_tags_to_snapshot Copy tags to backup snapshots bool false no
db_cluster_instance_class This setting is required to create a provisioned Multi-AZ DB cluster string null no
db_name Database name (default is not to create a database) string "" no
db_port Database port number 3306 no
deletion_protection If the DB instance should have deletion protection enabled bool false no
delimiter Delimiter to be used between ID elements.
Defaults to - (hyphen). Set to "" to use no delimiter at all.
string null no
descriptor_formats Describe additional descriptors to be output in the descriptors output map.
Map of maps. Keys are names of descriptors. Values are maps of the form
{<br> format = string<br> labels = list(string)<br>}
(Type is any so the map values can later be enhanced to provide additional options.)
format is a Terraform format string to be passed to the format() function.
labels is a list of labels, in order, to pass to format() function.
Label values will be normalized before being passed to format() so they will be
identical to how they appear in id.
Default is {} (descriptors output will be empty).
any {} no
egress_enabled Whether or not to apply the egress security group rule to default security group, defaults to true bool true no
enable_http_endpoint Enable HTTP endpoint (data API). Only valid when engine_mode is set to serverless bool false no
enabled Set to false to prevent the module from creating any resources bool null no
enabled_cloudwatch_logs_exports List of log types to export to cloudwatch. The following log types are supported: audit, error, general, slowquery list(string) [] no
engine The name of the database engine to be used for this DB cluster. Valid values: aurora, aurora-mysql, aurora-postgresql string "aurora" no
engine_mode The database engine mode. Valid values: parallelquery, provisioned, serverless string "provisioned" no
engine_version The version of the database engine to use. See aws rds describe-db-engine-versions string "" no
enhanced_monitoring_attributes The attributes for the enhanced monitoring IAM role list(string)
[
"enhanced-monitoring"
]
no
enhanced_monitoring_role_enabled A boolean flag to enable/disable the creation of the enhanced monitoring IAM role. If set to false, the module will not create a new role and will use rds_monitoring_role_arn for enhanced monitoring bool false no
environment ID element. Usually used for region e.g. 'uw2', 'us-west-2', OR role 'prod', 'staging', 'dev', 'UAT' string null no
global_cluster_identifier ID of the Aurora global cluster string "" no
iam_database_authentication_enabled Specifies whether or mappings of AWS Identity and Access Management (IAM) accounts to database accounts is enabled bool false no
iam_roles Iam roles for the Aurora cluster list(string) [] no
id_length_limit Limit id to this many characters (minimum 6).
Set to 0 for unlimited length.
Set to null for keep the existing setting, which defaults to 0.
Does not affect id_full.
number null no
instance_availability_zone Optional parameter to place cluster instances in a specific availability zone. If left empty, will place randomly string "" no
instance_parameters List of DB instance parameters to apply
list(object({
apply_method = string
name = string
value = string
}))
[] no
instance_type Instance type to use string "db.t2.small" no
intra_security_group_traffic_enabled Whether to allow traffic between resources inside the database's security group. bool false no
iops The amount of provisioned IOPS. Setting this implies a storage_type of 'io1'. This setting is required to create a Multi-AZ DB cluster. Check TF docs for values based on db engine number null no
kms_key_arn The ARN for the KMS encryption key. When specifying kms_key_arn, storage_encrypted needs to be set to true string "" no
label_key_case Controls the letter case of the tags keys (label names) for tags generated by this module.
Does not affect keys of tags passed in via the tags input.
Possible values: lower, title, upper.
Default value: title.
string null no
label_order The order in which the labels (ID elements) appear in the id.
Defaults to ["namespace", "environment", "stage", "name", "attributes"].
You can omit any of the 6 labels ("tenant" is the 6th), but at least one must be present.
list(string) null no
label_value_case Controls the letter case of ID elements (labels) as included in id,
set as tag values, and output by this module individually.
Does not affect values of tags passed in via the tags input.
Possible values: lower, title, upper and none (no transformation).
Set this to title and set delimiter to "" to yield Pascal Case IDs.
Default value: lower.
string null no
labels_as_tags Set of labels (ID elements) to include as tags in the tags output.
Default is to include all labels.
Tags with empty values will not be included in the tags output.
Set to [] to suppress all generated tags.
Notes:
The value of the name tag, if included, will be the id, not the name.
Unlike other null-label inputs, the initial setting of labels_as_tags cannot be
changed in later chained modules. Attempts to change it will be silently ignored.
set(string)
[
"default"
]
no
maintenance_window Weekly time range during which system maintenance can occur, in UTC string "wed:03:00-wed:04:00" no
name ID element. Usually the component or solution name, e.g. 'app' or 'jenkins'.
This is the only ID element not also included as a tag.
The "name" tag is set to the full id string. There is no tag with the value of the name input.
string null no
namespace ID element. Usually an abbreviation of your organization name, e.g. 'eg' or 'cp', to help ensure generated IDs are globally unique string null no
performance_insights_enabled Whether to enable Performance Insights bool false no
performance_insights_kms_key_id The ARN for the KMS key to encrypt Performance Insights data. When specifying performance_insights_kms_key_id, performance_insights_enabled needs to be set to true string "" no
performance_insights_retention_period Amount of time in days to retain Performance Insights data. Either 7 (7 days) or 731 (2 years) number null no
publicly_accessible Set to true if you want your cluster to be publicly accessible (such as via QuickSight) bool false no
rds_monitoring_interval The interval, in seconds, between points when enhanced monitoring metrics are collected for the DB instance. To disable collecting Enhanced Monitoring metrics, specify 0. The default is 0. Valid Values: 0, 1, 5, 10, 15, 30, 60 number 0 no
rds_monitoring_role_arn The ARN for the IAM role that permits RDS to send enhanced monitoring metrics to CloudWatch Logs string null no
reader_dns_name Name of the reader endpoint CNAME record to create in the parent DNS zone specified by zone_id. If left empty, the name will be auto-asigned using the format replicas.var.name string "" no
regex_replace_chars Terraform regular expression (regex) string.
Characters matching the regex will be removed from the ID elements.
If not set, "/[^a-zA-Z0-9-]/" is used to remove all characters other than hyphens, letters and digits.
string null no
replication_source_identifier ARN of a source DB cluster or DB instance if this DB cluster is to be created as a Read Replica string "" no
restore_to_point_in_time List point-in-time recovery options. Only valid actions are source_cluster_identifier, restore_type and use_latest_restorable_time
list(object({
source_cluster_identifier = string
restore_type = string
use_latest_restorable_time = bool
}))
[] no
retention_period Number of days to retain backups for number 5 no
s3_import Restore from a Percona Xtrabackup in S3. The bucket_name is required to be in the same region as the resource.
object({
bucket_name = string
bucket_prefix = string
ingestion_role = string
source_engine = string
source_engine_version = string
})
null no
scaling_configuration List of nested attributes with scaling properties. Only valid when engine_mode is set to serverless
list(object({
auto_pause = bool
max_capacity = number
min_capacity = number
seconds_until_auto_pause = number
timeout_action = string
}))
[] no
security_groups List of security groups to be allowed to connect to the DB instance list(string) [] no
serverlessv2_scaling_configuration serverlessv2 scaling properties
object({
min_capacity = number
max_capacity = number
})
null no
skip_final_snapshot Determines whether a final DB snapshot is created before the DB cluster is deleted bool true no
snapshot_identifier Specifies whether or not to create this cluster from a snapshot string null no
source_region Source Region of primary cluster, needed when using encrypted storage and region replicas string "" no
stage ID element. Usually used to indicate role, e.g. 'prod', 'staging', 'source', 'build', 'test', 'deploy', 'release' string null no
storage_encrypted Specifies whether the DB cluster is encrypted. The default is false for provisioned engine_mode and true for serverless engine_mode bool false no
storage_type One of 'standard' (magnetic), 'gp2' (general purpose SSD), or 'io1' (provisioned IOPS SSD) string null no
subnet_group_name Database subnet group name. Will use generated label ID if not supplied. string "" no
subnets List of VPC subnet IDs list(string) n/a yes
tags Additional tags (e.g. {'BusinessUnit': 'XYZ'}).
Neither the tag keys nor the tag values will be modified by this module.
map(string) {} no
tenant ID element _(Rarely used, not included by default)_. A customer identifier, indicating who this instance of a resource is for string null no
timeouts_configuration List of timeout values per action. Only valid actions are create, update and delete
list(object({
create = string
update = string
delete = string
}))
[] no
vpc_id VPC ID to create the cluster in (e.g. vpc-a22222ee) string n/a yes
vpc_security_group_ids Additional security group IDs to apply to the cluster, in addition to the provisioned default security group with ingress traffic from existing CIDR blocks and existing security groups list(string) [] no
zone_id Route53 DNS Zone ID as list of string (0 or 1 items). If empty, no custom DNS name will be published.
If the list contains a single Zone ID, a custom DNS name will be pulished in that zone.
Can also be a plain string, but that use is DEPRECATED because of Terraform issues.
any [] no

Outputs

Name Description
activity_stream_arn Activity Stream ARN
activity_stream_name Activity Stream Name
arn Amazon Resource Name (ARN) of the cluster
cluster_identifier Cluster Identifier
cluster_resource_id The region-unique, immutable identifie of the cluster
cluster_security_groups Default RDS cluster security groups
database_name Database name
dbi_resource_ids List of the region-unique, immutable identifiers for the DB instances in the cluster
endpoint The DNS address of the RDS instance
master_host DB Master hostname
master_username Username for the master DB user
reader_endpoint A read-only endpoint for the Aurora cluster, automatically load-balanced across replicas
replicas_host Replicas hostname
security_group_arn Security Group ARN
security_group_id Security Group ID
security_group_name Security Group name

Share the Love

Like this project? Please give it a β˜… on our GitHub! (it helps us a lot)

Are you using this project or any of our other projects? Consider leaving a testimonial. =)

Related Projects

Check out these related projects.

Help

Got a question? We got answers.

File a GitHub issue, send us an email or join our Slack Community.

README Commercial Support

DevOps Accelerator for Startups

We are a DevOps Accelerator. We'll help you build your cloud infrastructure from the ground up so you can own it. Then we'll show you how to operate it and stick around for as long as you need us.

Learn More

Work directly with our team of DevOps experts via email, slack, and video conferencing.

We deliver 10x the value for a fraction of the cost of a full-time engineer. Our track record is not even funny. If you want things done right and you need it done FAST, then we're your best bet.

  • Reference Architecture. You'll get everything you need from the ground up built using 100% infrastructure as code.
  • Release Engineering. You'll have end-to-end CI/CD with unlimited staging environments.
  • Site Reliability Engineering. You'll have total visibility into your apps and microservices.
  • Security Baseline. You'll have built-in governance with accountability and audit logs for all changes.
  • GitOps. You'll be able to operate your infrastructure via Pull Requests.
  • Training. You'll receive hands-on training so your team can operate what we build.
  • Questions. You'll have a direct line of communication between our teams via a Shared Slack channel.
  • Troubleshooting. You'll get help to triage when things aren't working.
  • Code Reviews. You'll receive constructive feedback on Pull Requests.
  • Bug Fixes. We'll rapidly work with you to fix any bugs in our projects.

Slack Community

Join our Open Source Community on Slack. It's FREE for everyone! Our "SweetOps" community is where you get to talk with others who share a similar vision for how to rollout and manage infrastructure. This is the best place to talk shop, ask questions, solicit feedback, and work together as a community to build totally sweet infrastructure.

Discourse Forums

Participate in our Discourse Forums. Here you'll find answers to commonly asked questions. Most questions will be related to the enormous number of projects we support on our GitHub. Come here to collaborate on answers, find solutions, and get ideas about the products and services we value. It only takes a minute to get started! Just sign in with SSO using your GitHub account.

Newsletter

Sign up for our newsletter that covers everything on our technology radar. Receive updates on what we're up to on GitHub as well as awesome new projects we discover.

Office Hours

Join us every Wednesday via Zoom for our weekly "Lunch & Learn" sessions. It's FREE for everyone!

zoom

Contributing

Bug Reports & Feature Requests

Please use the issue tracker to report any bugs or file feature requests.

Developing

If you are interested in being a contributor and want to get involved in developing this project or help out with our other projects, we would love to hear from you! Shoot us an email.

In general, PRs are welcome. We follow the typical "fork-and-pull" Git workflow.

  1. Fork the repo on GitHub
  2. Clone the project to your own machine
  3. Commit changes to your own branch
  4. Push your work back up to your fork
  5. Submit a Pull Request so that we can review your changes

NOTE: Be sure to merge the latest changes from "upstream" before making a pull request!

Copyright

Copyright Β© 2017-2023 Cloud Posse, LLC

License

License

See LICENSE for full details.

Licensed to the Apache Software Foundation (ASF) under one
or more contributor license agreements.  See the NOTICE file
distributed with this work for additional information
regarding copyright ownership.  The ASF licenses this file
to you under the Apache License, Version 2.0 (the
"License"); you may not use this file except in compliance
with the License.  You may obtain a copy of the License at

  https://www.apache.org/licenses/LICENSE-2.0

Unless required by applicable law or agreed to in writing,
software distributed under the License is distributed on an
"AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
KIND, either express or implied.  See the License for the
specific language governing permissions and limitations
under the License.

Trademarks

All other trademarks referenced herein are the property of their respective owners.

About

This project is maintained and funded by Cloud Posse, LLC. Like it? Please let us know by leaving a testimonial!

Cloud Posse

We're a DevOps Professional Services company based in Los Angeles, CA. We ❀️ Open Source Software.

We offer paid support on all of our projects.

Check out our other projects, follow us on twitter, apply for a job, or hire us to help with your cloud strategy and implementation.

Contributors

Erik Osterman
Erik Osterman
Igor Rodionov
Igor Rodionov
Andriy Knysh
Andriy Knysh
Sarkis Varozian
Sarkis Varozian
Mike Crowe
Mike Crowe
Sergey Vasilyev
Sergey Vasilyev
Todor Todorov
Todor Todorov
Lee Huffman
Lee Huffman
RB
RB

README Footer Beacon

More Repositories

1

geodesic

πŸš€ Geodesic is a DevOps Linux Toolbox in Docker
Shell
915
star
2

bastion

πŸ”’Secure Bastion implemented as Docker Container running Alpine Linux with Google Authenticator & DUO MFA support
Shell
623
star
3

terraform-null-label

Terraform Module to define a consistent naming convention by (namespace, stage, name, [attributes])
HCL
516
star
4

atmos

πŸ‘½ Workflow automation tool for DevOps. Keep configuration DRY with hierarchical imports of configurations, inheritance, and WAY more. Native support for Terraform and Helmfile.
Go
490
star
5

terraform-aws-eks-cluster

Terraform module for provisioning an EKS cluster
HCL
453
star
6

terraform-aws-components

Opinionated, self-contained Terraform root modules that each solve one, specific problem
HCL
403
star
7

build-harness

Collection of Makefiles to facilitate building Golang projects, Dockerfiles, Helm charts, and more
Makefile
348
star
8

terraform-aws-tfstate-backend

Terraform module that provision an S3 bucket to store the `terraform.tfstate` file and a DynamoDB table to lock the state file to prevent concurrent modifications and state corruption.
HCL
344
star
9

terraform-aws-ecs-container-definition

Terraform module to generate well-formed JSON documents (container definitions) that are passed to the aws_ecs_task_definition Terraform resource
HCL
316
star
10

terraform-aws-elastic-beanstalk-environment

Terraform module to provision an AWS Elastic Beanstalk Environment
HCL
292
star
11

terraform-aws-cloudfront-s3-cdn

Terraform module to easily provision CloudFront CDN backed by an S3 origin
HCL
255
star
12

helmfiles

Comprehensive Distribution of Helmfiles for Kubernetes
Makefile
250
star
13

terraform-aws-jenkins

Terraform module to build Docker image with Jenkins, save it to an ECR repo, and deploy to Elastic Beanstalk running Docker stack
HCL
250
star
14

terraform-aws-vpc

Terraform Module that defines a VPC with public/private subnets across multiple AZs with Internet Gateways
HCL
212
star
15

terraform-aws-elasticsearch

Terraform module to provision an Elasticsearch cluster with built-in integrations with Kibana and Logstash.
HCL
211
star
16

terraform-aws-ecs-web-app

Terraform module that implements a web app on ECS and supports autoscaling, CI/CD, monitoring, ALB integration, and much more.
HCL
206
star
17

terraform-aws-cloudtrail-cloudwatch-alarms

Terraform module for creating alarms for tracking important changes and occurrences from cloudtrail.
HCL
193
star
18

tfmask

Terraform utility to mask select output from `terraform plan` and `terraform apply`
Go
191
star
19

terraform-aws-cicd

Terraform Module for CI/CD with AWS Code Pipeline and Code Build
HCL
185
star
20

copyright-header

Β© Copyright Header is a utility to manipulate software licenses on source code.
Ruby
177
star
21

terraform-aws-ecr

Terraform Module to manage Docker Container Registries on AWS ECR
HCL
170
star
22

terraform-aws-dynamic-subnets

Terraform module for public and private subnets provisioning in existing VPC
HCL
165
star
23

prometheus-to-cloudwatch

Utility for scraping Prometheus metrics from a Prometheus client endpoint and publishing them to CloudWatch
Go
159
star
24

reference-architectures

[WIP] Get up and running quickly with one of our reference architecture using our fully automated cold-start process.
HCL
154
star
25

charts

The "Cloud Posse" Distribution of Kubernetes Applications
Mustache
149
star
26

terraform-aws-s3-bucket

Terraform module that creates an S3 bucket with an optional IAM user for external CI/CD systems
HCL
147
star
27

terraform-null-ansible

Terraform Module to run ansible playbooks
HCL
146
star
28

terraform-aws-ec2-instance

Terraform module for provisioning a general purpose EC2 host
HCL
143
star
29

terraform-aws-key-pair

Terraform Module to Automatically Generate SSH Key Pairs (Public/Private Keys)
HCL
141
star
30

terraform-aws-ecs-codepipeline

Terraform Module for CI/CD with AWS Code Pipeline and Code Build for ECS https://cloudposse.com/
HCL
139
star
31

terraform-aws-rds

Terraform module to provision AWS RDS instances
HCL
134
star
32

github-authorized-keys

Use GitHub teams to manage system user accounts and authorized_keys
Go
131
star
33

terraform-aws-ecs-alb-service-task

Terraform module which implements an ECS service which exposes a web service via ALB.
HCL
129
star
34

terraform-aws-elasticache-redis

Terraform module to provision an ElastiCache Redis Cluster
HCL
129
star
35

packages

Cloud Posse DevOps distribution of linux packages for native apps, binaries, alpine packages, debian packages, and redhat packages.
Shell
125
star
36

terraform-example-module

Example Terraform Module Scaffolding
HCL
125
star
37

terraform-aws-ec2-bastion-server

Terraform module to define a generic Bastion host with parameterized user_data and support for AWS SSM Session Manager for remote access with IAM authentication.
HCL
124
star
38

tfenv

Transform environment variables for use with Terraform (e.g. `HOSTNAME` ⇨ `TF_VAR_hostname`)
Go
123
star
39

terraform-terraform-label

Terraform Module to define a consistent naming convention by (namespace, stage, name, [attributes])
HCL
116
star
40

terraform-aws-s3-website

Terraform Module for Creating S3 backed Websites and Route53 DNS
HCL
114
star
41

terraform-aws-ec2-autoscale-group

Terraform module to provision Auto Scaling Group and Launch Template on AWS
HCL
113
star
42

terraform-aws-vpc-peering-multi-account

Terraform module to provision a VPC peering across multiple VPCs in different accounts by using multiple providers
HCL
108
star
43

terraform-aws-vpc-peering

Terraform module to create a peering connection between two VPCs in the same AWS account.
HCL
105
star
44

github-commenter

Command line utility for creating GitHub comments on Commits, Pull Request Reviews or Issues
Go
104
star
45

terraform-aws-rds-cloudwatch-sns-alarms

Terraform module that configures important RDS alerts using CloudWatch and sends them to an SNS topic
HCL
103
star
46

terraform-aws-s3-log-storage

This module creates an S3 bucket suitable for receiving logs from other AWS services such as S3, CloudFront, and CloudTrail
HCL
103
star
47

terraform-aws-iam-role

A Terraform module that creates IAM role with provided JSON IAM polices documents.
HCL
101
star
48

github-status-updater

Command line utility for updating GitHub commit statuses and enabling required status checks for pull requests
Go
100
star
49

terraform-aws-codebuild

Terraform Module to easily leverage AWS CodeBuild for Continuous Integration
HCL
96
star
50

terraform-aws-alb

Terraform module to provision a standard ALB for HTTP/HTTP traffic
HCL
94
star
51

terraform-aws-cloudfront-cdn

Terraform Module that implements a CloudFront Distribution (CDN) for a custom origin.
HCL
93
star
52

terraform-aws-ssm-parameter-store

Terraform module to populate AWS Systems Manager (SSM) Parameter Store with values from Terraform. Works great with Chamber.
HCL
93
star
53

terraform-aws-acm-request-certificate

Terraform module to request an ACM certificate for a domain name and create a CNAME record in the DNS zone to complete certificate validation
HCL
93
star
54

terraform-provider-utils

The Cloud Posse Terraform Provider for various utilities (e.g. deep merging, stack configuration management)
Go
93
star
55

terraform-aws-multi-az-subnets

Terraform module for multi-AZ public and private subnets provisioning
HCL
90
star
56

terraform-aws-cloudtrail

Terraform module to provision an AWS CloudTrail and an encrypted S3 bucket with versioning to store CloudTrail logs
HCL
90
star
57

sudosh

Shell wrapper to run a login shell with `sudo` as the current user for the purpose of audit logging
Go
88
star
58

terraform-aws-backup

Terraform module to provision AWS Backup, a fully managed backup service that makes it easy to centralize and automate the back up of data across AWS services such as EBS volumes, RDS databases, DynamoDB tables, EFS file systems, and AWS Storage Gateway volumes.
HCL
87
star
59

terraform-aws-eks-workers

Terraform module to provision an AWS AutoScaling Group, IAM Role, and Security Group for EKS Workers
HCL
84
star
60

terraform-aws-eks-node-group

Terraform module to provision a fully managed AWS EKS Node Group
HCL
82
star
61

terraform-aws-efs

Terraform Module to define an EFS Filesystem (aka NFS)
HCL
79
star
62

terraform-datadog-platform

Terraform module to configure and provision Datadog monitors, custom RBAC roles with permissions, Datadog synthetic tests, Datadog child organizations, and other Datadog resources from a YAML configuration, complete with automated tests.
HCL
79
star
63

terraform-aws-iam-system-user

Terraform Module to Provision a Basic IAM System User Suitable for CI/CD Systems (E.g. TravisCI, CircleCI)
HCL
76
star
64

terraform-aws-sso

Terraform module to configure AWS Single Sign-On (SSO)
HCL
76
star
65

terraform-aws-dynamodb

Terraform module that implements AWS DynamoDB with support for AutoScaling
HCL
72
star
66

terraform-aws-emr-cluster

Terraform module to provision an Elastic MapReduce (EMR) cluster on AWS
HCL
70
star
67

terraform-aws-msk-apache-kafka-cluster

Terraform module to provision AWS MSK
HCL
66
star
68

terraform-yaml-config

Terraform module to convert local and remote YAML configuration templates into Terraform lists and maps
HCL
66
star
69

terraform-aws-iam-user

Terraform Module to provision a basic IAM user suitable for humans.
HCL
66
star
70

slack-notifier

Command line utility to send messages with attachments to Slack channels via Incoming Webhooks
Go
65
star
71

terraform-aws-cloudwatch-logs

Terraform Module to Provide a CloudWatch Logs Endpoint
HCL
61
star
72

terraform-aws-kms-key

Terraform module to provision a KMS key with alias
HCL
61
star
73

actions

Our Library of GitHub Actions
TypeScript
57
star
74

terraform-aws-iam-s3-user

Terraform module to provision a basic IAM user with permissions to access S3 resources, e.g. to give the user read/write/delete access to the objects in an S3 bucket
HCL
53
star
75

load-testing

A collection of best practices, workflows, scripts and scenarios that Cloud Posse uses for load and performance testing of websites and applications (in particular those deployed on Kubernetes clusters)
JavaScript
52
star
76

docs

πŸ“˜ SweetOps documentation for the Cloud Posse way of doing Infrastructure as Code. https://docs.cloudposse.com
Python
51
star
77

terraform-aws-documentdb-cluster

Terraform module to provision a DocumentDB cluster on AWS
HCL
51
star
78

terraform-aws-iam-policy-document-aggregator

Terraform module to aggregate multiple IAM policy documents into single policy document.
HCL
50
star
79

terraform-aws-vpn-connection

Terraform module to provision a site-to-site VPN connection between a VPC and an on-premises network
HCL
49
star
80

terraform-aws-route53-alias

Terraform Module to Define Vanity Host/Domain (e.g. `brand.com`) as an ALIAS record
HCL
48
star
81

terraform-aws-ecs-atlantis

Terraform module for deploying Atlantis as an ECS Task
HCL
47
star
82

terraform-aws-cloudtrail-s3-bucket

S3 bucket with built in IAM policy to allow CloudTrail logs
HCL
47
star
83

terraform-yaml-stack-config

Terraform module that loads an opinionated "stack" configuration from local or remote YAML sources. It supports deep-merged variables, settings, ENV variables, backend config, and remote state outputs for Terraform and helmfile components.
HCL
47
star
84

terraform-aws-transit-gateway

Terraform module to provision AWS Transit Gateway, AWS Resource Access Manager (AWS RAM) Resource, and share the Transit Gateway with the Organization or another AWS Account.
HCL
46
star
85

terraform-aws-route53-cluster-zone

Terraform module to easily define consistent cluster domains on Route53 (e.g. `prod.ourcompany.com`)
HCL
46
star
86

terraform-aws-named-subnets

Terraform module for named subnets provisioning.
HCL
45
star
87

terraform-aws-route53-cluster-hostname

Terraform module to define a consistent AWS Route53 hostname
HCL
45
star
88

terraform-aws-elastic-beanstalk-application

Terraform Module to define an ElasticBeanstalk Application
HCL
44
star
89

terraform-aws-config

This module configures AWS Config, a service that enables you to assess, audit, and evaluate the configurations of your AWS resources.
HCL
43
star
90

terraform-aws-eks-fargate-profile

Terraform module to provision an EKS Fargate Profile
HCL
42
star
91

terraform-aws-efs-backup

Terraform module designed to easily backup EFS filesystems to S3 using DataPipeline
HCL
41
star
92

terraform-aws-sns-topic

Terraform Module to Provide an Amazon Simple Notification Service (SNS)
HCL
40
star
93

terraform-aws-service-control-policies

Terraform module to provision Service Control Policies (SCP) for AWS Organizations, Organizational Units, and AWS accounts
HCL
38
star
94

terraform-aws-cloudformation-stack

Terraform module to provision CloudFormation Stack
HCL
38
star
95

terraform-aws-ec2-client-vpn

HCL
37
star
96

terraform-provider-awsutils

Terraform provider to help with various AWS automation tasks (mostly all that stuff we cannot accomplish with the official AWS terraform provider)
Go
36
star
97

terraform-aws-ecs-cloudwatch-sns-alarms

Terraform module to create CloudWatch Alarms on ECS Service level metrics.
HCL
36
star
98

terraform-aws-iam-assumed-roles

Terraform Module for Assumed Roles on AWS with IAM Groups Requiring MFA
HCL
33
star
99

terraform-aws-mq-broker

Terraform module for provisioning an AmazonMQ broker
HCL
33
star
100

terraform-tls-ssh-key-pair

Terraform module for generating an SSH public/private key file.
HCL
33
star