Skip to content

Latest commit

 

History

History
173 lines (117 loc) · 4.98 KB

File metadata and controls

173 lines (117 loc) · 4.98 KB

TerracidTango

Terraform Template for AWS Infraestructure

Overview

This template deploys a VPC accross two availability zones in the region specified (eu-west-1 is the default region). For each AZ there is a private and a public subnet. There is one Bastion Host in the first public subnet to access the instances in the private subnets. There are two t2.micro instances deployed, one in each private subnet, running RancherOS 1.5. Finally, all the network plumbing and an Application Load Balancer are deployed to route the traffic to http port 80 to the EC2 instances.

Architecture Overview

""

Setup

Dependencies

  • AWS Credentials File (link)
  • terraform (link)
  • Create a ssh-key in your AWS account

For more info on how Terraform works visit the Getting Starte Guide and the Docs. For more specific information about AWS resources go to the AWS Provider Documentation.

Usage

Clone the repository

git clone [email protected]:acidtango/terracid_tango.git

Rename the terraform.tfvars.example file to be terraform.tfvars:

mv terraform.tfvars.example terraform.tfvars

Modify the file to use your specific credentials:

# File where your AWS credentials are stored
aws_credentials_file = "~/.aws/credentials"

# SSH key pair used to create the EC2 instances.
aws_key_name = "your-aws-key"

# AWS CLI profile to use. Defaults to 'default'.
aws_cli_profile = "your-profile"

Initialize terraform

$ terraform init

If everything is correct the following command should output an execution plan

$ terraform plan

First apply - Swarm Initialization

Execute the infrastructure

$ terraform apply

If everything goes right, you should found one instace called Swarm Manager in the EC2 dashboard. This instance will be the Swarm Manager Leader, and after a few moments it should be ready to accept new nodes.

Second apply - Configure AutoScalingGroups

Connect to the instance (trough the Bastion Host) and verify that the Swarm is initialized and that there is a script called cluster-init.sh created in /home/rancher

$ docker node ls

ID                            HOSTNAME                STATUS              AVAILABILITY        MANAGER STATUS      ENGINE VERSION
zwqenokj41vl3oho7ulv3d2bp *   example-hostname        Ready               Active              Leader              18.09.1

$ ls -l /home/rancher

total 4
-rwx--x--x    1 rancher  root           517 Feb 13 10:41 cluster-init.sh

This script uses the Swarm init compose file to deploy Traefik and Swarmpit in our cluster.

Run the script and take note of the tokens for the managers and the workers.

$ ./cluster-init.sh
Unable to find image 'bwits/docker-git-alpine:latest' locally
latest: Pulling from bwits/docker-git-alpine
...
Cloning into 'swarm_init'...
...
Creating network infrastructure_net
Creating service infrastructure_swarmpit
Creating service infrastructure_swarmpit_db
Creating service infrastructure_swarmpit_agent
Creating service infrastructure_traefik

====== MANAGER TOKEN ======
To add a manager to this swarm, run the following command:

    docker swarm join --token SWMTKN-1-3x.. # Rest of the manager token

===========================


====== WORKER  TOKEN ======
To add a worker to this swarm, run the following command:

    docker swarm join --token SWMTKN-1-3x.. # Rest of the worker token

===========================

Copy both commands and add them in the variables.tf file, in the swarm_managers_user_data and swarm_workers_user_data

Next, go to the ec2_instances.tf file. Go to the block resource "aws_launch_template" "swarm_managers", and change the user_data key from "${base64encode("${local.swarm_managers_init_user_data}")}" to "${base64encode("${local.swarm_managers_user_data}")}". Finally, uncomment the two commented blocks in the end of the file.

After all this changes are done, apply the infrastructure once again

$ terraform apply

Testing your deployment

If the terraform apply command completes succesfully, it will output the DNS Address of the Load Balancer

$ terraform apply
...

Apply complete! Resources: 0 added, 0 changed, 0 destroyed.

Outputs:

  alb_dns_name = some.dns.name

Make sure that both traefik.your-domain.com and swarmpit.your-domain.com resolve to the load balancer address.

After a few moments, connecting to traefik.your-domain.com should show the Traefik dashboard and swarmpit.your-domain.com should show you the swarmpit dashboard (default credentials are admin/admin).

If you want to bring down the infrastructure

terraform destroy