.. | ||
code | ||
README.md |
Introduction
So far we've learnt how to create, update and destroy infrastructure and use variables in our code so we can reuse what we've written over multiple enviroments. One thing you may have noticed while running the examples is that theres a file (sometimes two) created in you working directory in CloudShell called terraform.tfstate
(and maybe terraform.tfstate.backup
). This is the state files that terraform/tofu uses to keep track of whats actually been deployed and any data that might returned by the provider, such as public_ip address which we used in our previous example.
Now the problem of this state file being local is that only you have access to it. If you work in a team or even from different computers your self you are going to want to store that statefile somewhere it can be easily read and updated. Now you mightthing great I'll commitit to git, but this is a bad idea, it can easily get out of sync and that can lead to all kinds of ugly problems. Fortunately terraform/tofu supplies us with a solution, these are called backends and there are lots of them, you can store your state in a consul cluster for example, an HTTP endpoint in Gitlab and the one we are going to use as these examples are in AWS is DynamoDB and S3. S3 will store the actual state file and we are going use DynamoDB to provide a lock flag that stops multiple people trying to update the stack at the same time, which would end in tears!
Setting up AWS
As this is a terraform/tofu workshop we aren't going to use clickops to make these resources we are going to use IaC of course. So start by creating a new directory in the 3-remote-states/
directory called state
mkdir state
cd state
Now lets make some really simple terraform for this, create a new file called main.tf
vi main.tf
and lets add the following code:
provider "aws" {
region = "eu-west-1"
}
resource "aws_s3_bucket" "tfstate_bucket" {
bucket = var.bucket_name
acl = "private"
versioning {
enabled = true
}
lifecycle {
prevent_destroy = false
}
tags = {
Name = "${var.environment}-s3"
}
}
resource "aws_dynamodb_table" "remotestate_table" {
name = var.table_name
hash_key = "LockID"
billing_mode = "PAY_PER_REQUEST"
attribute {
name = "LockID"
type = "S"
}
tags = {
Name = "${var.environment}-Dynamo"
}
}
You'll notice in here we are creating an S3 bucket and a table in dynamoDB. These are going to be used to save our state from our main stack. It also includes some variables and we'll define those in a moment.
Now save and exit and create a new file called versions.tf
vi versions.tf
Lets add the following content:
terraform {
required_version = ">= 1.0"
required_providers {
aws = {
source = "hashicorp/aws"
version = ">= 4.66.1"
}
}
}
Save and exit and now lets create our variables.tf
file:
vi variables.tf
An we want to set up the following, I've delibarately left out the default values for the bucket_name
and the table_name
. This is because I want you to use unquie values, you'll be prompted for these soon!
Note
S3 buckets are in a global name space so your bucket_name must be unquie!
variable "environment" {
description = "Default environment"
type = string
default = "demo"
}
variable "bucket_name" {
description = "Name for S3 Bucket"
type = string
}
variable "table_name" {
description = "Name for DynamoDB Table"
type = string
}