Git Product home page Git Product logo

best-practices's People

Contributors

bensojona avatar clstokes avatar jedineeper avatar joshuajabbour avatar kfishner avatar moofish32 avatar mootpt avatar rcgenova avatar smithamax avatar xueshanf avatar

Watchers

 avatar  avatar

best-practices's Issues

Checklist

Setup

Take all instructions from Setup forward and paste into a new "Issue" on your repository, this will allow you to check items off the list as they're completed and track your progress.

  • Create a GitHub account or use existing
  • Fork the best-practices repo
    • If you're working with a HashiCorp SE, add them as a collaborator on your forked repo
      • "Settings" -> "Collaborators & Teams" -> "Add collaborator"
      • Make sure they have permissions to create "Issues" on the repository
  • Create an AWS account or use existing
  • Generate AWS keys
    • These credentials will need FullAccess to EC2, S3, Route53, and IAM in order for Terraform to create all of the resources
    • Depending on your security requirements, keys can be created in AWS IAM specific to their use case (Packer Builds, Terraform Environments, etc.)
  • Create an Atlas account
    • If you're working with a HashiCorp SE, add them to the "owners" team in your organization
      • You may need to create a new or migrate an existing organization
      • If you would not like to add the SE to the owner team, you can alternatively create a new team and make sure that team is added to all appropriate resources
      • Settings -> Your Organization -> "Teams" -> "Manage" or "Create" -> "Add user"
  • Generate an Atlas API token

Note: Terraform creates real resources in AWS that cost money. Don't forget to destroy your PoC environment when finished to avoid unnecessary expenses.

Set Local Environment Variables

Set the below environment variables if you'll be using Packer or Terraform locally.

$ export AWS_ACCESS_KEY_ID=YOUR_AWS_ACCESS_KEY_ID
$ export AWS_SECRET_ACCESS_KEY=YOUR_AWS_SECRET_ACCESS_KEY
$ export AWS_DEFAULT_REGION=us-east-1
$ export ATLAS_USERNAME=YOUR_ORGANIZATION_NAME
$ export ATLAS_TOKEN=YOUR_ATLAS_TOKEN

Note: The environment variable ATLAS_USERNAME can be set to your individual username or your
organization name in Atlas. Typically, this should be set to your organization name - e.g. hashicorp.

Generate Keys and Certs

There are certain resources in this project that require the use of keys and certs to validate identity, such as Terraform's remote-exec provisioners and TLS in Consul/Vault. For the sake of quicker & easier onboarding, we've created a gen_key.sh and gen_cert.sh script that can generate these for you.

Note: While using this for PoC purposes, these keys and certs should suffice. However, as you start to move your actual applications into this infrastructure, you'll likely want to replace these self-signed certs with certs that are signed by a CA and use keys that are created with your security principles in mind.

  • Generate site keys
    • Run sh gen_key.sh site in setup
      • If you enter the path of an existing private key as an optional second parameter, it will create a public (.pub) key from the existing private (.pem) key specified (e.g. sh gen_key.sh site ~/.ssh/my-existing-private-key.pem)
      • This will generate a public (.pub) and private (.pem) key in the setup/. directory
  • Generate site and vault certs
    • Run sh gen_cert.sh YOUR_DOMAIN YOUR_COMPANY in setup (e.g. sh gen_cert.sh hashicorpdemo.com HashiCorp)
      • If you don't have a domain currently, you can make one up, or grab one from a service like NameCheap to do your testing on
      • This will generate 2 certs, one named site (external self-signed cert for browsers) and one named vault (internal self-signed cert for Consul/Vault TLS), both within the setup/. directory
  • Move all keys & certs created here out of the repository and into a secure location
    • No keys or certs should ever be checked into version control

Create and Configure Artifacts

Use the New Build Configuration tool to create each new Build Configuration below. Enter the names provided as you go through the checklist and be sure to leave the Automatically build on version uploads and Connect build configuration to a GitHub repository boxes unchecked for each.

After creating each Build Configuration, there is some additional configuration you'll need to do. The summary of what will need to be completed for each Build Configuration is below, the relevant values are provided as you go through the checklist.

Add Environment Variables
  • Go into "Variables" in the left navigation of the Build Configuration and set the below Environment Variables with their appropriate values
    • ATLAS_USERNAME
    • AWS_ACCESS_KEY_ID
    • AWS_SECRET_ACCESS_KEY
    • AWS_DEFAULT_REGION: us-east-1
Integrate with GitHub
  • Go into "Integrations" in the left navigation of the Build Configuration
  • Select the best-practices GitHub repository you just forked
  • Enter packer for Packer Directory
  • Enter the appropriate Packer template (provided below) and click Associate
Queue Build

You can then go to "Builds" in the left navigation of each of the Build Configuration(s) and click Queue build, this should create new artifact(s). You'll need to wait for the base artifact to be created before you queue any of the child builds as we take advantage of Base Artifact Variable Injection.

You do NOT want to queue builds for aws-us-east-1-ubuntu-nodejs because this Build Template will be used by the application. Queueing a build for aws-us-east-1-ubuntu-nodejs will fail with the error * Bad source 'app/': stat app/: no such file or directory.

Base Artifact

Child Artifacts

Wait until the Base Artifact has been created before moving on to the child Build Configurations. These will fail with an error of * A source_ami must be specified until the Base Artifact has been created and selected.

For child Build Configurations, there is one additional step you need to take. In "Settings", set Inject artifact ID during build to aws-us-east-1-ubuntu-base for each.

  • Create aws-us-east-1-ubuntu-consul Artifact
  • Create aws-us-east-1-ubuntu-vault Artifact
  • Create aws-us-east-1-ubuntu-haproxy Artifact
  • Create aws-us-east-1-ubuntu-nodejs Build Configuration
  • In "GitHub": Ingress Packer templates
  • Click Queue build in "Builds" for each of the below Build Configurations to create new artifacts for each (remember, we will not do this for aws-us-east-1-ubuntu-nodejs)
    • aws-us-east-1-ubuntu-consul
    • aws-us-east-1-ubuntu-vault
    • aws-us-east-1-ubuntu-haproxy

We built artifacts for the us-east-1 region in this walkthrough. If you'd like to add another region, follow the Multi-Region setup instructions below.

If you decide to update any of the artifact names, be sure those name changes are reflected in your terraform.tfvars file(s).

Provision the aws-global Environment

  • Use the Import Terraform Configuration from GitHub tool to import the aws-global Environment from GitHub
    • Name the environment: YOUR_ATLAS_ORG/aws-global
    • GitHub repository: YOUR_GITHUB_USERNAME/best-practices
    • Path to directory of Terraform files: terraform
  • terraform push your environment to Atlas to set the Terraform variables, the GitHub Ingress does not currently pull in variables
    • Set local environment variables
    • From the root directory, navigate to the global folder: cd terraform/providers/aws/global/.
    • Configure & pull remote state: terraform remote config -backend-config name=$ATLAS_USERNAME/aws-global
    • Get latest modules: terraform get
    • Push to Atlas: terraform push -name $ATLAS_USERNAME/aws-global -var "atlas_token=$ATLAS_TOKEN" -var "atlas_username=$ATLAS_USERNAME" ../../../../terraform/
      • The path ../../../../terraform/ must be provided so that the Terraform command and project
        files are referenced correctly.
      • The first plan in Atlas will fail, this is okay. We'll correct this in the steps below.
  • Navigate to the aws-global environment
  • In "Settings": check Plan on artifact uploads and click Save
  • In "Variables": add the below Environment Variables with appropriate values
    • ATLAS_USERNAME
    • AWS_ACCESS_KEY_ID
    • AWS_SECRET_ACCESS_KEY
    • AWS_DEFAULT_REGION: us-east-1
    • TF_ATLAS_DIR: providers/aws/global
      • Atlas uses the TF_ATLAS_DIR variable to identify where it should run Terraform commands within the repo
  • In "Variables": update all Terraform variables containing the value REPLACE_IN_ATLAS
    • Update domain with your domain (e.g. hashicorpdemo.com)
      • If you don't have a domain currently, you can make one up, or grab one from a service like NameCheap to do your testing on
      • We use this domain to create S3 buckets to host a static website, so if you're making a domain up, try to make it unique to your company to avoid S3 bucket naming conflicts
    • Update atlas_username with your Atlas username
    • Update iam_admins with a comma separated list of users you'd like added to the admin group in IAM (e.g. cameron,jay,jon,kevin)
      • This is defaulted to a user named global-admin
      • Be sure that you don't use a name that already exists in IAM for this AWS account or you will see conflict errors
      • If you don't want any admin users to be created, just leave this blank
  • In "Integrations": under "GitHub Integration" click Update GitHub settings to pull the latest configuration from master
  • In "Changes": click Queue plan if one has not already been queued, then Confirm & Apply to provision the aws-global environment

Provision the aws-us-east-1-prod Environment

  • Use the Import Terraform Configuration from GitHub tool to import the aws-us-east-1-prod Environment from GitHub
    • Name the environment: YOUR_ATLAS_ORG/aws-us-east-1-prod
    • GitHub repository: YOUR_GITHUB_USERNAME/best-practices
    • Path to directory of Terraform files: terraform
  • terraform push your environment to Atlas to set the Terraform variables, the GitHub Ingress does not currently pull in variables
    • Set local environment variables
    • From the root directory, navigate to the us_east_1_prod folder: cd terraform/providers/aws/us_east_1_prod/.
    • Configure & pull remote state: terraform remote config -backend-config name=$ATLAS_USERNAME/aws-us-east-1-prod
    • Get latest modules: terraform get
    • Push to Atlas: terraform push -name $ATLAS_USERNAME/aws-us-east-1-prod -var "atlas_token=$ATLAS_TOKEN" -var "atlas_username=$ATLAS_USERNAME" ../../../../terraform/
      • The path ../../../../terraform/ must be provided so that the Terraform command and project
        files are referenced correctly.
      • The first plan in Atlas will fail, this is okay. We'll correct this in the steps below.
  • Navigate to the aws-us-east-1-prod environment
  • In "Settings": check Plan on artifact uploads and click Save
  • In "Variables": add the below Environment Variables with appropriate values
    • ATLAS_USERNAME
    • AWS_ACCESS_KEY_ID
    • AWS_SECRET_ACCESS_KEY
    • AWS_DEFAULT_REGION: us-east-1
    • TF_ATLAS_DIR: providers/aws/us_east_1_prod
      • Atlas uses the TF_ATLAS_DIR variable to identify where it should run Terraform commands within the repo
  • In "Variables": update all Terraform Variables containing the value REPLACE_IN_ATLAS, you will use the contents of the keys and certs created in Generate Keys and Certs as values for most of these variables
    • Update atlas_token with your Atlas token
    • Update atlas_username with your Atlas username
    • Update site_public_key with the contents of site.pub
    • Update site_private_key with the contents of site.pem
    • Update site_ssl_cert with the contents of site.crt
    • Update site_ssl_key with the contents of site.key
    • Update vault_ssl_cert with the contents of vault.crt
    • Update vault_ssl_key with the contents of vault.key
  • In "Integrations": under "GitHub Integration" click Update GitHub settings to pull the latest configuration from master
  • In "Changes": click Queue plan if one has not already been queued, then Confirm & Apply to provision the aws-us-east-1-prod environment
    • Note: You'll likely see an error related to the OpenVPN AMI during the apply, you'll need to follow the link to the AWS Marketplace provided and complete the steps to opt-in to the OpenVPN AMI for the apply to complete successfully
    • On a successful apply, there will be instructions output in a green font that will tell you how to interact with your new infrastructure

This same process can be repeated for the aws-us-east-1-staging environment as well as any other regions you would like to deploy infrastructure into. If you are deploying into a new region, be sure you have Artifacts created for it by following the Multi-Region steps below.

Setup Vault

A HA Vault should have already been provisioned, but you'll need to initialize and unseal Vault to make it work. To do so, SSH into each of the newly provisioned Vault instances and follow the below instructions. The output from your apply in Atlas will tell you how to SSH into Vault.

  • Initialize Vault

    $ vault init | tee /tmp/vault.init > /dev/null

  • Retrieve the unseal keys and root token from /tmp/vault.init and store these in a safe place

  • Shred keys and token once they are stored in a safe place

    $ shred /tmp/vault.init

  • Use the unseal keys you just retrieved to unseal Vault

    $ vault unseal YOUR_UNSEAL_KEY_1
    $ vault unseal YOUR_UNSEAL_KEY_2
    $ vault unseal YOUR_UNSEAL_KEY_3
    
  • Authenticate with Vault by entering your root token retrieved earlier

    $ vault auth

  • Shred the token

    $ shred -u -z ~/.vault-token

After Vault is initialized and unsealed, update the below variable(s) and apply the changes. Next time you deploy your application, you should see the Vault/Consul Template integration working in your Node.js website!

  • In "Variables" of the aws-us-east-1-prod environment: Update vault_token with the root-token
  • Commit a new change (git commit --allow-empty -m "Force a change in Atlas") to your demo-app-nodejs repo, this should trigger a new "plan" in aws-us-east-1-prod after a new artifact is built
  • In "Changes" of the the aws-us-east-1-prod environment: Queue a new plan and apply the changes to deploy the new application to see the Vault/Consul Template integration at work

You'll eventually want to configure Vault specific to your needs and setup appropriate ACLs.

Multi-Region

If you'd like to expand outside of us-east-1, there are a few changes you need to make. We'll use the region us-west-2 as an example of how to do this.

In the base.json Packer template...

Add a new variable for the new region's AMI and a new variable for the new Build name. Note that the AMI will need to be from the region you intend to use.

"us_west_2_ami":   "ami-8ee605bd",
"us_west_2_name":  "aws-us-west-2-ubuntu-base",

Add an additional builder for the new region

{
  "name":            "aws-us-west-2-ubuntu-base",
  "type":            "amazon-ebs",
  "access_key":      "{{user `aws_access_key`}}",
  "secret_key":      "{{user `aws_secret_key`}}",
  "region":          "us-west-2",
  "vpc_id":          "",
  "subnet_id":       "",
  "source_ami":      "{{user `us_west_2_ami`}}",
  "instance_type":   "t2.micro",
  "ssh_username":    "{{user `ssh_username`}}",
  "ssh_timeout":     "10m",
  "ami_name":        "{{user `us_west_2_name`}} {{timestamp}}",
  "ami_description": "{{user `us_west_2_name`}} AMI",
  "run_tags":        { "ami-create": "{{user `us_west_2_name`}}" },
  "tags":            { "ami": "{{user `us_west_2_name`}}" },
  "ssh_private_ip":  false,
  "associate_public_ip_address": true
}

Add an additional post-processor for the new region

{
  "type": "atlas",
  "only": ["aws-us-west-2-ubuntu-base"],
  "artifact": "{{user `atlas_username`}}/{{user `us_west_2_name`}}",
  "artifact_type": "amazon.image",
  "metadata": {
    "created_at": "{{timestamp}}"
  }
}

Once the updates to base.json have been completed and pushed to master (this should trigger a new Build Configuration to be sent to Atlas), complete the Child Artifact steps with the new region instead of us-east-1 to build new artifacts in that region.

To deploy these new artifacts...

In each of the new "us_west_2" terraform.tfvars files...

  • Replace all instances of us-east-1 with us-west-2.
  • Update the OpenVPN ami from ami-5fe36434 to ami-9fe2f2af
  • You may need to update the azs variable depending on what the subnets in that region support

Finally, push these new environments to master and follow the same steps you completed to deploy your environments in us-east-1.

Terraform Destroy

If you want to destroy the environment, run the following command in the appropriate environment's directory

$ terraform destroy -var "atlas_token=$ATLAS_TOKEN" -var "atlas_username=$ATLAS_USERNAME"

There is currently an issue when destroying the aws_internet_gateway resource that requires you to run terraform destroy a second time as it fails the first.

Note: terraform destroy deletes real resources, it is important that you take extra precaution when using this command. Verify that you are in the correct environment, verify that you are using the correct keys, and set any extra configuration necessary to prevent someone from accidentally destroying infrastructure.

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.