This is an example repository that compliaments a walk-thru video of provisioning AWS networking, a public Jumpbox, a private BOSH/UAA/CredHub/Concourse (BUCC), and an example 5-node ZooKeeper cluster.
Go to file
starkandwayne-bot 41b3b003f1 Updating CODE_OF_CONDUCT.md 2019-01-15 15:42:53 -05:00
envs remove dir is based on local root dir name 2018-03-16 17:31:21 -04:00
src using bucc develop branch which has 'bucc down' fix 2018-03-14 08:17:31 -04:00
.gitignore ignore tmp 2018-03-14 21:11:34 -04:00
.gitmodules AWS terraform plan + instructions 2018-03-13 18:14:37 -04:00
CODE_OF_CONDUCT.md Updating CODE_OF_CONDUCT.md 2019-01-15 15:42:53 -05:00
README.md newer way to source env 2018-03-15 11:42:10 -04:00

README.md

Walk thru of BUCC on AWS

This is an example repository that compliaments a walk-thru video of provisioning AWS networking, a public Jumpbox, a private BOSH/UAA/CredHub/Concourse (BUCC), and an example 5-node ZooKeeper cluster.

All the state files created from terraform and bosh are listed in .gitignore. After going thru this walk thru, you would copy + paste liberally into your own private Git repository that would allow you to commit your state files.

Clone this repo

git clone https://github.com/starkandwayne/bucc-walk-thru-aws
cd bucc-walk-thru-aws

Configure and terraform AWS

cp envs/aws/aws.tfvars{.example,}

Populate envs/aws/aws.tfvars with your AWS API credentials.

Create a key pair called bucc-walk-thru

The private key will automatically be downloaded by your browser. Copy it into envs/ssh/bucc-walk-thru.pem:

mv ~/Downloads/bucc-walk-thru.pem envs/ssh/bucc-walk-thru.pem

Allocate an elastic IP and store the IP in envs/aws/aws.tfvars at jumpbox_ip = "<your-ip>". This will be used for your jumpbox/bastion host later.

envs/aws/bin/update

This will create a new VPC, a NAT to allow egress Internet access, and two subnets - a public DMZ for the jumpbox, and a private subnet for BUCC.

Deploy Jumpbox

This repository already has a submodule to jumpbox-deployment, and some wrapper scripts. You are ready to go.

git submodule update --init
envs/jumpbox/bin/update

This will create a tiny EC2 VM, with a 64G persistent disk, and a jumpbox user whose home folder is placed upon that large persistent disk.

Looking inside envs/jumpbox/bin/update you can see that it is a wrapper script to use bosh create-env. Variables from the terraform output are consumed via the envs/jumpbox/bin/jumpbox-vars-from-terraform.sh helper script.

bosh create-env src/jumpbox-deployment/jumpbox.yml \
  -o src/jumpbox-deployment/aws/cpi.yml \
  -l <(envs/jumpbox/bin/vars-from-terraform.sh) \
  ...

If you ever need to enlarge the disk, edit envs/jumpbox/operators/persistent-homes.yml, change disk_size: 65_536 to a larger number, and run envs/jumpbox/bin/update again. That's the beauty of bosh create-env.

SSH into Jumpbox

To SSH into jumpbox we will need to store the private key of the jumpbox into a file, etc. There is a helper script for this:

envs/jumpbox/bin/ssh

You can see that the jumpbox user's home directory is placed on the /var/vcap/store persistent disk:

jumpbox/0:~$ pwd
/var/vcap/store/home/jumpbox

SOCKS5 magic tunnel thru Jumpbox

envs/jumpbox/bin/socks5-tunnel

The output will look like:

Starting SOCKS5 on port 9999...
export BOSH_ALL_PROXY=socks5://localhost:9999
export CREDHUB_PROXY=socks5://localhost:9999
Unauthorized use is strictly prohibited. All access and activity
is subject to logging and monitoring.

Deploying BUCC

source <(envs/bucc/bin/env)
bucc up --cpi aws --spot-instance

This will create envs/bucc/vars.yml stub for you to populate. Fortunately, we have a helper script to get most of it:

envs/bucc/bin/vars-from-terraform.sh > envs/bucc/vars.yml

Now run bucc up again:

bucc up

BUT... you are about download a few hundred GB of BOSH releases, AND THEN upload them thru the jumpbox to your new BUCC/BOSH VM on AWS. Either it will take a few hours or you can move to the jumpbox and run the commands there.

So let's use the jumpbox. For your convenience, there is a nice wrapper script which uploads this project to your jumpbox, runs bucc up, and then downloads the modified state files created by bucc up/bosh create-env back into this project locally:

envs/bucc/bin/update-upon-jumpbox

Inside the SSH session:

cd ~/workspace/walk-thru
envs/bucc/bin/update

To access your BOSH/CredHub, remember to have your SOCKS5 magic tunnel running in another terminal:

envs/jumpbox/bin/socks5-tunnel

After it has bootstrapped your BUCC/BOSH from either the jumpbox or your local machine:

export BOSH_ALL_PROXY=socks5://localhost:9999
export CREDHUB_PROXY=socks5://localhost:9999

source <(envs/bucc/bin/env)

bosh env

If you get a connection error like below, your SOCKS5 tunnel is no longer up. Run it again.

Fetching info:
  Performing request GET 'https://10.10.1.4:25555/info':
    Performing GET request:
      Retry: Get https://10.10.1.4:25555/info: dial tcp [::1]:9999: getsockopt: connection refused

Instead, the output should look like:

Using environment '10.10.1.4' as client 'admin'

Name      bucc-walk-thru
UUID      71fec1e3-d34b-4940-aba8-e53e8c848dd1
Version   264.7.0 (00000000)
CPI       aws_cpi
Features  compiled_package_cache: disabled
          config_server: enabled
          dns: disabled
          snapshots: disabled
User      admin

Succeeded

The envs/bucc/bin/update script also pre-populated a stemcell:

$ bosh stemcells
Using environment '10.10.1.4' as client 'admin'

Name                                     Version  OS             CPI  CID
bosh-aws-xen-hvm-ubuntu-trusty-go_agent  3541.9   ubuntu-trusty  -    ami-02f7c167 light

And it pre-populates the cloud config with our two subnets and with vm_types that all use AWS spot instances:

$ bosh cloud-config
...
vm_types:
- cloud_properties:
    ephemeral_disk:
      size: 25000
    instance_type: m4.large
    spot_bid_price: 10

Deploy something

Five-node cluster of Apache ZooKeeper:

source <(envs/bucc/bin/env)

bosh deploy -d zookeeper <(curl https://raw.githubusercontent.com/cppforlife/zookeeper-release/master/manifests/zookeeper.yml)

Upgrade Everything

cd src/bucc
git checkout develop
git pull
cd -

envs/bucc/bin/update-upon-jumpbox

Backup & Restore

You can start cutting backups of your BUCC and its BOSH deployments immediately.

Read https://www.starkandwayne.com/blog/bucc-bbr-finally/ to learn more about the inclusion of BOSH Backup & Restore (BBR) into your BUCC VM.

Backup within Jumpbox

BBR does not currently support SOCKS5, so we will request the backup from inside the jumpbox.

envs/jumpbox/bin/ssh
mkdir -p ~/workspace/backups
cd ~/workspace/backups
source <(~/workspace/walk-thru/envs/bucc/bin/env)
bucc bbr backup

Restore within Jumpbox

Our backup is meaningless if we haven't tried to restore it. So, let's do it.

From within the jumpbox session above, destroy the BUCC VM:

bucc down

Now re-deploy it without any fancy pre-populated stemcells etc:

bucc up

When it comes up, it is empty. It has no stemcells nor deployments.

$ bosh deployments
Name  Release(s)  Stemcell(s)  Team(s)  Cloud Config

0 deployments

But if you check the AWS console you'll see the ZooKeeper cluster is still running. Let's restore the BOSH/BUCC data.

cd ~/workspace/backups/
last_backup=$(find . -type d -regex ".+_.+Z" | sort -r | head -n1)
bucc bbr restore --artifact-path=${last_backup}

Our BOSH now remembers its ZooKeeper cluster:

$ bosh deployments
Using environment '10.10.1.4' as client 'admin'

Name       Release(s)       Stemcell(s)                                     Team(s)  Cloud Config
zookeeper  zookeeper/0.0.7  bosh-aws-xen-hvm-ubuntu-trusty-go_agent/3541.9  -        latest

Sync State Files

In the example above we rebuilt our BUCC/BOSH from within the Jumpbox. This means that our local laptop does not have the updated state files. From your laptop, sync them back:

envs/jumpbox/bin/rsync from . walk-thru

Destroy Everything

To discover all your running deployments

source <(envs/bucc/bin/env)

bosh deployments

To delete each one:

bosh delete-deployment -d zookeeper

To destroy your BUCC/BOSH VM and its persistent disk, and the orphaned disks from your deleted zookeeper deployment:

bosh clean-up --all
bucc down

To destroy your jumpbox and its persistent disk:

envs/jumpbox/bin/update delete-env

To destroy your AWS VPC, subnets, etc:

cd envs/aws
make destroy
cd -