I have an EC2 ASG on AWS and I'm interested in storing the shell script that's used to instantiate any given instance in an S3 bucket and have it downloaded and run upon instantiation, but it all feels a little rickety even though I'm using an IAM Instance Role
, transferring via HTTPS, and encrypting the script itself while at rest in the S3 bucket using using KMS
S3 Server Side Encryption
(because the KMS
method was throwing an 'Unknown' error).
The Setup
- Created an
IAM Instance Role
that gets assigned to any instance in my ASG upon instantiation, resulting in my AWS creds being baked into the instance asENV
vars - Uploaded and encrypted my
Instance-Init.sh
script to S3 resulting in a private endpoint like so : https://s3.amazonaws.com/super-secret-bucket/Instance-Init.sh
In The User-Data
Field
I input the following into the User Data
field when creating the Launch Configuration
I want my ASG to use:
#!/bin/bash
apt-get update
apt-get -y install python-pip
apt-get -y install awscli
cd /home/ubuntu
aws s3 cp s3://super-secret-bucket/Instance-Init.sh . --region us-east-1
chmod +x Instance-Init.sh
. Instance-Init.sh
shred -u -z -n 27 Instance-Init.sh
The above does the following:
- Updates package lists
- Installs Python (required to run
aws-cli
) - Installs
aws-cli
- Changes to the
/home/ubuntu
user directory - Uses the
aws-cli
to download theInstance-Init.sh
file fromS3
. Due to theIAM Role
assigned to my instance, my AWS creds are automagically discovered byaws-cli
. TheIAM Role
also grants my instance the permissions necessary to decrypt the file. - Makes it executable
- Runs the script
- Deletes the script after it's completed.
The Instance-Init.sh
Script
The script itself will do stuff like setting env
vars and docker run
the containers that I need deployed on my instance. Kinda like so:
#!/bin/bash
export MONGO_USER='MyMongoUserName'
export MONGO_PASS='Top-Secret-Dont-Tell-Anyone'
docker login -u <username> -p <password> -e <email>
docker run - e MONGO_USER=${MONGO_USER} -e MONGO_PASS=${MONGO_PASS} --name MyContainerName quay.io/myQuayNameSpace/MyAppName:latest
Very Handy
This creates a very handy way to update User-Data
scripts without the need to create a new Launch Config
every time you need to make a minor change. And it does a great job of getting env
vars out of your codebase and into a narrow, controllable space (the Instance-Init.sh
script itself).
But it all feels a little insecure. The idea of putting my master DB creds into a file on S3 is unsettling to say the least.
The Questions
- Is this a common practice or am I dreaming up a bad idea here?
- Does the fact that the file is downloaded and stored (albeit briefly) on the fresh instance constitute a vulnerability at all?
- Is there a better method for deleting the file in a more secure way?
- Does it even matter whether the file is deleted after it's run? Considering the secrets are being transferred to
env
vars it almost seems redundant to delete theInstance-Init.sh
file. - Is there something that I'm missing in my nascent days of ops?
Thanks for any help in advance.
S3 Server Side Encryption
. AWS offers this encryption and from what I understand each object is encrypted with its own key, which itself is encrypted but he AWS master key, which they are responsible for. The object keys are rotated (don't remember how often) so, to be honest, they've got a better process in-place than I could roll myself. – Archegonium