Commit f42750b61ad4e22d787a19592892dd81b6dac49c
Committed by
Earth Ugat

1 parent
e835e351
Bump to v0.4.0.
- Install mariadb::client before mariadb::default. Backup script will now compress first before encrypting. - Modify the way the backup script is called. Optimize cookbook a bit. Modify reload_from_s3 recipe to account for new compression-then-encryption order.
Showing
9 changed files
with
66 additions
and
60 deletions
... | ... | @@ -8,16 +8,16 @@ driver: |
8 | 8 | subnet_id: subnet-d530d8b1 |
9 | 9 | instance_type: t2.micro |
10 | 10 | associate_public_ip: true |
11 | - require_chef_omnibus: true | |
11 | + require_chef_omnibus: 12.12.15 | |
12 | 12 | shared_credentials_profile: earth |
13 | 13 | |
14 | 14 | provisioner: |
15 | - name: chef_solo | |
15 | + name: chef_zero | |
16 | 16 | |
17 | 17 | platforms: |
18 | 18 | - name: ubuntu-14.04 |
19 | 19 | driver: |
20 | - image_id: ami-50946030 | |
20 | + image_id: ami-d732f0b7 | |
21 | 21 | transport: |
22 | 22 | username: ubuntu |
23 | 23 | ssh_key: ~/.ssh/cfe_stg_20160222.pem | ... | ... |
1 | +## 0.4.0 | |
2 | +### Fixed | |
3 | +- Install `mariadb::client` first, before doing `mariadb::default`. Otherswise, it might error out. | |
4 | +- Fixed recipes to account for the above change, including `reload_from_s3`. | |
5 | +- Backup script now properly compresses the file before encrypting it. | |
6 | + | |
7 | +### Changed | |
8 | +- Optimized certain variables and recipt for better understandability. Some variables are now incompatible with previous versions. | |
9 | + | |
1 | 10 | # 0.3.1 |
2 | 11 | |
3 | 12 | Use encoding utf8 and collation utf8_general_ci as defaults when creating databases | ... | ... |
... | ... | @@ -8,6 +8,9 @@ The server is assumed to be using an IAM role with S3 bucket read/write access, |
8 | 8 | |
9 | 9 | When encryption is enabled for DB backups, the private and public keys are shared across all databases in the `db_map` attribute. Encryption is enabled separately for each individual database (see Usage below). |
10 | 10 | |
11 | + | |
12 | +Note that enabling encryption can cause the backup process to take a very long time depending on how large the database is. That said, it is still very much recommended to do so. | |
13 | + | |
11 | 14 | ## Supported Platforms |
12 | 15 | |
13 | 16 | Ubuntu 14.04 | ... | ... |
... | ... | @@ -47,17 +47,15 @@ else |
47 | 47 | default['cfe-mariadb']['backup']['aws_bin'] = '/usr/local/bin/aws' |
48 | 48 | default['cfe-mariadb']['backup']['mysqldump_bin'] = '/usr/bin/mysqldump' |
49 | 49 | end |
50 | + | |
50 | 51 | # Path to directory where the backup script should be placed. |
51 | 52 | # Uncomment to set custom locations. |
52 | 53 | #default['cfe-mariadb']['backup']['script_dir'] = ::File.join( |
53 | 54 | # node['mariadb']['configuration']['path'], 'scripts' |
54 | 55 | #) |
56 | +default['cfe-mariadb']['backup']['log_dir'] = '/var/log/backup_db_to_s3' | |
55 | 57 | |
56 | -default['cfe-mariadb']['backup']['cron']['min'] = '0' | |
57 | -default['cfe-mariadb']['backup']['cron']['hour'] = '0' | |
58 | -default['cfe-mariadb']['backup']['cron']['day'] = '*' | |
59 | -default['cfe-mariadb']['backup']['cron']['mon'] = '*' | |
60 | -default['cfe-mariadb']['backup']['cron']['wday'] = '*' | |
58 | +default['cfe-mariadb']['backup']['cron']['sched'] = '0 0 * * *' | |
61 | 59 | default['cfe-mariadb']['backup']['cron']['mailto'] = "''" # Empty single quotes |
62 | 60 | |
63 | 61 | # Basic options for logrotate |
... | ... | @@ -91,11 +89,6 @@ default['cfe-mariadb']['backup']['logrotate']['options'] = %w{ |
91 | 89 | # |
92 | 90 | # The key file will be stored in the same directory |
93 | 91 | # as the script as 'pub.key'. |
94 | -# NOTE: | |
95 | -# Enabling encryption will result in HUGE file sizes and, | |
96 | -# depending on how large a database is, can take a LOT of time | |
97 | -# during the backup process. That said, it is still recommended to | |
98 | -# enforce encryption on DB backups. | |
99 | 92 | default['cfe-mariadb']['encrypt']['priv_key'] = nil |
100 | 93 | default['cfe-mariadb']['encrypt']['pub_key'] = nil |
101 | 94 | ... | ... |
... | ... | @@ -4,7 +4,7 @@ maintainer_email 'sysadmin @ chromedia.com' |
4 | 4 | license 'Apache License' |
5 | 5 | description 'Simplifies setup of MariaDB in Chromedia.' |
6 | 6 | long_description IO.read(File.join(File.dirname(__FILE__), 'README.md')) |
7 | -version '0.3.1' | |
7 | +version '0.4.0' | |
8 | 8 | |
9 | 9 | { |
10 | 10 | 'mariadb' => '0.3.1', | ... | ... |
... | ... | @@ -50,6 +50,10 @@ file pub_key_file do |
50 | 50 | only_if { md['encrypt']['pub_key'] } |
51 | 51 | end |
52 | 52 | |
53 | +directory mdb['log_dir'] do | |
54 | + recursive true | |
55 | +end | |
56 | + | |
53 | 57 | template "#{mdb['script_dir']}/backup_db_to_s3" do |
54 | 58 | variables( |
55 | 59 | :db_map => md['db_map'], |
... | ... | @@ -66,13 +70,15 @@ end |
66 | 70 | # Chef built-in resource 'cron' seems to have a problem with |
67 | 71 | # the syntax here. But they suggest using 'cron_d', instead, from |
68 | 72 | # the cron cookbook (https://docs.chef.io/resource_cron.html). |
73 | +sched = mdbc['sched'].split(' ') | |
69 | 74 | cron_d 'backup_db_to_s3' do |
70 | - command "bash #{mdb['script_dir']}/backup_db_to_s3" | |
71 | - minute mdbc['min'] | |
72 | - hour mdbc['hour'] | |
73 | - day mdbc['day'] | |
74 | - month mdbc['mon'] | |
75 | - weekday mdbc['wday'] | |
75 | + command "bash #{mdb['script_dir']}/backup_db_to_s3 "\ | |
76 | + ">> #{mdb['log_dir']}/backup_db_to_s3.log 2>&1" | |
77 | + minute sched[0] | |
78 | + hour sched[1] | |
79 | + day sched[2] | |
80 | + month sched[3] | |
81 | + weekday sched[4] | |
76 | 82 | mailto mdbc['mailto'] |
77 | 83 | path '/usr/local/sbin:/usr/local/bin:/sbin:/bin:/usr/sbin:/usr/bin' |
78 | 84 | end | ... | ... |
... | ... | @@ -30,6 +30,8 @@ chef_gem 'chef-rewind' do |
30 | 30 | end |
31 | 31 | require 'chef/rewind' |
32 | 32 | |
33 | +# mysql2_chef_gem seems to want the client to be installed first. | |
34 | +include_recipe 'mariadb::client' | |
33 | 35 | include_recipe 'mariadb' |
34 | 36 | |
35 | 37 | # It looks like the service is sometimes not being restarted | ... | ... |
... | ... | @@ -49,11 +49,10 @@ node['cfe-mariadb']['db_map'].each do |dbx| |
49 | 49 | dbx_name = dbx[:db_name] |
50 | 50 | end |
51 | 51 | |
52 | - keyname = "#{dbx[:bak_filename]}#{dbx[:bak_encrypted] ? '.enc.gz' : '.gz'}" | |
53 | - filegz = "#{tmp_dir}/#{keyname}" | |
54 | - filesql = "#{tmp_dir}/#{dbx[:bak_filename]}" | |
52 | + keyname = "#{dbx[:bak_filename]}.gz#{dbx[:bak_encrypted] ? '.enc' : ''}" | |
53 | + filepath = "#{tmp_dir}/#{dbx[:bak_filename]}" | |
55 | 54 | |
56 | - awscli_s3_file filegz do | |
55 | + awscli_s3_file "#{tmp_dir}/#{keyname}" do | |
57 | 56 | region node['cfe-mariadb']['s3_region'] |
58 | 57 | bucket node['cfe-mariadb']['s3_bucket'] |
59 | 58 | key keyname |
... | ... | @@ -64,30 +63,28 @@ node['cfe-mariadb']['db_map'].each do |dbx| |
64 | 63 | end |
65 | 64 | end |
66 | 65 | |
67 | - execute "unpack_#{filegz}" do | |
68 | - command "gzip -d #{filegz}" | |
69 | - end | |
70 | - | |
71 | - execute "decrypt_#{filesql}.enc" do | |
66 | + execute "decrypt_#{filepath}.gz.enc" do | |
72 | 67 | command "openssl smime -decrypt -binary -inkey #{priv_key_file} "\ |
73 | - "-in #{filesql}.enc -out #{filesql} -inform DEM" | |
74 | - only_if { dbx[:bak_encrypted] } | |
75 | - notifies :delete, "file[#{filesql}.enc]" | |
68 | + "-in #{filepath}.gz.enc -out #{filepath}.gz -inform DEM" | |
69 | + only_if { ::File.exist?("#{filepath}.gz.enc") } | |
70 | + notifies :delete, "file[#{filepath}.gz.enc]" | |
76 | 71 | end |
77 | 72 | |
78 | - execute "reload_#{filesql}" do | |
73 | + execute "gzip -d #{filepath}.gz" | |
74 | + | |
75 | + execute "reload_#{filepath}" do | |
79 | 76 | command "mysql -h #{node['mariadb']['mysqld']['bind_address']} "\ |
80 | 77 | "-P #{node['mariadb']['mysqld']['port']} -u #{dbx[:db_user]} "\ |
81 | - "-p'#{dbx[:db_pass]}' -D #{dbx_name} < #{filesql}" | |
82 | - notifies :delete, "file[#{filesql}]" | |
78 | + "-p'#{dbx[:db_pass]}' -D #{dbx_name} < #{filepath}" | |
79 | + notifies :delete, "file[#{filepath}]" | |
83 | 80 | sensitive true |
84 | 81 | end |
85 | 82 | |
86 | - file "#{filesql}.enc" do | |
83 | + file "#{filepath}.gz.enc" do | |
87 | 84 | action :nothing |
88 | 85 | end |
89 | 86 | |
90 | - file filesql do | |
87 | + file filepath do | |
91 | 88 | action :nothing |
92 | 89 | end |
93 | 90 | end | ... | ... |
... | ... | @@ -10,7 +10,10 @@ |
10 | 10 | set -e |
11 | 11 | |
12 | 12 | suffix=.backup_db_to_s3 |
13 | -[ -f /tmp/*"$suffix" ] && exit 200 | |
13 | +if [ -f /tmp/*"$suffix" ] ; then | |
14 | + ( >&2 echo "[ERROR] Another operation might still be in progress" ) | |
15 | + exit 200 | |
16 | +fi | |
14 | 17 | tmp_file=$( mktemp --suffix "$suffix" ) |
15 | 18 | |
16 | 19 | <% bak_dir = "#{Chef::Config[:file_cache_path]}/backup_db_to_s3" -%> |
... | ... | @@ -24,15 +27,6 @@ aws_bin=<%= @aws_bin %> |
24 | 27 | mysqldump_bin=<%= @mysqldump_bin %> |
25 | 28 | pub_key_file=<%= @pub_key_file %> |
26 | 29 | |
27 | -log_dir=/var/log/backup_db_to_s3 | |
28 | -if [[ ! -d "$log_dir" ]] ; then | |
29 | - mkdir -p "$log_dir" | |
30 | -fi | |
31 | - | |
32 | -exec 3>&1 4>&2 | |
33 | -trap 'exec 2>&4 1>&3' 0 1 2 3 | |
34 | -exec 1>>"${log_dir}/backup_db_to_s3.log" 2>&1 | |
35 | - | |
36 | 30 | if [[ ! -d "$bak_dir" ]] ; then |
37 | 31 | echo "$(date) : Create backup directory." |
38 | 32 | mkdir -p "$bak_dir" |
... | ... | @@ -55,7 +49,7 @@ export_db() { |
55 | 49 | # Encrypt a file using OpenSSL and a given public key. |
56 | 50 | # The original file will be replaced by a new file, suffixed with '.enc'. |
57 | 51 | # Args: |
58 | -# $1 = dump file filename, e.g. 'mydb.sql' | |
52 | +# $1 = compressed dump file filename, e.g. 'mydb.sql.gz' | |
59 | 53 | encrypt_file() { |
60 | 54 | echo "$(date) : Encrypt file ${1}." |
61 | 55 | openssl smime -encrypt -binary -text -aes256 -in "${bak_dir}/${1}" \ |
... | ... | @@ -65,7 +59,7 @@ encrypt_file() { |
65 | 59 | |
66 | 60 | # Compress the backup file with gzip. |
67 | 61 | # Args: |
68 | -# $1 = dump file filename, e.g. 'mydb.sql', 'mydb.sql.enc' | |
62 | +# $1 = dump file filename, e.g. 'mydb.sql' | |
69 | 63 | compress_backup_file() { |
70 | 64 | echo "$(date) : Gzip file ${1}." |
71 | 65 | gzip "${bak_dir}/${1}" |
... | ... | @@ -73,15 +67,14 @@ compress_backup_file() { |
73 | 67 | |
74 | 68 | # Rotate the current backups in S3. |
75 | 69 | # Args: |
76 | -# $1 = dump file filename, e.g. 'mydb.sql', 'mydb.sql.enc' | |
70 | +# $1 = resulting dump filename, e.g. 'mydb.sql.gz', 'mydb.sql.gz.enc' | |
77 | 71 | # $2 = max number of backup files to store at a time |
78 | 72 | increment_backup_names() { |
79 | - bak_keyname="${1}.gz" | |
73 | + bak_keyname=$1 | |
80 | 74 | max_backups=$2 |
81 | 75 | |
82 | 76 | baks=$( "$aws_bin" --output text --region "$region" \ |
83 | - s3api list-objects --bucket "$bucket" \ | |
84 | - | grep '^CONTENTS' | cut -f3 | grep "^${bak_keyname}" || echo "" ) | |
77 | + s3 ls "s3://${bucket}/" | awk '{ printf("%s\n", $4); }' || echo "" ) | |
85 | 78 | |
86 | 79 | echo "$(date) : Backup rotation for ${bak_keyname}." |
87 | 80 | start=$((max_backups - 1)) |
... | ... | @@ -106,15 +99,16 @@ increment_backup_names() { |
106 | 99 | |
107 | 100 | # Upload the compressed db backup file. |
108 | 101 | # Args: |
109 | -# $1 = dump file filename, e.g. 'mydb.sql', 'mydb.sql.enc' | |
102 | +# $1 = resulting dump filename, e.g. 'mydb.sql.gz', 'mydb.sql.gz.enc' | |
110 | 103 | upload_to_s3() { |
111 | - echo "$(date) : Upload ${1}.gz to S3 bucket ${bucket}." | |
104 | + echo "$(date) : Upload ${1} to S3 bucket ${bucket}." | |
112 | 105 | "$aws_bin" --region "$region" \ |
113 | - s3 mv "${bak_dir}/${1}.gz" "s3://${bucket}/${1}.gz" | |
106 | + s3 mv "${bak_dir}/${1}" "s3://${bucket}/${1}" | |
114 | 107 | } |
115 | 108 | |
116 | 109 | # First, perform mysqldump on each database (and encrypt if desired): |
117 | 110 | |
111 | +<% bfname = '' -%> | |
118 | 112 | <% @db_map.each do |db| -%> |
119 | 113 | <% if db.is_a?(Array) -%> |
120 | 114 | <% db_name = db[0] -%> |
... | ... | @@ -123,17 +117,19 @@ upload_to_s3() { |
123 | 117 | <% db_name = db[:db_name] -%> |
124 | 118 | <% end -%> |
125 | 119 | export_db <%= db_name %> <%= db[:db_user] %> '<%= db[:db_pass] %>' <%= db[:bak_filename] %> |
120 | +compress_backup_file <%= db[:bak_filename] %> | |
126 | 121 | <% if db[:bak_encrypted] -%> |
127 | -encrypt_file <%= db[:bak_filename] %> | |
122 | +<% bfname = "#{db[:bak_filename]}.gz.enc" -%> | |
123 | +encrypt_file <%= db[:bak_filename] %>.gz | |
124 | +<% else -%> | |
125 | +<% bfname = "#{db[:bak_filename]}.gz" -%> | |
128 | 126 | <% end -%> |
129 | 127 | <% end -%> |
130 | 128 | |
131 | -# Then compress and upload the backup files one by one: | |
129 | +# Then upload the backup files one by one: | |
132 | 130 | |
133 | 131 | <% @db_map.each do |db| -%> |
134 | 132 | <% if db.is_a?(Array) then db = db[1] end -%> |
135 | -<% bfname = db[:bak_encrypted] ? "#{db[:bak_filename]}.enc" : db[:bak_filename] -%> | |
136 | -compress_backup_file <%= bfname %> | |
137 | 133 | increment_backup_names <%= bfname %> <%= db[:bak_maxcopies] %> |
138 | 134 | upload_to_s3 <%= bfname %> |
139 | 135 | ... | ... |