You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
128 lines
3.4 KiB
128 lines
3.4 KiB
#! /bin/sh
|
|
|
|
set -e
|
|
|
|
if [ "${S3_ACCESS_KEY_ID}" == "**None**" ]; then
|
|
echo "Warning: You did not set the S3_ACCESS_KEY_ID environment variable."
|
|
fi
|
|
|
|
if [ "${S3_SECRET_ACCESS_KEY}" == "**None**" ]; then
|
|
echo "Warning: You did not set the S3_SECRET_ACCESS_KEY environment variable."
|
|
fi
|
|
|
|
if [ "${S3_BUCKET}" == "**None**" ]; then
|
|
echo "You need to set the S3_BUCKET environment variable."
|
|
exit 1
|
|
fi
|
|
|
|
if [ "${MYSQL_HOST}" == "**None**" ]; then
|
|
echo "You need to set the MYSQL_HOST environment variable."
|
|
exit 1
|
|
fi
|
|
|
|
if [ "${MYSQL_USER}" == "**None**" ]; then
|
|
echo "You need to set the MYSQL_USER environment variable."
|
|
exit 1
|
|
fi
|
|
|
|
if [ "${MYSQL_PASSWORD}" == "**None**" ]; then
|
|
echo "You need to set the MYSQL_PASSWORD environment variable or link to a container named MYSQL."
|
|
exit 1
|
|
fi
|
|
|
|
if [ "${S3_IAMROLE}" != "true" ]; then
|
|
# env vars needed for aws tools - only if an IAM role is not used
|
|
export AWS_ACCESS_KEY_ID=$S3_ACCESS_KEY_ID
|
|
export AWS_SECRET_ACCESS_KEY=$S3_SECRET_ACCESS_KEY
|
|
export AWS_DEFAULT_REGION=$S3_REGION
|
|
fi
|
|
|
|
MYSQL_HOST_OPTS="-h $MYSQL_HOST -P $MYSQL_PORT -u$MYSQL_USER -p$MYSQL_PASSWORD"
|
|
DUMP_START_TIME=$(date +"%Y-%m-%dT%H%M%SZ")
|
|
|
|
copy_s3 () {
|
|
SRC_FILE=$1
|
|
DEST_FILE=$2
|
|
|
|
if [ "${S3_ENDPOINT}" == "**None**" ]; then
|
|
AWS_ARGS=""
|
|
else
|
|
AWS_ARGS="--endpoint-url ${S3_ENDPOINT}"
|
|
fi
|
|
|
|
echo "Uploading ${DEST_FILE} on S3..."
|
|
|
|
cat $SRC_FILE | aws $AWS_ARGS s3 cp - s3://$S3_BUCKET/$S3_PREFIX/$DEST_FILE
|
|
|
|
if [ $? != 0 ]; then
|
|
>&2 echo "Error uploading ${DEST_FILE} on S3"
|
|
fi
|
|
|
|
rm $SRC_FILE
|
|
}
|
|
# Multi file: yes
|
|
if [ ! -z "$(echo $MULTI_FILES | grep -i -E "(yes|true|1)")" ]; then
|
|
if [ "${MYSQLDUMP_DATABASE}" == "--all-databases" ]; then
|
|
DATABASES=`mysql $MYSQL_HOST_OPTS -e "SHOW DATABASES;" | grep -Ev "(Database|information_schema|performance_schema|mysql|sys|innodb)"`
|
|
else
|
|
DATABASES=$MYSQLDUMP_DATABASE
|
|
fi
|
|
|
|
for DB in $DATABASES; do
|
|
echo "Creating individual dump of ${DB} from ${MYSQL_HOST}..."
|
|
|
|
DUMP_FILE="/tmp/${DB}.sql.gz"
|
|
|
|
mysqldump $MYSQL_HOST_OPTS $MYSQLDUMP_OPTIONS --databases $DB | gzip > $DUMP_FILE
|
|
|
|
if [ $? == 0 ]; then
|
|
if [ "${S3_FILENAME}" == "**None**" ]; then
|
|
S3_FILE="${DUMP_START_TIME}.${DB}.sql.gz"
|
|
else
|
|
S3_FILE="${S3_FILENAME}.${DB}.sql.gz"
|
|
fi
|
|
|
|
copy_s3 $DUMP_FILE $S3_FILE
|
|
else
|
|
>&2 echo "Error creating dump of ${DB}"
|
|
fi
|
|
done
|
|
# Multi file: no
|
|
else
|
|
echo "Creating dump for ${MYSQLDUMP_DATABASE} from ${MYSQL_HOST}..."
|
|
|
|
DUMP_FILE="/tmp/dump.sql.gz"
|
|
mysqldump $MYSQL_HOST_OPTS $MYSQLDUMP_OPTIONS $MYSQLDUMP_DATABASE | gzip > $DUMP_FILE
|
|
|
|
if [ $? == 0 ]; then
|
|
if [ "${S3_FILENAME}" == "**None**" ]; then
|
|
S3_FILE="${DUMP_START_TIME}.dump.sql.gz"
|
|
else
|
|
S3_FILE="${S3_FILENAME}.sql.gz"
|
|
fi
|
|
|
|
copy_s3 $DUMP_FILE $S3_FILE
|
|
else
|
|
>&2 echo "Error creating dump of all databases"
|
|
fi
|
|
fi
|
|
|
|
if [ "${DELETE_OLDER_THAN}" != "**None**" ]; then
|
|
aws $AWS_ARGS s3 ls s3://$S3_BUCKET/$S3_PREFIX/ | grep " PRE " -v | while read -r line;
|
|
do
|
|
created=`echo $line|awk {'print $1" "$2'}`
|
|
created=`date -d "$created" +%s`
|
|
older_than=`date -d "$DELETE_OLDER_THAN" +%s`
|
|
if [ $created -lt $older_than ]
|
|
then
|
|
fileName=`echo $line|awk {'print $4'}`
|
|
if [ $fileName != "" ]
|
|
then
|
|
printf 'Deleting "%s"\n' $fileName
|
|
aws $AWS_ARGS s3 rm s3://$S3_BUCKET/$S3_PREFIX/$fileName
|
|
fi
|
|
fi
|
|
done;
|
|
fi
|
|
|
|
echo "SQL backup finished"
|