You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
|
|
#! /bin/sh
set -e
if [ -n "${MYSQL_PASSWORD_FILE}" ]; then MYSQL_PASSWORD=$(cat "$MYSQL_PASSWORD_FILE") export MYSQL_PASSWORD fi
if [ -n "${S3_ACCESS_KEY_ID_FILE}" ]; then S3_ACCESS_KEY_ID=$(cat "$S3_ACCESS_KEY_ID_FILE") export S3_ACCESS_KEY_ID fi
if [ -n "${S3_SECRET_ACCESS_KEY_FILE}" ]; then S3_SECRET_ACCESS_KEY=$(cat "$S3_SECRET_ACCESS_KEY_FILE") export S3_SECRET_ACCESS_KEY fi
if [ "${S3_ACCESS_KEY_ID}" == "**None**" ]; then echo "Warning: You did not set the S3_ACCESS_KEY_ID environment variable." fi
if [ "${S3_SECRET_ACCESS_KEY}" == "**None**" ]; then echo "Warning: You did not set the S3_SECRET_ACCESS_KEY environment variable." fi
if [ "${S3_BUCKET}" == "**None**" ]; then echo "You need to set the S3_BUCKET environment variable." exit 1 fi
if [ "${MYSQL_HOST}" == "**None**" ]; then echo "You need to set the MYSQL_HOST environment variable." exit 1 fi
if [ "${MYSQL_USER}" == "**None**" ]; then echo "You need to set the MYSQL_USER environment variable." exit 1 fi
if [ "${MYSQL_PASSWORD}" == "**None**" ]; then echo "You need to set the MYSQL_PASSWORD environment variable or link to a container named MYSQL." exit 1 fi
if [ "${S3_IAMROLE}" != "true" ]; then # env vars needed for aws tools - only if an IAM role is not used export AWS_ACCESS_KEY_ID=$S3_ACCESS_KEY_ID export AWS_SECRET_ACCESS_KEY=$S3_SECRET_ACCESS_KEY export AWS_DEFAULT_REGION=$S3_REGION fi
MYSQL_HOST_OPTS="-h $MYSQL_HOST -P $MYSQL_PORT -u$MYSQL_USER -p$MYSQL_PASSWORD" DUMP_START_TIME=$(date +"%Y-%m-%dT%H%M%SZ")
copy_s3 () { SRC_FILE=$1 DEST_FILE=$2
if [ "${S3_ENDPOINT}" == "**None**" ]; then AWS_ARGS="" else AWS_ARGS="--endpoint-url ${S3_ENDPOINT}" fi
echo "$AWS_ARGS" echo "Uploading ${DEST_FILE} to S3..."
cat $SRC_FILE | aws $AWS_ARGS s3 cp - s3://$S3_BUCKET/$S3_PREFIX/$DEST_FILE
if [ $? != 0 ]; then >&2 echo "Error uploading ${DEST_FILE} on S3" fi
rm $SRC_FILE }
# Multi file: yes if [ ! -z "$(echo $MULTI_FILES | grep -i -E "(yes|true|1)")" ]; then if [ "${MYSQLDUMP_DATABASE}" == "--all-databases" ]; then DATABASES=`mysql $MYSQL_HOST_OPTS -e "SHOW DATABASES;" | grep -Ev "(Database|information_schema|performance_schema|mysql|sys|innodb)"` else DATABASES=$MYSQLDUMP_DATABASE fi
for DB in $DATABASES; do echo "Creating individual dump of ${DB} from ${MYSQL_HOST}..."
DUMP_FILE="/tmp/${DB}.sql.gz"
mysqldump $MYSQL_HOST_OPTS $MYSQLDUMP_OPTIONS --databases $DB | gzip > $DUMP_FILE
if [ $? == 0 ]; then if [ "${S3_FILENAME}" == "**None**" ]; then S3_FILE="${DUMP_START_TIME}.${DB}.sql.gz" else S3_FILE="${S3_FILENAME}.${DB}.sql.gz" fi
copy_s3 $DUMP_FILE $S3_FILE else >&2 echo "Error creating dump of ${DB}" fi done # Multi file: no else echo "Creating dump for ${MYSQLDUMP_DATABASE} from ${MYSQL_HOST}..."
DUMP_FILE="/tmp/dump.sql.gz" mysqldump $MYSQL_HOST_OPTS $MYSQLDUMP_OPTIONS $MYSQLDUMP_DATABASE | gzip > $DUMP_FILE
if [ $? == 0 ]; then if [ "${S3_FILENAME}" == "**None**" ]; then S3_FILE="${DUMP_START_TIME}.dump.sql.gz" else S3_FILE="${S3_FILENAME}.sql.gz" fi
copy_s3 $DUMP_FILE $S3_FILE else >&2 echo "Error creating dump of all databases" fi fi
echo "SQL backup finished"
|