You can run scheduled and on-demand backups of your Redis Cloud databases to a remote storage location. In Redis Cloud Essentials, a scheduled backups run every 24 hours; in Redis Cloud Pro, you can choose the backup interval and time.

Note -
Backups are only available for paid subscriptions. Customers with free subscriptions must upgrade to a paid subscriptions to use backups. Subscription upgrade does not cause downtime.

Can I export my Redis data from Redis Cloud?

Absolutely! There is no lock-in with Redis Cloud. Using the instructions on this page, you can export your latest RDB backup file from your cloud storage, FTP or HTTP server to any Redis server of your choice.

Backing up to Amazon Simple Storage Service (AWS S3)

To use an S3 bucket for storing backups, first access your AWS Management Console and follow these steps:

  1. Go to the AWS S3 Management Console in: Services > Storage > S3
  2. Configure the permissions for your S3 bucket:
    • For a new bucket:
      1. Click Create Bucket.
      2. Enter a name and region for the bucket and click Next.
      3. Set the bucket properties that are required by your company standards, and click Next.
      4. Set Public access properties according to your company standards, and click Next.
      5. Review the properties, and click Create bucket.
      6. Find your bucket in the bucket list, and click on the bucket name.
      7. In Permissions > Access Control, click Add account:
        1. In the Account field enter: fd1b05415aa5ea3a310265ddb13b156c7c76260dbc87e037a8fc290c3c86b614
        2. Select the Write objects, Read bucket permissions, and Write bucket permissions, and click Save.
    • For an existing bucket, click on the S3 bucket and go to the Permissions tab.
      1. Click Add account.
      2. In the Access for other AWS accounts section, enter:
        1. In the Account field enter: fd1b05415aa5ea3a310265ddb13b156c7c76260dbc87e037a8fc290c3c86b614
        2. Select the permissions that you need.
      3. Click Save.

After the permissions are set for your bucket, set the Backup Path to the path of your S3 bucket and click Apply. For example, if the name of your bucket is backups-bucket, use the path: s3://backups-bucket

Backing up to Google Cloud Storage (GCS)

For Google Cloud Platform (GCP) console subscriptions, to use a GCS bucket for storing your resources’ backups:

  1. Login to your account on Google Cloud Platform
  2. Navigate to Storage -> Browser
  3. Click on the three dot button (1) on your relevant bucket name and choose Edit bucket permissions (2). GCS bucketpermissions
  4. Under Add members, enter:
  5. For the role, select Storage Legacy -> Storage Legacy Bucket Writer. Google Cloud StoragePermissions
  6. Click on the Add button.

Once your bucket’s permissions are set, you can use it with your resource by setting its Backup Path to the path of your GCS bucket and clicking the Activate button. For example, if your backups bucket’s name is backups-bucket, use the path:


Backing up to Azure Blob Storage (ABS)

To use an ABS container for storing your resources’ backups, follow these steps in your Microsoft Azure Management Portal:

Create an Azure Storage account

Create a container

Manage storage account access keys

Set your resource’s Backup Path to the path of your ABS storage account and clicking the Apply button using the following syntax:



  • storage_account_access_key: the primary access key to the storage account
  • storage_account_name: the storage account name
  • container_name: the name of the container, if needed.
  • path: the backups path, if needed.

Backing up to an FTP Server

To store your resource backups on an FTP server, set its Backup Path using the following syntax:



  • protocol: the server’s protocol, can be either ftp or ftps.
  • username: your username, if needed.
  • password: your password, if needed.
  • hostname: the hostname or IP address of the server.
  • port: the port number of the server, if needed.
  • path: the backups path, if needed.