Your submission was sent successfully! Close


The state of your Kubernetes cluster is kept in the etcd datastore. This page shows how to backup and restore the etcd included in Charmed Kubernetes.

Backing up application specific data, normally stored in a persistent volume, is not currently supported by native Kubernetes. Various third party solutions are available - please refer to their own documentation for details.

Creating an etcd snapshot


Snapshots can only be restored on the **same version of etcd**.

etcd is a distributed key/value store. To create a snapshot, all that is required is to run the snapshot action on one of the units running etcd:

juju run-action etcd/0 snapshot keys-version=v3 --wait

By specifying --wait, the console will wait to return the result of running the action, which in this case includes the path and filename of the generated snapshot file. For example:

 id: 3d6a505e-07d7-4697-8471-60156f87b1b4
     cmd: juju scp etcd/0:/home/ubuntu/etcd-snapshots/etcd-snapshot-2018-09-26-18.04.02.tar.gz
     path: /home/ubuntu/etcd-snapshots/etcd-snapshot-2018-09-26-18.04.02.tar.gz
     sha256: e85ae4d49b6a889de2063031379ab320cc8f09b6e328cdff2fb9179fc641eee9
     size: 68K
     version: |-
       etcdctl version: 3.2.10
       API version: 2
 status: completed
   completed: 2018-09-26 18:04:04 +0000 UTC
   enqueued: 2018-09-26 18:04:04 +0000 UTC
   started: 2018-09-26 18:04:03 +0000 UTC
 unit: etcd/0

This path/filename relates to the unit where the action was run. As we will likely want to use the snapshot on a different unit, we should fetch the snapshot to the local machine. The command to perform this is also helpfully supplied in the copy section of the output (see above):

  juju scp etcd/0:/home/ubuntu/etcd-snapshots/etcd-snapshot-2018-09-26-18.04.02.tar.gz .

It is also wise to check the sha256 checksum of the file you have copied against the value in the previous output:

sha256sum etcd-snapshot-2018-09-26-18.04.02.tar.gz

Restoring a snapshot


Restoring a snapshot should not be performed when there is more than one unit of **etcd** running.

As restoring only works when there is a single unit of etcd, it is usual to deploy a new instance of the application first.

juju deploy etcd new-etcd --series=focal --config channel=3.4/stable

The --series option is included here to illustrate how to specify which series the new unit should be running on. The --config option is required to specify the same channel of etcd as the original unit.

Next we upload and identify the snapshot file to this new unit:

juju attach new-etcd snapshot=./etcd-snapshot-2018-09-26-18.04.02.tar.gz

Then run the restore action:

juju run-action new-etcd/0 restore --wait

Once the restore action has finished, you should see output confirming that the operation is completed. The new etcd application will need to be connected to the rest of the deployment:

juju add-relation new-etcd [calico|flannel|$cni]
juju add-relation new-etcd kubernetes-control-plane

To restore the cluster capabilities of etcd, you can now add more units:

juju add-unit new-etcd -n 2

Once the deployment has settled and all new-etcd units report ready, verify the cluster health with:

 juju run-action new-etcd/0 health --wait

which should return something similar to:

  id: 27fe2081-6513-4968-869d-6c2c092210a1
      message: |-
        member 3c149609bfcf7692 is healthy: got healthy result from
        cluster is healthy
  status: completed
    completed: 2018-10-26 15:16:33 +0000 UTC
    enqueued: 2018-10-26 15:16:32 +0000 UTC
    started: 2018-10-26 15:16:33 +0000 UTC
  unit: new-etcd/0

We appreciate your feedback on the documentation. You can edit this page or file a bug here.