Application Level Backup and Restore SingleStore database using KubeStash

KubeStash offers application-level backup and restore functionality for SingleStore databases. It captures both manifest and logical data backups of any SingleStore database in a single snapshot. During the restore process, KubeStash first applies the SingleStore manifest to the cluster and then restores the data into it.

This guide will give you how you can take application-level backup and restore your SingleStore databases using Kubestash.

Before You Begin

  • At first, you need to have a Kubernetes cluster, and the kubectl command-line tool must be configured to communicate with your cluster. If you do not already have a cluster, you can create one by using Minikube or Kind.
  • Install KubeDB in your cluster following the steps here.
  • Install KubeStash in your cluster following the steps here.
  • Install KubeStash kubectl plugin following the steps here.
  • If you are not familiar with how KubeStash backup and restore SingleStore databases, please check the following guide here.

You should be familiar with the following KubeStash concepts:

To keep everything isolated, we are going to use a separate namespace called demo throughout this tutorial.

$ kubectl create ns demo
namespace/demo created

Note: YAML files used in this tutorial are stored in docs/guides/singlestore/backup/kubestash/application-level/examples directory of kubedb/docs repository.

Backup SingleStore

KubeStash supports backups for SingleStore instances across different configurations, including Standalone, and Cluster setups. In this demonstration, we’ll focus on a SingleStore database using singlestore clustering. The backup and restore process is similar for Standalone and Cluster configurations as well.

This section will demonstrate how to take application-level backup of a SingleStore database. Here, we are going to deploy a SingleStore database using KubeDB. Then, we are going to back up the database at the application level to a GCS bucket. Finally, we will restore the entire SingleStore database.

Create SingleStore License Secret

We need SingleStore License to create SingleStore Database. So, Ensure that you have acquired a license and then simply pass the license by secret.

$ kubectl create secret generic -n demo license-secret \
                --from-literal=username=license \
                --from-literal=password='your-license-set-here'
secret/license-secret created

Deploy Sample SingleStore Database

Let’s deploy a sample SingleStore database and insert some data into it.

Create SingleStore CR:

Below is the YAML of a sample SingleStore CR that we are going to create for this tutorial:

apiVersion: kubedb.com/v1alpha2
kind: Singlestore
metadata:
  name: sample-singlestore
  namespace: demo
spec:
  version: "8.7.10"
  topology:
    aggregator:
      replicas: 2
      podTemplate:
        spec:
          containers:
          - name: singlestore
            resources:
              limits:
                memory: "2Gi"
                cpu: "600m"
              requests:
                memory: "2Gi"
                cpu: "600m"
      storage:
        accessModes:
        - ReadWriteOnce
        resources:
          requests:
            storage: 1Gi
    leaf:
      replicas: 3
      podTemplate:
        spec:
          containers:
            - name: singlestore
              resources:
                limits:
                  memory: "2Gi"
                  cpu: "600m"
                requests:
                  memory: "2Gi"
                  cpu: "600m"                      
      storage:
        accessModes:
          - ReadWriteOnce
        resources:
          requests:
            storage: 10Gi
  licenseSecret:
    name: license-secret
  storageType: Durable
  deletionPolicy: WipeOut

Here,

  • spec.version is the name of the SinglestoreVersion CRD where the docker images are specified. In this tutorial, a SingleStore 8.7.10 database is going to be created.
  • spec.topology specifies that it will be used as cluster mode. If this field is nil it will be work as standalone mode.
  • spec.topology.aggregator.replicas or spec.topology.leaf.replicas specifies that the number replicas that will be used for aggregator or leaf.
  • spec.storageType specifies the type of storage that will be used for SingleStore database. It can be Durable or Ephemeral. Default value of this field is Durable. If Ephemeral is used then KubeDB will create SingleStore database using EmptyDir volume. In this case, you don’t have to specify spec.storage field. This is useful for testing purposes.
  • spec.topology.aggregator.storage or spec.topology.leaf.storage specifies the StorageClass of PVC dynamically allocated to store data for this database. This storage spec will be passed to the PetSet created by KubeDB operator to run database pods. You can specify any StorageClass available in your cluster with appropriate resource requests.
  • spec.deletionPolicy gives flexibility whether to nullify(reject) the delete operation of Singlestore crd or which resources KubeDB should keep or delete when you delete Singlestore crd. If admission webhook is enabled, It prevents users from deleting the database as long as the spec.deletionPolicy is set to DoNotTerminate. Learn details of all DeletionPolicy here

Note: spec.storage section is used to create PVC for database pod. It will create PVC with storage size specified in storage.resources.requests field. Don’t specify limits here. PVC does not get resized automatically.

Create the above SingleStore CR,

$ kubectl apply -f https://github.com/kubedb/docs/raw/v2024.9.30/docs/guides/singlestore/backup/kubestash/application-level/examples/sample-singlestore.yaml
singlestore.kubedb.com/sample-singlestore created

KubeDB will deploy a SingleStore database according to the above specification. It will also create the necessary Secrets and Services to access the database.

Let’s check if the database is ready to use,

$ kubectl get singlestores.kubedb.com -n demo
NAME                   VERSION   STATUS    AGE
sample-singlestore      8.7.10   Ready     4m22s

The database is Ready. Verify that KubeDB has created a Secret and a Service for this database using the following commands,

$ kubectl get secret -n demo -l=app.kubernetes.io/instance=sample-singlestore
NAME                           TYPE                       DATA   AGE
sample-singlestore-root-cred   kubernetes.io/basic-auth   2      4m58s

$ kubectl get service -n demo -l=app.kubernetes.io/instance=sample-singlestore
NAME                       TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)             AGE
sample-singlestore        ClusterIP   10.128.230.168   <none>        3306/TCP,8081/TCP   5m10s
sample-singlestore-pods   ClusterIP   None             <none>        3306/TCP            5m10s

Here, we have to use service sample-singlestore and secret sample-singlestore-root-cred to connect with the database. KubeDB creates an AppBinding CR that holds the necessary information to connect with the database.

Verify AppBinding:

Verify that the AppBinding has been created successfully using the following command,

$ kubectl get appbindings -n demo
NAME                  AGE
sample-singlestore    9m24s

Let’s check the YAML of the above AppBinding,

$ kubectl get appbindings -n demo sample-singlestore -o yaml
apiVersion: appcatalog.appscode.com/v1alpha1
kind: AppBinding
metadata:
  annotations:
    kubectl.kubernetes.io/last-applied-configuration: |
      {"apiVersion":"kubedb.com/v1alpha2","kind":"Singlestore","metadata":{"annotations":{},"name":"sample-singlestore","namespace":"demo"},"spec":{"deletionPolicy":"WipeOut","licenseSecret":{"name":"license-secret"},"storageType":"Durable","topology":{"aggregator":{"podTemplate":{"spec":{"containers":[{"name":"singlestore","resources":{"limits":{"cpu":"0.6","memory":"2Gi"},"requests":{"cpu":"0.6","memory":"2Gi"}}}]}},"replicas":2,"storage":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"1Gi"}}}},"leaf":{"podTemplate":{"spec":{"containers":[{"name":"singlestore","resources":{"limits":{"cpu":"0.6","memory":"2Gi"},"requests":{"cpu":"0.6","memory":"2Gi"}}}]}},"replicas":3,"storage":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"10Gi"}}}}},"version":"8.7.10"}}      
  creationTimestamp: "2024-09-11T07:03:44Z"
  generation: 1
  labels:
    app.kubernetes.io/component: database
    app.kubernetes.io/instance: sample-singlestor
    app.kubernetes.io/managed-by: kubedb.com
    app.kubernetes.io/name: singlestores.kubedb.com
  name: sample-singlestore
  namespace: demo
  ownerReferences:
  - apiVersion: kubedb.com/v1alpha2
    blockOwnerDeletion: true
    controller: true
    kind: Singlestore
    name: sample-singlestore
    uid: e08e1f37-d869-437d-9b15-14c6aef3f406
  resourceVersion: "4904220"
  uid: 92b2b318-1874-4471-97ec-d789c6e16809
spec:
  appRef:
    apiGroup: kubedb.com
    kind: Singlestore
    name: sample-singlestore
    namespace: demo
  clientConfig:
    service:
      name: sample-singlestore
      path: /
      port: 3306
      scheme: tcp
    url: tcp(sample-singlestore.demo.svc:3306)/
  parameters:
    apiVersion: config.kubedb.com/v1alpha1
    kind: SinglestoreConfiguration
    masterAggregator: sample-singlestore-aggregator-0.sample-singlestore-pods.demo.svc
    stash:
      addon:
        backupTask:
          name: ""
        restoreTask:
          name: ""
  secret:
    name: sample-singlestore-root-cred
  type: kubedb.com/singlestore
  version: 8.7.10

KubeStash uses the AppBinding CR to connect with the target database. It requires the following two fields to set in AppBinding’s .spec section.

  • .spec.parameters.masterAggregator specifies the dns of master aggregator node that we have to mention in mysqldump command when taken backup or restore.
  • .spec.clientConfig.service.name specifies the name of the Service that connects to the database.
  • .spec.secret specifies the name of the Secret that holds necessary credentials to access the database.
  • spec.type specifies the types of the app that this AppBinding is pointing to. KubeDB generated AppBinding follows the following format: <app group>/<app resource type>.

Insert Sample Data:

Now, we are going to exec into the any aggregator pod and create some sample data. At first, find out the database Pod using the following command,

$ kubectl get pods -n demo --selector="app.kubernetes.io/instance=sample-singlestore"
NAME                             READY    STATUS    RESTARTS   AGE
sample-singlestore-aggregator-0   2/2     Running   0          15m
sample-singlestore-aggregator-1   2/2     Running   0          15m
sample-singlestore-leaf-0         2/2     Running   0          15m
sample-singlestore-leaf-1         2/2     Running   0          15m
sample-singlestore-leaf-2         2/2     Running   0          15m

And copy the username and password of the root user to access into memsql shell.

$ kubectl get secret -n demo  sample-singlestore-root-cred -o jsonpath='{.data.username}'| base64 -d
root⏎           

kubectl get secret -n demo  sample-singlestore-root-cred -o jsonpath='{.data.password}'| base64 -d
xEJv73q3w_m1~H.G⏎ 

Now, Lets exec into the any aggregator Pod to enter into mysql shell and create a database and a table,

$ kubectl exec -it -n demo sample-singlestore-aggregator-0 -- singlestore --user=root --password=xEJv73q3w_m1~H.G
Defaulted container "singlestore" out of: singlestore, singlestore-coordinator, singlestore-init (init)
singlestore-client: [Warning] Using a password on the command line interface can be insecure.
Welcome to the MySQL monitor.  Commands end with ; or \g.
Your MySQL connection id is 300070
Server version: 5.7.32 SingleStoreDB source distribution (compatible; MySQL Enterprise & MySQL Commercial)

Copyright (c) 2000, 2022, Oracle and/or its affiliates.

Oracle is a registered trademark of Oracle Corporation and/or its
affiliates. Other names may be trademarks of their respective
owners.

Type 'help;' or '\h' for help. Type '\c' to clear the current input statement.

singlestore> CREATE DATABASE playground partitions 2;
Query OK, 1 row affected (0.01 sec)

singlestore> show databases;
+--------------------+
| Database           |
+--------------------+
| cluster            |
| demo               |
| det                |
| information_schema |
| memsql             |
| playground         |
| singlestore_health |
| test               |
+--------------------+
8 rows in set (0.00 sec)

singlestore> CREATE TABLE playground.equipment ( id INT NOT NULL AUTO_INCREMENT, type VARCHAR(50), quant INT, color VARCHAR(25), PRIMARY KEY(id));
Query OK, 0 rows affected, 1 warning (0.27 sec)

singlestore> SHOW TABLES IN playground;
+----------------------+
| Tables_in_playground |
+----------------------+
| equipment            |
+----------------------+
1 row in set (0.00 sec)

singlestore> INSERT INTO playground.equipment (type, quant, color) VALUES ("slide", 2, "blue");
Query OK, 1 row affected (1.15 sec)

singlestore> SELECT * FROM playground.equipment;
+----+-------+-------+-------+
| id | type  | quant | color |
+----+-------+-------+-------+
|  1 | slide |     2 | blue  |
+----+-------+-------+-------+
1 row in set (0.14 sec)

singlestore> exit
Bye

Now, we are ready to backup the database.

Prepare Backend

We are going to store our backed up data into a GCS bucket. We have to create a Secret with necessary credentials and a BackupStorage CR to use this backend. If you want to use a different backend, please read the respective backend configuration doc from here.

Create Secret:

Let’s create a secret called gcs-secret with access credentials to our desired GCS bucket,

$ echo -n '<your-project-id>' > GOOGLE_PROJECT_ID
$ cat /path/to/downloaded-sa-key.json > GOOGLE_SERVICE_ACCOUNT_JSON_KEY
$ kubectl create secret generic -n demo gcs-secret \
    --from-file=./GOOGLE_PROJECT_ID \
    --from-file=./GOOGLE_SERVICE_ACCOUNT_JSON_KEY
secret/gcs-secret created

Create BackupStorage:

Now, create a BackupStorage using this secret. Below is the YAML of BackupStorage CR we are going to create,

apiVersion: storage.kubestash.com/v1alpha1
kind: BackupStorage
metadata:
  name: gcs-storage
  namespace: demo
spec:
  storage:
    provider: gcs
    gcs:
      bucket: kubestash-qa
      prefix: demo
      secretName: gcs-secret
  usagePolicy:
    allowedNamespaces:
      from: All
  default: true
  deletionPolicy: Delete

Let’s create the BackupStorage we have shown above,

$ kubectl apply -f https://github.com/kubedb/docs/raw/v2024.9.30/docs/guides/singlestore/backup/kubestash/application-level/examples/backupstorage.yaml
backupstorage.storage.kubestash.com/gcs-storage created

Now, we are ready to backup our database to our desired backend.

Create RetentionPolicy:

Now, let’s create a RetentionPolicy to specify how the old Snapshots should be cleaned up.

Below is the YAML of the RetentionPolicy object that we are going to create,

apiVersion: storage.kubestash.com/v1alpha1
kind: RetentionPolicy
metadata:
  name: demo-retention
  namespace: demo
spec:
  default: true
  failedSnapshots:
    last: 2
  maxRetentionPeriod: 2mo
  successfulSnapshots:
    last: 5
  usagePolicy:
    allowedNamespaces:
      from: All

Let’s create the above RetentionPolicy,

$ kubectl apply -f https://github.com/kubedb/docs/raw/v2024.9.30/docs/guides/singlestore/backup/kubestash/application-level/examples/retentionpolicy.yaml
retentionpolicy.storage.kubestash.com/demo-retention created

Backup

We have to create a BackupConfiguration targeting respective sample-singlestore SingleStore database. Then, KubeStash will create a CronJob for each session to take periodic backup of that database.

At first, we need to create a secret with a Restic password for backup data encryption.

Create Secret:

Let’s create a secret called encrypt-secret with the Restic password,

$ echo -n 'changeit' > RESTIC_PASSWORD
$ kubectl create secret generic -n demo encrypt-secret \
    --from-file=./RESTIC_PASSWORD \
secret "encrypt-secret" created

Create BackupConfiguration:

Below is the YAML for BackupConfiguration CR to take application-level backup of the sample-singlestore database that we have deployed earlier,

apiVersion: core.kubestash.com/v1alpha1
kind: BackupConfiguration
metadata:
  name: sample-singlestore-backup
  namespace: demo
spec:
  target:
    apiGroup: kubedb.com
    kind: Singlestore
    namespace: demo
    name: sample-singlestore
  backends:
    - name: gcs-backend
      storageRef:
        namespace: demo
        name: gcs-storage
      retentionPolicy:
        name: demo-retention
        namespace: demo
  sessions:
    - name: frequent-backup
      scheduler:
        schedule: "*/5 * * * *"
        jobTemplate:
          backoffLimit: 1
      repositories:
        - name: gcs-singlestore-repo
          backend: gcs-backend
          directory: /singlestore
          encryptionSecret:
            name: encrypt-secret
            namespace: demo
      addon:
        name: singlestore-addon
        tasks:
          - name: manifest-backup
          - name: logical-backup
  • .spec.sessions[*].schedule specifies that we want to backup at 5 minutes interval.
  • .spec.target refers to the targeted sample-singlestore SingleStore database that we created earlier.
  • .spec.sessions[*].addon.tasks[*].name[*] specifies that both the manifest-backup and logical-backup tasks will be executed.

Let’s create the BackupConfiguration CR that we have shown above,

$ kubectl apply -f https://github.com/kubedb/docs/raw/v2024.9.30/docs/guides/singlestore/backup/kubestash/application-level/examples/backupconfiguration.yaml
backupconfiguration.core.kubestash.com/sample-singlestore-backup created

Verify Backup Setup Successful

If everything goes well, the phase of the BackupConfiguration should be Ready. The Ready phase indicates that the backup setup is successful. Let’s verify the Phase of the BackupConfiguration,

$ kubectl get backupconfiguration -n demo
NAME                        PHASE   PAUSED   AGE
sample-singlestore-backup   Ready            2m50s

Additionally, we can verify that the Repository specified in the BackupConfiguration has been created using the following command,

$ kubectl get repo -n demo
NAME                     INTEGRITY   SNAPSHOT-COUNT   SIZE     PHASE   LAST-SUCCESSFUL-BACKUP   AGE
gcs-singlestore-repo                 0                0 B      Ready                            3m

KubeStash keeps the backup for Repository YAMLs. If we navigate to the GCS bucket, we will see the Repository YAML stored in the demo/singlestore directory.

Verify CronJob:

It will also create a CronJob with the schedule specified in spec.sessions[*].scheduler.schedule field of BackupConfiguration CR.

Verify that the CronJob has been created using the following command,

$ kubectl get cronjob -n demo
NAME                                               SCHEDULE      SUSPEND   ACTIVE   LAST SCHEDULE   AGE
trigger-sample-singlestore-backup-frequent-backup   */5 * * * *             0        2m45s          3m25s

Verify BackupSession:

KubeStash triggers an instant backup as soon as the BackupConfiguration is ready. After that, backups are scheduled according to the specified schedule.

Run the following command to watch BackupSession CR,

$ kubectl get backupsession -n demo -w

NAME                                                      INVOKER-TYPE          INVOKER-NAME                 PHASE       DURATION   AGE
sample-singlestore-backup-frequent-backup-1724065200   BackupConfiguration   sample-singlestore-backup    Succeeded              7m22s

We can see from the above output that the backup session has succeeded. Now, we are going to verify whether the backed up data has been stored in the backend.

Verify Backup:

Once a backup is complete, KubeStash will update the respective Repository CR to reflect the backup. Check that the repository sample-singlestore-backup has been updated by the following command,

$ kubectl get repository -n demo gcs-singlestore-repo
NAME                       INTEGRITY   SNAPSHOT-COUNT   SIZE    PHASE   LAST-SUCCESSFUL-BACKUP   AGE
gcs-singlestore-repo          true        1             806 B   Ready   8m27s                    9m18s

At this moment we have one Snapshot. Run the following command to check the respective Snapshot which represents the state of a backup run for an application.

$ kubectl get snapshots -n demo -l=kubestash.com/repo-name=gcs-demo-repo
NAME                                                            REPOSITORY            SESSION           SNAPSHOT-TIME          DELETION-POLICY   PHASE       AGE
gcs-singlestore-repo-sample-singlestore-backup-frequent-backup-1725359100   sample-singlestore-backup   frequent-backup   2024-01-23T13:10:54Z   Delete            Succeeded   16h

Note: KubeStash creates a Snapshot with the following labels:

  • kubestash.com/app-ref-kind: <target-kind>
  • kubestash.com/app-ref-name: <target-name>
  • kubestash.com/app-ref-namespace: <target-namespace>
  • kubestash.com/repo-name: <repository-name>

These labels can be used to watch only the Snapshots related to our target Database or Repository.

If we check the YAML of the Snapshot, we can find the information about the backed up components of the Database.

$ kubectl get snapshots -n demo gcs-singlestore-repo-sample-singlestore-backup-frequent-backup-1725359100 -oyaml
apiVersion: storage.kubestash.com/v1alpha1
kind: Snapshot
metadata:
  creationTimestamp: "2024-09-03T10:25:00Z"
  finalizers:
  - kubestash.com/cleanup
  generation: 1
  labels:
    kubestash.com/app-ref-kind: Singlestore
    kubestash.com/app-ref-name: sample-singlestore
    kubestash.com/app-ref-namespace: demo
    kubestash.com/repo-name: gcs-singlestore-repo
  annotations:
    kubedb.com/db-version: 8.7.10
  name: gcs-singlestore-repo-sample-singlestore-backup-frequent-backup-1725359100
  namespace: demo
  ownerReferences:
  - apiVersion: storage.kubestash.com/v1alpha1
    blockOwnerDeletion: true
    controller: true
    kind: Repository
    name: gcs-singlestore-repo
    uid: 1f5ba355-7f99-4b99-8bbf-9f9d4f31c52a
  resourceVersion: "213010"
  uid: 18cabb10-e594-4655-8763-3daa0872508e
spec:
  appRef:
    apiGroup: kubedb.com
    kind: Singlestore
    name: sample-singlestore
    namespace: demo
  backupSession: sample-singlestore-backup-frequent-backup-1725359100
  deletionPolicy: Delete
  repository: gcs-singlestore-repo
  session: frequent-backup
  snapshotID: 01J6VPN4TPHDFT1M9Q9YVGMTKF
  type: FullBackup
  version: v1
status:
  components:
    dump:
      driver: Restic
      duration: 7.393324414s
      integrity: true
      path: repository/v1/frequent-backup/dump
      phase: Succeeded
      resticStats:
      - hostPath: dumpfile.sql
        id: f2ffd1bdb98563e15c46d8927d7239873ce7094132d959e12134688e06984736
        size: 3.657 MiB
        uploaded: 706.081 KiB
      size: 893.009 KiB
    manifest:
      driver: Restic
      duration: 12.672292995s
      integrity: true
      path: repository/v1/frequent-backup/manifest
      phase: Succeeded
      resticStats:
      - hostPath: /kubestash-tmp/manifest
        id: ff99eb7ea769a365f7cdc83a252df610c262fc934ec0a3475499bbbb35ca6931
        size: 2.883 KiB
        uploaded: 1.440 KiB
      size: 3.788 KiB
  conditions:
  - lastTransitionTime: "2024-09-03T10:25:00Z"
    message: Recent snapshot list updated successfully
    reason: SuccessfullyUpdatedRecentSnapshotList
    status: "True"
    type: RecentSnapshotListUpdated
  - lastTransitionTime: "2024-09-03T10:25:49Z"
    message: Metadata uploaded to backend successfully
    reason: SuccessfullyUploadedSnapshotMetadata
    status: "True"
    type: SnapshotMetadataUploaded
  integrity: true
  phase: Succeeded
  size: 896.796 KiB
  snapshotTime: "2024-09-03T10:25:00Z"
  totalComponents: 2

KubeStash uses the mysqldump command to take backups of target SingleStore databases. Therefore, the component name for logical backups is set as dump. KubeStash set component name as manifest for the manifest backup of SingleStore databases.

Now, if we navigate to the GCS bucket, we will see the backed up data stored in the demo/singlestore/repository/v1/frequent-backup/dump directory. KubeStash also keeps the backup for Snapshot YAMLs, which can be found in the demo/dep/snapshots directory.

Note: KubeStash stores all dumped data encrypted in the backup directory, meaning it remains unreadable until decrypted.

Restore

In this section, we are going to restore the entire database from the backup that we have taken in the previous section.

For this tutorial, we will restore the database in a separate namespace called dev.

First, create the namespace by running the following command:

$ kubectl create ns dev
namespace/dev created

Create RestoreSession:

We need to create a RestoreSession CR.

Below, is the contents of YAML file of the RestoreSession CR that we are going to create to restore the entire database.

apiVersion: core.kubestash.com/v1alpha1
kind: RestoreSession
metadata:
  name: restore-sample-singlestore
  namespace: demo
spec:
  manifestOptions:
    singlestore:
        restoreNamespace: dev
  dataSource:
    repository: gcs-singlestore-repo
    snapshot: latest
    encryptionSecret:
      name: encrypt-secret
      namespace: demo
  addon:
    name: singlestore-addon
    tasks:
      - name: logical-backup-restore
      - name: manifest-restore

Here,

  • .spec.manifestOptions.singlestore.db specifies whether to restore the DB manifest or not.
  • .spec.dataSource.repository specifies the Repository object that holds the backed up data.
  • .spec.dataSource.snapshot specifies to restore from latest Snapshot.
  • .spec.addon.tasks[*] specifies that both the manifest-restore and logical-backup-restore tasks.

Let’s create the RestoreSession CRD object we have shown above,

$ kubectl apply -f https://github.com/kubedb/docs/raw/v2024.9.30/docs/guides/singlestore/backup/kubestash/application-level/examples/restoresession.yaml
restoresession.core.kubestash.com/sample-singlestore-restore created

Once, you have created the RestoreSession object, KubeStash will create restore Job. Run the following command to watch the phase of the RestoreSession object,

$ watch kubectl get restoresession -n demo
Every 2.0s: kubectl get restores... AppsCode-PC-03: Wed Aug 21 10:44:05 2024

NAME             REPOSITORY        FAILURE-POLICY   PHASE       DURATION   AGE
sample-restore   gcs-demo-repo                      Succeeded   3s         53s

The Succeeded phase means that the restore process has been completed successfully.

Verify Restored SingleStore Manifest:

In this section, we will verify whether the desired SingleStore database manifest has been successfully applied to the cluster.

$ kubectl get singlestores.kubedb.com -n dev
NAME                  VERSION    STATUS   AGE
sample-singlestore    8.7.10     Ready    39m

The output confirms that the SingleStore database has been successfully created with the same configuration as it had at the time of backup.

Verify Restored Data:

In this section, we are going to verify whether the desired data has been restored successfully. We are going to connect to the database server and check whether the database and the table we created earlier in the original database are restored.

At first, check if the database has gone into Ready state by the following command,

$ kubectl get sdb -n dev sample-singlestore
NAME                   VERSION   STATUS  AGE
sample-singlestore     8.7.10     Ready   4m

Now, find out the database Pod by the following command,

$ kubectl get pods -n demo --selector="app.kubernetes.io/instance=sample-singlestore"
NAME                             READY    STATUS    RESTARTS   AGE
sample-singlestore-aggregator-0   2/2     Running   0          15m
sample-singlestore-aggregator-1   2/2     Running   0          15m
sample-singlestore-leaf-0         2/2     Running   0          15m
sample-singlestore-leaf-1         2/2     Running   0          15m
sample-singlestore-leaf-2         2/2     Running   0          15m

And copy the username and password of the root user to access into mysql shell.

$ kubectl get secret -n demo  sample-singlestore-root-cred -o jsonpath='{.data.username}'| base64 -d
root⏎           

kubectl get secret -n demo  sample-singlestore-root-cred -o jsonpath='{.data.password}'| base64 -d
xEJv73q3w_m1~H.G⏎ 

Now, Lets exec into the any aggregator Pod to enter into mysql shell and create a database and a table,

$ kubectl exec -it -n demo sample-singlestore-aggregator-0 -- singlestore --user=root --password=xEJv73q3w_m1~H.G
Defaulted container "singlestore" out of: singlestore, singlestore-coordinator, singlestore-init (init)
singlestore-client: [Warning] Using a password on the command line interface can be insecure.
Welcome to the MySQL monitor.  Commands end with ; or \g.
Your MySQL connection id is 300070
Server version: 5.7.32 SingleStoreDB source distribution (compatible; MySQL Enterprise & MySQL Commercial)

Copyright (c) 2000, 2022, Oracle and/or its affiliates.

Oracle is a registered trademark of Oracle Corporation and/or its
affiliates. Other names may be trademarks of their respective
owners.

Type 'help;' or '\h' for help. Type '\c' to clear the current input statement.

singlestore> show databases;
+--------------------+
| Database           |
+--------------------+
| cluster            |
| demo               |
| det                |
| information_schema |
| memsql             |
| playground         |
| singlestore_health |
| test               |
+--------------------+
8 rows in set (0.00 sec)

singlestore> SHOW TABLES IN playground;
+----------------------+
| Tables_in_playground |
+----------------------+
| equipment            |
+----------------------+
1 row in set (0.00 sec)

singlestore> SELECT * FROM playground.equipment;
+----+-------+-------+-------+
| id | type  | quant | color |
+----+-------+-------+-------+
|  1 | slide |     2 | blue  |
+----+-------+-------+-------+
1 row in set (0.14 sec)

singlestore> exit
Bye

So, from the above output, we can see that the playground database and the equipment table we have created earlier in the original database and now, they are restored successfully.

Cleanup

To cleanup the Kubernetes resources created by this tutorial, run:

kubectl delete backupconfigurations.core.kubestash.com  -n demo sample-singlestore-backup
kubectl delete backupstorage -n demo gcs-storage
kubectl delete secret -n demo gcs-secret
kubectl delete secret -n demo encrypt-secret
kubectl delete retentionpolicies.storage.kubestash.com -n demo demo-retention
kubectl delete restoresessions.core.kubestash.com -n demo restore-sample-singlestore
kubectl delete my -n demo sample-singlestore
kubectl delete my -n dev sample-singlestore