You are looking at the documentation of a prior release. To read the documentation of the latest release, please
visit here.
Application Level Backup and Restore SingleStore database using KubeStash
KubeStash offers application-level backup and restore functionality for SingleStore
databases. It captures both manifest and logical data backups of any SingleStore
database in a single snapshot. During the restore process, KubeStash first applies the SingleStore
manifest to the cluster and then restores the data into it.
This guide will give you how you can take application-level backup and restore your SingleStore
databases using Kubestash
.
Before You Begin
- At first, you need to have a Kubernetes cluster, and the
kubectl
command-line tool must be configured to communicate with your cluster. If you do not already have a cluster, you can create one by usingMinikube
orKind
. - Install
KubeDB
in your cluster following the steps here. - Install
KubeStash
in your cluster following the steps here. - Install KubeStash
kubectl
plugin following the steps here. - If you are not familiar with how KubeStash backup and restore SingleStore databases, please check the following guide here.
You should be familiar with the following KubeStash
concepts:
To keep everything isolated, we are going to use a separate namespace called demo
throughout this tutorial.
$ kubectl create ns demo
namespace/demo created
Note: YAML files used in this tutorial are stored in docs/guides/singlestore/backup/kubestash/application-level/examples directory of kubedb/docs repository.
Backup SingleStore
KubeStash supports backups for SingleStore
instances across different configurations, including Standalone, and Cluster setups. In this demonstration, we’ll focus on a SingleStore
database using singlestore clustering. The backup and restore process is similar for Standalone and Cluster configurations as well.
This section will demonstrate how to take application-level backup of a SingleStore
database. Here, we are going to deploy a SingleStore
database using KubeDB. Then, we are going to back up the database at the application level to a GCS
bucket. Finally, we will restore the entire SingleStore
database.
Create SingleStore License Secret
We need SingleStore License to create SingleStore Database. So, Ensure that you have acquired a license and then simply pass the license by secret.
$ kubectl create secret generic -n demo license-secret \
--from-literal=username=license \
--from-literal=password='your-license-set-here'
secret/license-secret created
Deploy Sample SingleStore Database
Let’s deploy a sample SingleStore
database and insert some data into it.
Create SingleStore CR:
Below is the YAML of a sample SingleStore
CR that we are going to create for this tutorial:
apiVersion: kubedb.com/v1alpha2
kind: Singlestore
metadata:
name: sample-singlestore
namespace: demo
spec:
version: "8.7.10"
topology:
aggregator:
replicas: 2
podTemplate:
spec:
containers:
- name: singlestore
resources:
limits:
memory: "2Gi"
cpu: "600m"
requests:
memory: "2Gi"
cpu: "600m"
storage:
accessModes:
- ReadWriteOnce
resources:
requests:
storage: 1Gi
leaf:
replicas: 3
podTemplate:
spec:
containers:
- name: singlestore
resources:
limits:
memory: "2Gi"
cpu: "600m"
requests:
memory: "2Gi"
cpu: "600m"
storage:
accessModes:
- ReadWriteOnce
resources:
requests:
storage: 10Gi
licenseSecret:
name: license-secret
storageType: Durable
deletionPolicy: WipeOut
Here,
spec.version
is the name of the SinglestoreVersion CRD where the docker images are specified. In this tutorial, a SingleStore8.7.10
database is going to be created.spec.topology
specifies that it will be used as cluster mode. If this field is nil it will be work as standalone mode.spec.topology.aggregator.replicas
orspec.topology.leaf.replicas
specifies that the number replicas that will be used for aggregator or leaf.spec.storageType
specifies the type of storage that will be used for SingleStore database. It can beDurable
orEphemeral
. Default value of this field isDurable
. IfEphemeral
is used then KubeDB will create SingleStore database usingEmptyDir
volume. In this case, you don’t have to specifyspec.storage
field. This is useful for testing purposes.spec.topology.aggregator.storage
orspec.topology.leaf.storage
specifies the StorageClass of PVC dynamically allocated to store data for this database. This storage spec will be passed to the PetSet created by KubeDB operator to run database pods. You can specify any StorageClass available in your cluster with appropriate resource requests.spec.deletionPolicy
gives flexibility whether tonullify
(reject) the delete operation ofSinglestore
crd or which resources KubeDB should keep or delete when you deleteSinglestore
crd. If admission webhook is enabled, It prevents users from deleting the database as long as thespec.deletionPolicy
is set toDoNotTerminate
. Learn details of allDeletionPolicy
here
Note:
spec.storage
section is used to create PVC for database pod. It will create PVC with storage size specified instorage.resources.requests
field. Don’t specify limits here. PVC does not get resized automatically.
Create the above SingleStore
CR,
$ kubectl apply -f https://github.com/kubedb/docs/raw/v2024.11.18/docs/guides/singlestore/backup/kubestash/application-level/examples/sample-singlestore.yaml
singlestore.kubedb.com/sample-singlestore created
KubeDB will deploy a SingleStore database according to the above specification. It will also create the necessary Secrets and Services to access the database.
Let’s check if the database is ready to use,
$ kubectl get singlestores.kubedb.com -n demo
NAME VERSION STATUS AGE
sample-singlestore 8.7.10 Ready 4m22s
The database is Ready
. Verify that KubeDB has created a Secret
and a Service
for this database using the following commands,
$ kubectl get secret -n demo -l=app.kubernetes.io/instance=sample-singlestore
NAME TYPE DATA AGE
sample-singlestore-root-cred kubernetes.io/basic-auth 2 4m58s
$ kubectl get service -n demo -l=app.kubernetes.io/instance=sample-singlestore
NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE
sample-singlestore ClusterIP 10.128.230.168 <none> 3306/TCP,8081/TCP 5m10s
sample-singlestore-pods ClusterIP None <none> 3306/TCP 5m10s
Here, we have to use service sample-singlestore
and secret sample-singlestore-root-cred
to connect with the database. KubeDB
creates an AppBinding CR that holds the necessary information to connect with the database.
Verify AppBinding:
Verify that the AppBinding
has been created successfully using the following command,
$ kubectl get appbindings -n demo
NAME AGE
sample-singlestore 9m24s
Let’s check the YAML of the above AppBinding
,
$ kubectl get appbindings -n demo sample-singlestore -o yaml
apiVersion: appcatalog.appscode.com/v1alpha1
kind: AppBinding
metadata:
annotations:
kubectl.kubernetes.io/last-applied-configuration: |
{"apiVersion":"kubedb.com/v1alpha2","kind":"Singlestore","metadata":{"annotations":{},"name":"sample-singlestore","namespace":"demo"},"spec":{"deletionPolicy":"WipeOut","licenseSecret":{"name":"license-secret"},"storageType":"Durable","topology":{"aggregator":{"podTemplate":{"spec":{"containers":[{"name":"singlestore","resources":{"limits":{"cpu":"0.6","memory":"2Gi"},"requests":{"cpu":"0.6","memory":"2Gi"}}}]}},"replicas":2,"storage":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"1Gi"}}}},"leaf":{"podTemplate":{"spec":{"containers":[{"name":"singlestore","resources":{"limits":{"cpu":"0.6","memory":"2Gi"},"requests":{"cpu":"0.6","memory":"2Gi"}}}]}},"replicas":3,"storage":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"10Gi"}}}}},"version":"8.7.10"}}
creationTimestamp: "2024-09-11T07:03:44Z"
generation: 1
labels:
app.kubernetes.io/component: database
app.kubernetes.io/instance: sample-singlestor
app.kubernetes.io/managed-by: kubedb.com
app.kubernetes.io/name: singlestores.kubedb.com
name: sample-singlestore
namespace: demo
ownerReferences:
- apiVersion: kubedb.com/v1alpha2
blockOwnerDeletion: true
controller: true
kind: Singlestore
name: sample-singlestore
uid: e08e1f37-d869-437d-9b15-14c6aef3f406
resourceVersion: "4904220"
uid: 92b2b318-1874-4471-97ec-d789c6e16809
spec:
appRef:
apiGroup: kubedb.com
kind: Singlestore
name: sample-singlestore
namespace: demo
clientConfig:
service:
name: sample-singlestore
path: /
port: 3306
scheme: tcp
url: tcp(sample-singlestore.demo.svc:3306)/
parameters:
apiVersion: config.kubedb.com/v1alpha1
kind: SinglestoreConfiguration
masterAggregator: sample-singlestore-aggregator-0.sample-singlestore-pods.demo.svc
stash:
addon:
backupTask:
name: ""
restoreTask:
name: ""
secret:
name: sample-singlestore-root-cred
type: kubedb.com/singlestore
version: 8.7.10
KubeStash uses the AppBinding
CR to connect with the target database. It requires the following two fields to set in AppBinding’s .spec
section.
.spec.parameters.masterAggregator
specifies the dns of master aggregator node that we have to mention in mysqldump command when taken backup or restore..spec.clientConfig.service.name
specifies the name of the Service that connects to the database..spec.secret
specifies the name of the Secret that holds necessary credentials to access the database.spec.type
specifies the types of the app that this AppBinding is pointing to. KubeDB generated AppBinding follows the following format:<app group>/<app resource type>
.
Insert Sample Data:
Now, we are going to exec into the any aggregator pod and create some sample data. At first, find out the database Pod
using the following command,
$ kubectl get pods -n demo --selector="app.kubernetes.io/instance=sample-singlestore"
NAME READY STATUS RESTARTS AGE
sample-singlestore-aggregator-0 2/2 Running 0 15m
sample-singlestore-aggregator-1 2/2 Running 0 15m
sample-singlestore-leaf-0 2/2 Running 0 15m
sample-singlestore-leaf-1 2/2 Running 0 15m
sample-singlestore-leaf-2 2/2 Running 0 15m
And copy the username and password of the root
user to access into memsql
shell.
$ kubectl get secret -n demo sample-singlestore-root-cred -o jsonpath='{.data.username}'| base64 -d
root⏎
kubectl get secret -n demo sample-singlestore-root-cred -o jsonpath='{.data.password}'| base64 -d
xEJv73q3w_m1~H.G⏎
Now, Lets exec into the any aggregator Pod
to enter into mysql
shell and create a database and a table,
$ kubectl exec -it -n demo sample-singlestore-aggregator-0 -- singlestore --user=root --password=xEJv73q3w_m1~H.G
Defaulted container "singlestore" out of: singlestore, singlestore-coordinator, singlestore-init (init)
singlestore-client: [Warning] Using a password on the command line interface can be insecure.
Welcome to the MySQL monitor. Commands end with ; or \g.
Your MySQL connection id is 300070
Server version: 5.7.32 SingleStoreDB source distribution (compatible; MySQL Enterprise & MySQL Commercial)
Copyright (c) 2000, 2022, Oracle and/or its affiliates.
Oracle is a registered trademark of Oracle Corporation and/or its
affiliates. Other names may be trademarks of their respective
owners.
Type 'help;' or '\h' for help. Type '\c' to clear the current input statement.
singlestore> CREATE DATABASE playground partitions 2;
Query OK, 1 row affected (0.01 sec)
singlestore> show databases;
+--------------------+
| Database |
+--------------------+
| cluster |
| demo |
| det |
| information_schema |
| memsql |
| playground |
| singlestore_health |
| test |
+--------------------+
8 rows in set (0.00 sec)
singlestore> CREATE TABLE playground.equipment ( id INT NOT NULL AUTO_INCREMENT, type VARCHAR(50), quant INT, color VARCHAR(25), PRIMARY KEY(id));
Query OK, 0 rows affected, 1 warning (0.27 sec)
singlestore> SHOW TABLES IN playground;
+----------------------+
| Tables_in_playground |
+----------------------+
| equipment |
+----------------------+
1 row in set (0.00 sec)
singlestore> INSERT INTO playground.equipment (type, quant, color) VALUES ("slide", 2, "blue");
Query OK, 1 row affected (1.15 sec)
singlestore> SELECT * FROM playground.equipment;
+----+-------+-------+-------+
| id | type | quant | color |
+----+-------+-------+-------+
| 1 | slide | 2 | blue |
+----+-------+-------+-------+
1 row in set (0.14 sec)
singlestore> exit
Bye
Now, we are ready to backup the database.
Prepare Backend
We are going to store our backed up data into a GCS bucket. We have to create a Secret with necessary credentials and a BackupStorage
CR to use this backend. If you want to use a different backend, please read the respective backend configuration doc from here.
Create Secret:
Let’s create a secret called gcs-secret
with access credentials to our desired GCS bucket,
$ echo -n '<your-project-id>' > GOOGLE_PROJECT_ID
$ cat /path/to/downloaded-sa-key.json > GOOGLE_SERVICE_ACCOUNT_JSON_KEY
$ kubectl create secret generic -n demo gcs-secret \
--from-file=./GOOGLE_PROJECT_ID \
--from-file=./GOOGLE_SERVICE_ACCOUNT_JSON_KEY
secret/gcs-secret created
Create BackupStorage:
Now, create a BackupStorage
using this secret. Below is the YAML of BackupStorage
CR we are going to create,
apiVersion: storage.kubestash.com/v1alpha1
kind: BackupStorage
metadata:
name: gcs-storage
namespace: demo
spec:
storage:
provider: gcs
gcs:
bucket: kubestash-qa
prefix: demo
secretName: gcs-secret
usagePolicy:
allowedNamespaces:
from: All
default: true
deletionPolicy: Delete
Let’s create the BackupStorage we have shown above,
$ kubectl apply -f https://github.com/kubedb/docs/raw/v2024.11.18/docs/guides/singlestore/backup/kubestash/application-level/examples/backupstorage.yaml
backupstorage.storage.kubestash.com/gcs-storage created
Now, we are ready to backup our database to our desired backend.
Create RetentionPolicy:
Now, let’s create a RetentionPolicy
to specify how the old Snapshots should be cleaned up.
Below is the YAML of the RetentionPolicy
object that we are going to create,
apiVersion: storage.kubestash.com/v1alpha1
kind: RetentionPolicy
metadata:
name: demo-retention
namespace: demo
spec:
default: true
failedSnapshots:
last: 2
maxRetentionPeriod: 2mo
successfulSnapshots:
last: 5
usagePolicy:
allowedNamespaces:
from: All
Let’s create the above RetentionPolicy
,
$ kubectl apply -f https://github.com/kubedb/docs/raw/v2024.11.18/docs/guides/singlestore/backup/kubestash/application-level/examples/retentionpolicy.yaml
retentionpolicy.storage.kubestash.com/demo-retention created
Backup
We have to create a BackupConfiguration
targeting respective sample-singlestore
SingleStore database. Then, KubeStash will create a CronJob
for each session to take periodic backup of that database.
At first, we need to create a secret with a Restic password for backup data encryption.
Create Secret:
Let’s create a secret called encrypt-secret
with the Restic password,
$ echo -n 'changeit' > RESTIC_PASSWORD
$ kubectl create secret generic -n demo encrypt-secret \
--from-file=./RESTIC_PASSWORD \
secret "encrypt-secret" created
Create BackupConfiguration:
Below is the YAML for BackupConfiguration
CR to take application-level backup of the sample-singlestore
database that we have deployed earlier,
apiVersion: core.kubestash.com/v1alpha1
kind: BackupConfiguration
metadata:
name: sample-singlestore-backup
namespace: demo
spec:
target:
apiGroup: kubedb.com
kind: Singlestore
namespace: demo
name: sample-singlestore
backends:
- name: gcs-backend
storageRef:
namespace: demo
name: gcs-storage
retentionPolicy:
name: demo-retention
namespace: demo
sessions:
- name: frequent-backup
scheduler:
schedule: "*/5 * * * *"
jobTemplate:
backoffLimit: 1
repositories:
- name: gcs-singlestore-repo
backend: gcs-backend
directory: /singlestore
encryptionSecret:
name: encrypt-secret
namespace: demo
addon:
name: singlestore-addon
tasks:
- name: manifest-backup
- name: logical-backup
.spec.sessions[*].schedule
specifies that we want to backup at5 minutes
interval..spec.target
refers to the targetedsample-singlestore
SingleStore database that we created earlier..spec.sessions[*].addon.tasks[*].name[*]
specifies that both themanifest-backup
andlogical-backup
tasks will be executed.
Let’s create the BackupConfiguration
CR that we have shown above,
$ kubectl apply -f https://github.com/kubedb/docs/raw/v2024.11.18/docs/guides/singlestore/backup/kubestash/application-level/examples/backupconfiguration.yaml
backupconfiguration.core.kubestash.com/sample-singlestore-backup created
Verify Backup Setup Successful
If everything goes well, the phase of the BackupConfiguration
should be Ready
. The Ready
phase indicates that the backup setup is successful. Let’s verify the Phase
of the BackupConfiguration,
$ kubectl get backupconfiguration -n demo
NAME PHASE PAUSED AGE
sample-singlestore-backup Ready 2m50s
Additionally, we can verify that the Repository
specified in the BackupConfiguration
has been created using the following command,
$ kubectl get repo -n demo
NAME INTEGRITY SNAPSHOT-COUNT SIZE PHASE LAST-SUCCESSFUL-BACKUP AGE
gcs-singlestore-repo 0 0 B Ready 3m
KubeStash keeps the backup for Repository
YAMLs. If we navigate to the GCS bucket, we will see the Repository
YAML stored in the demo/singlestore
directory.
Verify CronJob:
It will also create a CronJob
with the schedule specified in spec.sessions[*].scheduler.schedule
field of BackupConfiguration
CR.
Verify that the CronJob
has been created using the following command,
$ kubectl get cronjob -n demo
NAME SCHEDULE SUSPEND ACTIVE LAST SCHEDULE AGE
trigger-sample-singlestore-backup-frequent-backup */5 * * * * 0 2m45s 3m25s
Verify BackupSession:
KubeStash triggers an instant backup as soon as the BackupConfiguration
is ready. After that, backups are scheduled according to the specified schedule.
Run the following command to watch BackupSession
CR,
$ kubectl get backupsession -n demo -w
NAME INVOKER-TYPE INVOKER-NAME PHASE DURATION AGE
sample-singlestore-backup-frequent-backup-1724065200 BackupConfiguration sample-singlestore-backup Succeeded 7m22s
We can see from the above output that the backup session has succeeded. Now, we are going to verify whether the backed up data has been stored in the backend.
Verify Backup:
Once a backup is complete, KubeStash will update the respective Repository
CR to reflect the backup. Check that the repository sample-singlestore-backup
has been updated by the following command,
$ kubectl get repository -n demo gcs-singlestore-repo
NAME INTEGRITY SNAPSHOT-COUNT SIZE PHASE LAST-SUCCESSFUL-BACKUP AGE
gcs-singlestore-repo true 1 806 B Ready 8m27s 9m18s
At this moment we have one Snapshot
. Run the following command to check the respective Snapshot
which represents the state of a backup run for an application.
$ kubectl get snapshots -n demo -l=kubestash.com/repo-name=gcs-demo-repo
NAME REPOSITORY SESSION SNAPSHOT-TIME DELETION-POLICY PHASE AGE
gcs-singlestore-repo-sample-singlestore-backup-frequent-backup-1725359100 sample-singlestore-backup frequent-backup 2024-01-23T13:10:54Z Delete Succeeded 16h
Note: KubeStash creates a
Snapshot
with the following labels:
kubestash.com/app-ref-kind: <target-kind>
kubestash.com/app-ref-name: <target-name>
kubestash.com/app-ref-namespace: <target-namespace>
kubestash.com/repo-name: <repository-name>
These labels can be used to watch only the
Snapshot
s related to our target Database orRepository
.
If we check the YAML of the Snapshot
, we can find the information about the backed up components of the Database.
$ kubectl get snapshots -n demo gcs-singlestore-repo-sample-singlestore-backup-frequent-backup-1725359100 -oyaml
apiVersion: storage.kubestash.com/v1alpha1
kind: Snapshot
metadata:
creationTimestamp: "2024-09-03T10:25:00Z"
finalizers:
- kubestash.com/cleanup
generation: 1
labels:
kubestash.com/app-ref-kind: Singlestore
kubestash.com/app-ref-name: sample-singlestore
kubestash.com/app-ref-namespace: demo
kubestash.com/repo-name: gcs-singlestore-repo
annotations:
kubedb.com/db-version: 8.7.10
name: gcs-singlestore-repo-sample-singlestore-backup-frequent-backup-1725359100
namespace: demo
ownerReferences:
- apiVersion: storage.kubestash.com/v1alpha1
blockOwnerDeletion: true
controller: true
kind: Repository
name: gcs-singlestore-repo
uid: 1f5ba355-7f99-4b99-8bbf-9f9d4f31c52a
resourceVersion: "213010"
uid: 18cabb10-e594-4655-8763-3daa0872508e
spec:
appRef:
apiGroup: kubedb.com
kind: Singlestore
name: sample-singlestore
namespace: demo
backupSession: sample-singlestore-backup-frequent-backup-1725359100
deletionPolicy: Delete
repository: gcs-singlestore-repo
session: frequent-backup
snapshotID: 01J6VPN4TPHDFT1M9Q9YVGMTKF
type: FullBackup
version: v1
status:
components:
dump:
driver: Restic
duration: 7.393324414s
integrity: true
path: repository/v1/frequent-backup/dump
phase: Succeeded
resticStats:
- hostPath: dumpfile.sql
id: f2ffd1bdb98563e15c46d8927d7239873ce7094132d959e12134688e06984736
size: 3.657 MiB
uploaded: 706.081 KiB
size: 893.009 KiB
manifest:
driver: Restic
duration: 12.672292995s
integrity: true
path: repository/v1/frequent-backup/manifest
phase: Succeeded
resticStats:
- hostPath: /kubestash-tmp/manifest
id: ff99eb7ea769a365f7cdc83a252df610c262fc934ec0a3475499bbbb35ca6931
size: 2.883 KiB
uploaded: 1.440 KiB
size: 3.788 KiB
conditions:
- lastTransitionTime: "2024-09-03T10:25:00Z"
message: Recent snapshot list updated successfully
reason: SuccessfullyUpdatedRecentSnapshotList
status: "True"
type: RecentSnapshotListUpdated
- lastTransitionTime: "2024-09-03T10:25:49Z"
message: Metadata uploaded to backend successfully
reason: SuccessfullyUploadedSnapshotMetadata
status: "True"
type: SnapshotMetadataUploaded
integrity: true
phase: Succeeded
size: 896.796 KiB
snapshotTime: "2024-09-03T10:25:00Z"
totalComponents: 2
KubeStash uses the
mysqldump
command to take backups of target SingleStore databases. Therefore, the component name for logical backups is set asdump
. KubeStash set component name asmanifest
for themanifest backup
of SingleStore databases.
Now, if we navigate to the GCS bucket, we will see the backed up data stored in the demo/singlestore/repository/v1/frequent-backup/dump
directory. KubeStash also keeps the backup for Snapshot
YAMLs, which can be found in the demo/dep/snapshots
directory.
Note: KubeStash stores all dumped data encrypted in the backup directory, meaning it remains unreadable until decrypted.
Restore
In this section, we are going to restore the entire database from the backup that we have taken in the previous section.
For this tutorial, we will restore the database in a separate namespace called dev
.
First, create the namespace by running the following command:
$ kubectl create ns dev
namespace/dev created
Create RestoreSession:
We need to create a RestoreSession CR.
Below, is the contents of YAML file of the RestoreSession
CR that we are going to create to restore the entire database.
apiVersion: core.kubestash.com/v1alpha1
kind: RestoreSession
metadata:
name: restore-sample-singlestore
namespace: demo
spec:
manifestOptions:
singlestore:
restoreNamespace: dev
dataSource:
repository: gcs-singlestore-repo
snapshot: latest
encryptionSecret:
name: encrypt-secret
namespace: demo
addon:
name: singlestore-addon
tasks:
- name: logical-backup-restore
- name: manifest-restore
Here,
.spec.manifestOptions.singlestore.db
specifies whether to restore the DB manifest or not..spec.dataSource.repository
specifies the Repository object that holds the backed up data..spec.dataSource.snapshot
specifies to restore from latestSnapshot
..spec.addon.tasks[*]
specifies that both themanifest-restore
andlogical-backup-restore
tasks.
Let’s create the RestoreSession CRD object we have shown above,
$ kubectl apply -f https://github.com/kubedb/docs/raw/v2024.11.18/docs/guides/singlestore/backup/kubestash/application-level/examples/restoresession.yaml
restoresession.core.kubestash.com/sample-singlestore-restore created
Once, you have created the RestoreSession
object, KubeStash will create restore Job. Run the following command to watch the phase of the RestoreSession
object,
$ watch kubectl get restoresession -n demo
Every 2.0s: kubectl get restores... AppsCode-PC-03: Wed Aug 21 10:44:05 2024
NAME REPOSITORY FAILURE-POLICY PHASE DURATION AGE
sample-restore gcs-demo-repo Succeeded 3s 53s
The Succeeded
phase means that the restore process has been completed successfully.
Verify Restored SingleStore Manifest:
In this section, we will verify whether the desired SingleStore
database manifest has been successfully applied to the cluster.
$ kubectl get singlestores.kubedb.com -n dev
NAME VERSION STATUS AGE
sample-singlestore 8.7.10 Ready 39m
The output confirms that the SingleStore
database has been successfully created with the same configuration as it had at the time of backup.
Verify Restored Data:
In this section, we are going to verify whether the desired data has been restored successfully. We are going to connect to the database server and check whether the database and the table we created earlier in the original database are restored.
At first, check if the database has gone into Ready
state by the following command,
$ kubectl get sdb -n dev sample-singlestore
NAME VERSION STATUS AGE
sample-singlestore 8.7.10 Ready 4m
Now, find out the database Pod
by the following command,
$ kubectl get pods -n demo --selector="app.kubernetes.io/instance=sample-singlestore"
NAME READY STATUS RESTARTS AGE
sample-singlestore-aggregator-0 2/2 Running 0 15m
sample-singlestore-aggregator-1 2/2 Running 0 15m
sample-singlestore-leaf-0 2/2 Running 0 15m
sample-singlestore-leaf-1 2/2 Running 0 15m
sample-singlestore-leaf-2 2/2 Running 0 15m
And copy the username and password of the root
user to access into mysql
shell.
$ kubectl get secret -n demo sample-singlestore-root-cred -o jsonpath='{.data.username}'| base64 -d
root⏎
kubectl get secret -n demo sample-singlestore-root-cred -o jsonpath='{.data.password}'| base64 -d
xEJv73q3w_m1~H.G⏎
Now, Lets exec into the any aggregator Pod
to enter into mysql
shell and create a database and a table,
$ kubectl exec -it -n demo sample-singlestore-aggregator-0 -- singlestore --user=root --password=xEJv73q3w_m1~H.G
Defaulted container "singlestore" out of: singlestore, singlestore-coordinator, singlestore-init (init)
singlestore-client: [Warning] Using a password on the command line interface can be insecure.
Welcome to the MySQL monitor. Commands end with ; or \g.
Your MySQL connection id is 300070
Server version: 5.7.32 SingleStoreDB source distribution (compatible; MySQL Enterprise & MySQL Commercial)
Copyright (c) 2000, 2022, Oracle and/or its affiliates.
Oracle is a registered trademark of Oracle Corporation and/or its
affiliates. Other names may be trademarks of their respective
owners.
Type 'help;' or '\h' for help. Type '\c' to clear the current input statement.
singlestore> show databases;
+--------------------+
| Database |
+--------------------+
| cluster |
| demo |
| det |
| information_schema |
| memsql |
| playground |
| singlestore_health |
| test |
+--------------------+
8 rows in set (0.00 sec)
singlestore> SHOW TABLES IN playground;
+----------------------+
| Tables_in_playground |
+----------------------+
| equipment |
+----------------------+
1 row in set (0.00 sec)
singlestore> SELECT * FROM playground.equipment;
+----+-------+-------+-------+
| id | type | quant | color |
+----+-------+-------+-------+
| 1 | slide | 2 | blue |
+----+-------+-------+-------+
1 row in set (0.14 sec)
singlestore> exit
Bye
So, from the above output, we can see that the playground
database and the equipment
table we have created earlier in the original database and now, they are restored successfully.
Cleanup
To cleanup the Kubernetes resources created by this tutorial, run:
kubectl delete backupconfigurations.core.kubestash.com -n demo sample-singlestore-backup
kubectl delete backupstorage -n demo gcs-storage
kubectl delete secret -n demo gcs-secret
kubectl delete secret -n demo encrypt-secret
kubectl delete retentionpolicies.storage.kubestash.com -n demo demo-retention
kubectl delete restoresessions.core.kubestash.com -n demo restore-sample-singlestore
kubectl delete my -n demo sample-singlestore
kubectl delete my -n dev sample-singlestore