You are looking at the documentation of a prior release. To read the documentation of the latest release, please
visit here.
New to KubeDB? Please start here.
Horizontal Scale MongoDB Replicaset
This guide will show you how to use KubeDB
Enterprise operator to scale the replicaset of a MongoDB database.
Before You Begin
At first, you need to have a Kubernetes cluster, and the
kubectl
command-line tool must be configured to communicate with your cluster. If you do not already have a cluster, you can create one by using kind.Install
KubeDB
Community and Enterprise operator in your cluster following the steps here.You should be familiar with the following
KubeDB
concepts:
To keep everything isolated, we are going to use a separate namespace called demo
throughout this tutorial.
$ kubectl create ns demo
namespace/demo created
Note: YAML files used in this tutorial are stored in docs/examples/mongodb directory of kubedb/docs repository.
Apply Horizontal Scaling on Replicaset
Here, we are going to deploy a MongoDB
replicaset using a supported version by KubeDB
operator. Then we are going to apply horizontal scaling on it.
Prepare MongoDB Replicaset Database
Now, we are going to deploy a MongoDB
replicaset database with version 3.6.8
.
Deploy MongoDB replicaset
In this section, we are going to deploy a MongoDB replicaset database. Then, in the next section we will scale the database using MongoDBOpsRequest
CRD. Below is the YAML of the MongoDB
CR that we are going to create,
apiVersion: kubedb.com/v1alpha2
kind: MongoDB
metadata:
name: mg-replicaset
namespace: demo
spec:
version: "3.6.8-v1"
replicaSet:
name: "replicaset"
replicas: 3
storageType: Durable
storage:
storageClassName: "standard"
accessModes:
- ReadWriteOnce
resources:
requests:
storage: 1Gi
Let’s create the MongoDB
CR we have shown above,
$ kubectl create -f https://github.com/kubedb/docs/raw/v2020.11.12/docs/examples/mongodb/scaling/mg-replicaset.yaml
mongodb.kubedb.com/mg-replicaset created
Now, wait until mg-replicaset
has status Running
. i.e,
$ kubectl get mg -n demo 20:05:47
NAME VERSION STATUS AGE
mg-replicaset 3.6.8-v1 Running 2m36s
Let’s check the number of replicas this database has from the MongoDB object, number of pods the statefulset have,
$ kubectl get mongodb -n demo mg-replicaset -o json | jq '.spec.replicas' 11:02:09
3
$ kubectl get sts -n demo mg-replicaset -o json | jq '.spec.replicas' 11:03:27
3
We can see from both command that the database has 3 replicas in the replicaset.
Also, we can verify the replicas of the replicaset from an internal mongodb command by execing into a replica.
First we need to get the username and password to connect to a mongodb instance,
$ kubectl get secrets -n demo mg-replicaset-auth -o jsonpath='{.data.\username}' | base64 -d 11:09:51
root
$ kubectl get secrets -n demo mg-replicaset-auth -o jsonpath='{.data.\password}' | base64 -d 11:10:44
nrKuxni0wDSMrgwy
Now let’s connect to a mongodb instance and run a mongodb internal command to check the number of replicas,
$ kubectl exec -n demo mg-replicaset-0 -- mongo admin -u root -p nrKuxni0wDSMrgwy --eval "db.adminCommand( { replSetGetStatus : 1 } ).members" --quiet
[
{
"_id" : 0,
"name" : "mg-replicaset-0.mg-replicaset-gvr.demo.svc.cluster.local:27017",
"health" : 1,
"state" : 1,
"stateStr" : "PRIMARY",
"uptime" : 631,
"optime" : {
"ts" : Timestamp(1598418585, 1),
"t" : NumberLong(2)
},
"optimeDate" : ISODate("2020-08-26T05:09:45Z"),
"syncingTo" : "",
"syncSourceHost" : "",
"syncSourceId" : -1,
"infoMessage" : "",
"electionTime" : Timestamp(1598417963, 1),
"electionDate" : ISODate("2020-08-26T04:59:23Z"),
"configVersion" : 3,
"self" : true,
"lastHeartbeatMessage" : ""
},
{
"_id" : 1,
"name" : "mg-replicaset-1.mg-replicaset-gvr.demo.svc.cluster.local:27017",
"health" : 1,
"state" : 2,
"stateStr" : "SECONDARY",
"uptime" : 606,
"optime" : {
"ts" : Timestamp(1598418585, 1),
"t" : NumberLong(2)
},
"optimeDurable" : {
"ts" : Timestamp(1598418585, 1),
"t" : NumberLong(2)
},
"optimeDate" : ISODate("2020-08-26T05:09:45Z"),
"optimeDurableDate" : ISODate("2020-08-26T05:09:45Z"),
"lastHeartbeat" : ISODate("2020-08-26T05:09:49.489Z"),
"lastHeartbeatRecv" : ISODate("2020-08-26T05:09:50.484Z"),
"pingMs" : NumberLong(0),
"lastHeartbeatMessage" : "",
"syncingTo" : "mg-replicaset-0.mg-replicaset-gvr.demo.svc.cluster.local:27017",
"syncSourceHost" : "mg-replicaset-0.mg-replicaset-gvr.demo.svc.cluster.local:27017",
"syncSourceId" : 0,
"infoMessage" : "",
"configVersion" : 3
},
{
"_id" : 2,
"name" : "mg-replicaset-2.mg-replicaset-gvr.demo.svc.cluster.local:27017",
"health" : 1,
"state" : 2,
"stateStr" : "SECONDARY",
"uptime" : 590,
"optime" : {
"ts" : Timestamp(1598418585, 1),
"t" : NumberLong(2)
},
"optimeDurable" : {
"ts" : Timestamp(1598418585, 1),
"t" : NumberLong(2)
},
"optimeDate" : ISODate("2020-08-26T05:09:45Z"),
"optimeDurableDate" : ISODate("2020-08-26T05:09:45Z"),
"lastHeartbeat" : ISODate("2020-08-26T05:09:49.539Z"),
"lastHeartbeatRecv" : ISODate("2020-08-26T05:09:50.330Z"),
"pingMs" : NumberLong(0),
"lastHeartbeatMessage" : "",
"syncingTo" : "mg-replicaset-1.mg-replicaset-gvr.demo.svc.cluster.local:27017",
"syncSourceHost" : "mg-replicaset-1.mg-replicaset-gvr.demo.svc.cluster.local:27017",
"syncSourceId" : 1,
"infoMessage" : "",
"configVersion" : 3
}
]
We can see from the above output that the replicaset has 3 nodes.
We are now ready to apply the MongoDBOpsRequest
CR to scale this database.
Scale Up Replicas
Here, we are going to scale up the replicas of the replicaset to meet the desired number of replicas after scaling.
Create MongoDBOpsRequest
In order to scale up the replicas of the replicaset of the database, we have to create a MongoDBOpsRequest
CR with our desired replicas. Below is the YAML of the MongoDBOpsRequest
CR that we are going to create,
apiVersion: ops.kubedb.com/v1alpha1
kind: MongoDBOpsRequest
metadata:
name: mops-hscale-up-replicaset
namespace: demo
spec:
type: HorizontalScaling
databaseRef:
name: mg-replicaset
horizontalScaling:
replicas: 4
Here,
spec.databaseRef.name
specifies that we are performing horizontal scaling operation onmops-hscale-up-replicaset
database.spec.type
specifies that we are performingHorizontalScaling
on our database.spec.horizontalScaling.replicas
specifies the desired replicas after scaling.
Let’s create the MongoDBOpsRequest
CR we have shown above,
$ kubectl apply -f https://github.com/kubedb/docs/raw/v2020.11.12/docs/examples/mongodb/scaling/horizontal-scaling/mops-hscale-up-replicaset.yaml
mongodbopsrequest.ops.kubedb.com/mops-hscale-up-replicaset created
Verify Replicaset replicas scaled up successfully
If everything goes well, KubeDB
Enterprise operator will update the replicas of MongoDB
object and related StatefulSets
and Pods
.
Let’s wait for MongoDBOpsRequest
to be Successful
. Run the following command to watch MongoDBOpsRequest
CR,
$ watch kubectl get mongodbopsrequest -n demo
Every 2.0s: kubectl get mongodbopsrequest -n demo
NAME TYPE STATUS AGE
mops-hscale-up-replicaset HorizontalScaling Successful 106s
We can see from the above output that the MongoDBOpsRequest
has succeeded. If we describe the MongoDBOpsRequest
we will get an overview of the steps that were followed to scale the database.
$ kubectl describe mongodbopsrequest -n demo mops-hscale-up-replicaset
Name: mops-hscale-up-replicaset
Namespace: demo
Labels: <none>
Annotations: API Version: ops.kubedb.com/v1alpha1
Kind: MongoDBOpsRequest
Metadata:
Creation Timestamp: 2020-08-26T05:22:33Z
Finalizers:
kubedb.com
Generation: 1
Managed Fields:
API Version: ops.kubedb.com/v1alpha1
Fields Type: FieldsV1
fieldsV1:
f:metadata:
f:annotations:
.:
f:kubectl.kubernetes.io/last-applied-configuration:
f:spec:
.:
f:databaseRef:
.:
f:name:
f:horizontalScaling:
.:
f:replicas:
f:type:
Manager: kubectl
Operation: Update
Time: 2020-08-26T05:22:33Z
API Version: ops.kubedb.com/v1alpha1
Fields Type: FieldsV1
fieldsV1:
f:metadata:
f:finalizers:
f:status:
.:
f:conditions:
f:observedGeneration:
f:phase:
Manager: kubedb-enterprise
Operation: Update
Time: 2020-08-26T05:23:18Z
Resource Version: 5681626
Self Link: /apis/ops.kubedb.com/v1alpha1/namespaces/demo/mongodbopsrequests/mops-hscale-up-replicaset
UID: 8b9b03c4-d95d-41af-b418-312ad81c49de
Spec:
Database Ref:
Name: mg-replicaset
Horizontal Scaling:
Replicas: 4
Type: HorizontalScaling
Status:
Conditions:
Last Transition Time: 2020-08-26T05:22:33Z
Message: MongoDB ops request is being processed
Observed Generation: 1
Reason: Scaling
Status: True
Type: Scaling
Last Transition Time: 2020-08-26T05:22:33Z
Message: Successfully halted mongodb: mg-replicaset
Observed Generation: 1
Reason: HaltDatabase
Status: True
Type: HaltDatabase
Last Transition Time: 2020-08-26T05:23:18Z
Message: Successfully Scaled Up Replicas of StatefulSet
Observed Generation: 1
Reason: ScaleUpReplicaSet
Status: True
Type: ScaleUpReplicaSet
Last Transition Time: 2020-08-26T05:23:18Z
Message: Successfully Resumed mongodb: mg-replicaset
Observed Generation: 1
Reason: ResumeDatabase
Status: True
Type: ResumeDatabase
Last Transition Time: 2020-08-26T05:23:18Z
Message: Successfully completed the modification process
Observed Generation: 1
Reason: Successful
Status: True
Type: Successful
Observed Generation: 1
Phase: Successful
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal HaltDatabase 2m10s KubeDB Enterprise Operator Pausing Mongodb mg-replicaset in Namespace demo
Normal HaltDatabase 2m10s KubeDB Enterprise Operator Successfully Halted Mongodb mg-replicaset in Namespace demo
Normal HaltDatabase 2m10s KubeDB Enterprise Operator Pausing Mongodb mg-replicaset in Namespace demo
Normal HaltDatabase 2m10s KubeDB Enterprise Operator Successfully Halted Mongodb mg-replicaset in Namespace demo
Normal ScaleUpReplicaSet 85s KubeDB Enterprise Operator Successfully Scaled Up Replicas of StatefulSet
Normal ResumeDatabase 85s KubeDB Enterprise Operator Resuming MongoDB
Normal ResumeDatabase 85s KubeDB Enterprise Operator Successfully Resumed mongodb
Normal Successful 85s KubeDB Enterprise Operator Successfully Scaled Database
Now, we are going to verify the number of replicas this database has from the MongoDB object, number of pods the statefulset have,
$ kubectl get mongodb -n demo mg-replicaset -o json | jq '.spec.replicas' 11:26:38
4
$ kubectl get sts -n demo mg-replicaset -o json | jq '.spec.replicas' 11:27:13
4
Now let’s connect to a mongodb instance and run a mongodb internal command to check the number of replicas,
$ kubectl exec -n demo mg-replicaset-0 -- mongo admin -u root -p nrKuxni0wDSMrgwy --eval "db.adminCommand( { replSetGetStatus : 1 } ).members" --quiet 11:28:20
[
{
"_id" : 0,
"name" : "mg-replicaset-0.mg-replicaset-gvr.demo.svc.cluster.local:27017",
"health" : 1,
"state" : 1,
"stateStr" : "PRIMARY",
"uptime" : 1749,
"optime" : {
"ts" : Timestamp(1598419697, 4),
"t" : NumberLong(2)
},
"optimeDate" : ISODate("2020-08-26T05:28:17Z"),
"syncingTo" : "",
"syncSourceHost" : "",
"syncSourceId" : -1,
"infoMessage" : "",
"electionTime" : Timestamp(1598417963, 1),
"electionDate" : ISODate("2020-08-26T04:59:23Z"),
"configVersion" : 4,
"self" : true,
"lastHeartbeatMessage" : ""
},
{
"_id" : 1,
"name" : "mg-replicaset-1.mg-replicaset-gvr.demo.svc.cluster.local:27017",
"health" : 1,
"state" : 2,
"stateStr" : "SECONDARY",
"uptime" : 1724,
"optime" : {
"ts" : Timestamp(1598419697, 4),
"t" : NumberLong(2)
},
"optimeDurable" : {
"ts" : Timestamp(1598419697, 4),
"t" : NumberLong(2)
},
"optimeDate" : ISODate("2020-08-26T05:28:17Z"),
"optimeDurableDate" : ISODate("2020-08-26T05:28:17Z"),
"lastHeartbeat" : ISODate("2020-08-26T05:28:28.990Z"),
"lastHeartbeatRecv" : ISODate("2020-08-26T05:28:27.959Z"),
"pingMs" : NumberLong(0),
"lastHeartbeatMessage" : "",
"syncingTo" : "mg-replicaset-0.mg-replicaset-gvr.demo.svc.cluster.local:27017",
"syncSourceHost" : "mg-replicaset-0.mg-replicaset-gvr.demo.svc.cluster.local:27017",
"syncSourceId" : 0,
"infoMessage" : "",
"configVersion" : 4
},
{
"_id" : 2,
"name" : "mg-replicaset-2.mg-replicaset-gvr.demo.svc.cluster.local:27017",
"health" : 1,
"state" : 2,
"stateStr" : "SECONDARY",
"uptime" : 1708,
"optime" : {
"ts" : Timestamp(1598419697, 4),
"t" : NumberLong(2)
},
"optimeDurable" : {
"ts" : Timestamp(1598419697, 4),
"t" : NumberLong(2)
},
"optimeDate" : ISODate("2020-08-26T05:28:17Z"),
"optimeDurableDate" : ISODate("2020-08-26T05:28:17Z"),
"lastHeartbeat" : ISODate("2020-08-26T05:28:28.990Z"),
"lastHeartbeatRecv" : ISODate("2020-08-26T05:28:27.959Z"),
"pingMs" : NumberLong(0),
"lastHeartbeatMessage" : "",
"syncingTo" : "mg-replicaset-0.mg-replicaset-gvr.demo.svc.cluster.local:27017",
"syncSourceHost" : "mg-replicaset-0.mg-replicaset-gvr.demo.svc.cluster.local:27017",
"syncSourceId" : 0,
"infoMessage" : "",
"configVersion" : 4
},
{
"_id" : 3,
"name" : "mg-replicaset-3.mg-replicaset-gvr.demo.svc.cluster.local:27017",
"health" : 1,
"state" : 2,
"stateStr" : "SECONDARY",
"uptime" : 310,
"optime" : {
"ts" : Timestamp(1598419697, 4),
"t" : NumberLong(2)
},
"optimeDurable" : {
"ts" : Timestamp(1598419697, 4),
"t" : NumberLong(2)
},
"optimeDate" : ISODate("2020-08-26T05:28:17Z"),
"optimeDurableDate" : ISODate("2020-08-26T05:28:17Z"),
"lastHeartbeat" : ISODate("2020-08-26T05:28:29.153Z"),
"lastHeartbeatRecv" : ISODate("2020-08-26T05:28:28.379Z"),
"pingMs" : NumberLong(0),
"lastHeartbeatMessage" : "",
"syncingTo" : "mg-replicaset-2.mg-replicaset-gvr.demo.svc.cluster.local:27017",
"syncSourceHost" : "mg-replicaset-2.mg-replicaset-gvr.demo.svc.cluster.local:27017",
"syncSourceId" : 2,
"infoMessage" : "",
"configVersion" : 4
}
]
From all the above outputs we can see that the replicas of the replicaset is 4
. That means we have successfully scaled up the replicas of the MongoDB replicaset.
Scale Down Replicas
Here, we are going to scale down the replicas of the replicaset to meet the desired number of replicas after scaling.
Create MongoDBOpsRequest
In order to scale down the replicas of the replicaset of the database, we have to create a MongoDBOpsRequest
CR with our desired replicas. Below is the YAML of the MongoDBOpsRequest
CR that we are going to create,
apiVersion: ops.kubedb.com/v1alpha1
kind: MongoDBOpsRequest
metadata:
name: mops-hscale-down-replicaset
namespace: demo
spec:
type: HorizontalScaling
databaseRef:
name: mg-replicaset
horizontalScaling:
replicas: 3
Here,
spec.databaseRef.name
specifies that we are performing horizontal scaling down operation onmops-hscale-down-replicaset
database.spec.type
specifies that we are performingHorizontalScaling
on our database.spec.horizontalScaling.replicas
specifies the desired replicas after scaling.
Let’s create the MongoDBOpsRequest
CR we have shown above,
$ kubectl apply -f https://github.com/kubedb/docs/raw/v2020.11.12/docs/examples/mongodb/scaling/horizontal-scaling/mops-hscale-down-replicaset.yaml
mongodbopsrequest.ops.kubedb.com/mops-hscale-down-replicaset created
Verify Replicaset replicas scaled down successfully
If everything goes well, KubeDB
Enterprise operator will update the replicas of MongoDB
object and related StatefulSets
and Pods
.
Let’s wait for MongoDBOpsRequest
to be Successful
. Run the following command to watch MongoDBOpsRequest
CR,
$ watch kubectl get mongodbopsrequest -n demo
Every 2.0s: kubectl get mongodbopsrequest -n demo
NAME TYPE STATUS AGE
mops-hscale-down-replicaset HorizontalScaling Successful 2m32s
We can see from the above output that the MongoDBOpsRequest
has succeeded. If we describe the MongoDBOpsRequest
we will get an overview of the steps that were followed to scale the database.
$ kubectl describe mongodbopsrequest -n demo mops-hscale-down-replicaset
Name: mops-hscale-down-replicaset
Namespace: demo
Labels: <none>
Annotations: API Version: ops.kubedb.com/v1alpha1
Kind: MongoDBOpsRequest
Metadata:
Creation Timestamp: 2020-08-26T05:36:49Z
Finalizers:
kubedb.com
Generation: 1
Managed Fields:
API Version: ops.kubedb.com/v1alpha1
Fields Type: FieldsV1
fieldsV1:
f:metadata:
f:annotations:
.:
f:kubectl.kubernetes.io/last-applied-configuration:
f:spec:
.:
f:databaseRef:
.:
f:name:
f:horizontalScaling:
.:
f:replicas:
f:type:
Manager: kubectl
Operation: Update
Time: 2020-08-26T05:36:49Z
API Version: ops.kubedb.com/v1alpha1
Fields Type: FieldsV1
fieldsV1:
f:metadata:
f:finalizers:
f:status:
.:
f:conditions:
f:observedGeneration:
f:phase:
Manager: kubedb-enterprise
Operation: Update
Time: 2020-08-26T05:36:54Z
Resource Version: 5691961
Self Link: /apis/ops.kubedb.com/v1alpha1/namespaces/demo/mongodbopsrequests/mops-hscale-down-replicaset
UID: 9563b401-2d20-4624-8374-c008a83a58ad
Spec:
Database Ref:
Name: mg-replicaset
Horizontal Scaling:
Replicas: 3
Type: HorizontalScaling
Status:
Conditions:
Last Transition Time: 2020-08-26T05:36:49Z
Message: MongoDB ops request is being processed
Observed Generation: 1
Reason: Scaling
Status: True
Type: Scaling
Last Transition Time: 2020-08-26T05:36:49Z
Message: Successfully halted mongodb: mg-replicaset
Observed Generation: 1
Reason: HaltDatabase
Status: True
Type: HaltDatabase
Last Transition Time: 2020-08-26T05:36:54Z
Message: Successfully Scale Down Replicas of Replicaset
Observed Generation: 1
Reason: ScaleDownReplicaSet
Status: True
Type: ScaleDownReplicaSet
Last Transition Time: 2020-08-26T05:36:54Z
Message: Successfully Resumed mongodb: mg-replicaset
Observed Generation: 1
Reason: ResumeDatabase
Status: True
Type: ResumeDatabase
Last Transition Time: 2020-08-26T05:36:54Z
Message: Successfully completed the modification process
Observed Generation: 1
Reason: Successful
Status: True
Type: Successful
Observed Generation: 1
Phase: Successful
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal HaltDatabase 3m1s KubeDB Enterprise Operator Pausing Mongodb mg-replicaset in Namespace demo
Normal HaltDatabase 3m1s KubeDB Enterprise Operator Successfully Halted Mongodb mg-replicaset in Namespace demo
Normal ScalingDown 3m1s KubeDB Enterprise Operator Scaling Down Replicas of replicaSet
Normal ScalingDown 3m1s KubeDB Enterprise Operator Scaling Down Replicas of replicaSet
Normal ScaleDownReplicaSet 2m56s KubeDB Enterprise Operator Successfully Scale Down Replicas of Replicaset
Normal ResumeDatabase 2m56s KubeDB Enterprise Operator Resuming MongoDB
Normal ResumeDatabase 2m56s KubeDB Enterprise Operator Successfully Resumed mongodb
Normal Successful 2m56s KubeDB Enterprise Operator Successfully Scaled Database
Now, we are going to verify the number of replicas this database has from the MongoDB object, number of pods the statefulset have,
$ kubectl get mongodb -n demo mg-replicaset -o json | jq '.spec.replicas'
3
$ kubectl get sts -n demo mg-replicaset -o json | jq '.spec.replicas'
3
Now let’s connect to a mongodb instance and run a mongodb internal command to check the number of replicas,
$ kubectl exec -n demo mg-replicaset-0 -- mongo admin -u root -p nrKuxni0wDSMrgwy --eval "db.adminCommand( { replSetGetStatus : 1 } ).members" --quiet
[
{
"_id" : 0,
"name" : "mg-replicaset-0.mg-replicaset-gvr.demo.svc.cluster.local:27017",
"health" : 1,
"state" : 1,
"stateStr" : "PRIMARY",
"uptime" : 2475,
"optime" : {
"ts" : Timestamp(1598420435, 1),
"t" : NumberLong(2)
},
"optimeDate" : ISODate("2020-08-26T05:40:35Z"),
"syncingTo" : "",
"syncSourceHost" : "",
"syncSourceId" : -1,
"infoMessage" : "",
"electionTime" : Timestamp(1598417963, 1),
"electionDate" : ISODate("2020-08-26T04:59:23Z"),
"configVersion" : 5,
"self" : true,
"lastHeartbeatMessage" : ""
},
{
"_id" : 1,
"name" : "mg-replicaset-1.mg-replicaset-gvr.demo.svc.cluster.local:27017",
"health" : 1,
"state" : 2,
"stateStr" : "SECONDARY",
"uptime" : 2450,
"optime" : {
"ts" : Timestamp(1598420425, 1),
"t" : NumberLong(2)
},
"optimeDurable" : {
"ts" : Timestamp(1598420425, 1),
"t" : NumberLong(2)
},
"optimeDate" : ISODate("2020-08-26T05:40:25Z"),
"optimeDurableDate" : ISODate("2020-08-26T05:40:25Z"),
"lastHeartbeat" : ISODate("2020-08-26T05:40:34.917Z"),
"lastHeartbeatRecv" : ISODate("2020-08-26T05:40:33.976Z"),
"pingMs" : NumberLong(0),
"lastHeartbeatMessage" : "",
"syncingTo" : "mg-replicaset-0.mg-replicaset-gvr.demo.svc.cluster.local:27017",
"syncSourceHost" : "mg-replicaset-0.mg-replicaset-gvr.demo.svc.cluster.local:27017",
"syncSourceId" : 0,
"infoMessage" : "",
"configVersion" : 5
},
{
"_id" : 2,
"name" : "mg-replicaset-2.mg-replicaset-gvr.demo.svc.cluster.local:27017",
"health" : 1,
"state" : 2,
"stateStr" : "SECONDARY",
"uptime" : 2434,
"optime" : {
"ts" : Timestamp(1598420425, 1),
"t" : NumberLong(2)
},
"optimeDurable" : {
"ts" : Timestamp(1598420425, 1),
"t" : NumberLong(2)
},
"optimeDate" : ISODate("2020-08-26T05:40:25Z"),
"optimeDurableDate" : ISODate("2020-08-26T05:40:25Z"),
"lastHeartbeat" : ISODate("2020-08-26T05:40:34.917Z"),
"lastHeartbeatRecv" : ISODate("2020-08-26T05:40:33.976Z"),
"pingMs" : NumberLong(0),
"lastHeartbeatMessage" : "",
"syncingTo" : "mg-replicaset-0.mg-replicaset-gvr.demo.svc.cluster.local:27017",
"syncSourceHost" : "mg-replicaset-0.mg-replicaset-gvr.demo.svc.cluster.local:27017",
"syncSourceId" : 0,
"infoMessage" : "",
"configVersion" : 5
}
]
From all the above outputs we can see that the replicas of the replicaset is 3
. That means we have successfully scaled down the replicas of the MongoDB replicaset.
Cleaning Up
To clean up the Kubernetes resources created by this tutorial, run:
kubectl delete mg -n demo mg-replicaset
kubectl delete mongodbopsrequest -n demo mops-vscale-replicaset