Convert a Self-Managed Replica Set to a Sharded Cluster
Overview
This tutorial converts a single three-member replica set to a sharded cluster with two shards. Each shard is an independent three-member replica set. This tutorial is specific to MongoDB 6.0. For other versions of MongoDB, refer to the corresponding version of the MongoDB Manual.
You can convert to a sharded cluster in the UI for deployments hosted in MongoDB Atlas.
Considerations
Individual steps in these procedures note when downtime will occur.
Important
These procedures cause some downtime for your deployment.
Prerequisites
This tutorial uses a total of ten servers: one server for the
mongos
and three servers each for the first replica set, the second replica set, and the config server replica set.
Each server must have a resolvable domain, hostname, or IP address within your system.
The tutorial uses the default data directories (e.g. /data/db
and
/data/configdb
). Create the appropriate directories with
appropriate permissions. To use different paths, see
Self-Managed Configuration File Options .
Procedures
Deploy the Config Server Replica Set
Deploy a three-member replica set for the config servers. In this example, the config servers use the following hosts:
mongodb7.example.net
mongodb8.example.net
mongodb9.example.net
Configure the Config Servers
Configure a
mongod
instance on each config server host. Specify these options in the configuration file for eachmongod
instance:OptionValueconfigReplSet
configsvr
localhost
, followed by any other hostnames on which themongod
should listen for client connections.replication: replSetName: configReplSet sharding: clusterRole: configsvr net: bindIp: localhost,<hostname(s)> Include additional options as appropriate for your deployment.
Start the Config Servers
Deploy the
mongod
with your specified configuration:mongod --config <PATH_TO_CONFIG_FILE> The config servers use the default data directory
/data/configdb
and the default port27019
.Connect to one of the config servers.
Use
mongosh
to connect to one of the config servers. For example:mongosh "mongodb://mongodb7.example.net:27019" Initiate the config server replica set.
To initiate the replica set, run
rs.initiate()
:rs.initiate( { _id: "configReplSet", configsvr: true, members: [ { _id: 0, host: "mongodb7.example.net:27019" }, { _id: 1, host: "mongodb8.example.net:27019" }, { _id: 2, host: "mongodb9.example.net:27019" } ] } ) The preceding command uses the localhost exception to perform administrative actions without authentication.
Important
Run
rs.initiate()
on just one and only onemongod
instance for the replica set.
Restore Existing Users and Roles to the New Config
Restore the existing users and roles you got when you ran mongodump
.
mongorestore ./adminDump --nsInclude "admin.*" --host <configPrimaryURI>
The preceding command uses the localhost exception to perform administrative actions without authentication.
The output from running this command may be similar to:
0 document(s) restored successfully
This message does not indicate a problem. This output means that 0 documents other than the users and roles were restored.
Secure Config Server Replica Set
Reconfigure and restart the config server replica set.
Reconfigure the Config Servers
Select the tab for your authentication mechanism:
Restart a
mongod
instance on each of these hosts:mongodb7.example.net
mongodb8.example.net
mongodb9.example.net
Specify these options in the configuration file for each
mongod
instance:OptionValueThe path to the key file used for your initial replica set.security: keyFile: <PATH_TO_KEYFILE> replication: replSetName: configReplSet sharding: clusterRole: configsvr net: bindIp: localhost,<hostname(s)> Include additional options as appropriate for your deployment.
Restart a
mongod
instance on each of these hosts:mongodb7.example.net
mongodb8.example.net
mongodb9.example.net
Specify these options in the configuration file for each
mongod
instance in addition to the options you already configured:OptionValuex509
requireTLS
The absolute path to the.pem
file that contains both the TLS certificate and key.The absolute path to the.pem
file that contains the root certificate chain from the Certificate Authority.localhost
, followed by any other hostnames on which themongod
should listen for client connections.WARNING: Before binding to a non-localhost (e.g. publicly accessible) IP address, ensure you have secured your cluster from unauthorized access. For a complete list of security recommendations, see Security Checklist for Self-Managed Deployments. At minimum, consider enabling authentication and hardening network infrastructure.
sharding: clusterRole: configsvr replication: replSetName: configReplSet security: clusterAuthMode: x509 net: tls: mode: requireTLS certificateKeyFile: <FILE_WITH_COMBINED_CERT_AND_KEY> CAFile: <CA_FILE> bindIp: localhost,<hostname(s)> Include additional options as appropriate for your deployment, such as
net.tls.certificateKeyFilePassword
if your TLS certificate-key file is encrypted with a password.Restart MongoDB
Restart the
mongod
with your specified configuration:mongod --config <PATH_TO_CONFIG_FILE> --shutdown mongod --config <PATH_TO_CONFIG_FILE>
Deploy the mongos
The mongos
provides the interface between the client
applications and the sharded cluster.
Create a configuration file for the mongos.
Specify these options in your
mongos
configuration file:OptionValueconfigReplSet
, followed by a slash/
and at least one of the config server hostnames and ports.The path to the key file used for your initial replica set.localhost
, followed by any other hostnames on which themongos
should listen for client connections.sharding: configDB: configReplSet/mongodb7.example.net:27019,mongodb8.example.net:27019,mongodb9.example.net:27019 security: keyFile: <PATH_TO_KEYFILE> net: bindIp: localhost,<hostname(s)> Include additional options as appropriate for your deployment.
Specify these options in your
mongos
configuration file:OptionValueconfigReplSet
, followed by a slash/
and at least one of the config server hostnames and ports.x509
requireTLS
The absolute path to the.pem
file that contains both the TLS certificate and key.The absolute path to the.pem
file that contains the root certificate chain from the Certificate Authority.localhost
, followed by any other hostnames on which themongos
should listen for client connections.sharding: configDB: configReplSet/mongodb7.example.net:27019,mongodb8.example.net:27019,mongodb9.example.net:27019 security: clusterAuthMode: x509 net: tls: mode: requireTLS certificateKeyFile: <FILE_WITH_COMBINED_CERT_AND_KEY> CAFile: <CA_FILE> bindIp: localhost,<hostname(s)> Include any additional options appropriate for your deployment.
Deploy the mongos.
Deploy the
mongos
with your specified configuration:mongos --config <PATH_TO_CONFIG_FILE>
Restart the Initial Replica Set as a Shard
In this example, your initial replica set is a three-member replica set. This step updates the initial replica set so that it can be added as a shard to your sharded cluster.
The replica set runs on these hosts:
mongodb0.example.net:27017
mongodb1.example.net:27017
mongodb2.example.net:27017
For sharded clusters, you must set the role for each
mongod
instance in the shard to shardsvr
. To specify
the server role, set the sharding.clusterRole
setting in the
mongod
configuration file.
Note
Connect to a member of your initial replica set.
Use
mongosh
to connect to one of the members of your initial replica set.mongosh "mongodb://<username>@mongodb0.example.net:27017" If your deployment uses x.509 authentication, specify these
mongosh
options:For example:
mongosh "mongodb://<username>@mongodb0.example.net:27017" --tls --tlsCAFile <CA_FILE> --tlsCertificateKeyFile <filename> Determine the replica set's primary and secondaries.
Run
rs.status()
to determine the primary and secondaries:rs.status() In the command output, the
replSetGetStatus.members[n].stateStr
field indicates which member is the primary and which members are secondaries.Restart the secondaries with the
--shardsvr
option.Warning
This step requires some downtime for applications connected to the replica set secondaries.
After you restart a secondary, any applications that are connected to that secondary return a
CannotVerifyAndSignLogicalTime
error until you perform the steps in Add the Initial Replica Set as a Shard.You can also restart your application to stop it from receiving
CannotVerifyAndSignLogicalTime
errors.Connect to a secondary.
Use
mongosh
to connect to one of the secondaries.mongosh "mongodb://<username>@<host>:<port>" Shut down the secondary.
Run the following commands:
use admin db.shutdownServer() Edit the secondary's configuration file.
In the secondary's configuration file, set
sharding.clusterRole
toshardsvr
:security: keyFile: <PATH_TO_KEYFILE> replication: replSetName: rs0 sharding: clusterRole: shardsvr net: port: 27017 bindIp: localhost,<hostname(s)> Include additional options as appropriate for your deployment.
Restart the secondary as a shard server.
Run the following command on the host containing the secondary:
mongod --config <PATH_TO_CONFIG_FILE> Repeat the shut down and restart steps for the other secondary.
Connect to a secondary.
Use
mongosh
to connect to one of the secondaries.If your deployment uses x.509 authentication, specify these
mongosh
options:mongosh "mongodb://<username>@<host>:<port>" --tls --tlsCAFile <CA_FILE> --tlsCertificateKeyFile <filename> Shut down the secondary.
Run the following commands:
use admin db.shutdownServer() Edit the secondary's configuration file.
In the secondary's configuration file, set
sharding.clusterRole
toshardsvr
:replication: replSetName: rs0 sharding: clusterRole: shardsvr security: clusterAuthMode: x509 net: port: 27017 tls: mode: requireTLS certificateKeyFile: <FILE_WITH_COMBINED_CERT_AND_KEY> CAFile: <CA_FILE> bindIp: localhost,<hostname(s)> Include additional options as appropriate for your deployment, such as
net.tls.certificateKeyFilePassword
if your TLS certificate-key file is encrypted with a password.Restart the secondary as a shard server.
Run the following command on the host containing the secondary:
mongod --config <PATH_TO_CONFIG_FILE> Repeat the shut down and restart steps for the other secondary.
Restart the primary with the --shardsvr
option.
Warning
This step requires some downtime for applications connected to the primary of the replica set.
After you restart the primary, any applications that are
connected to the primary return a
CannotVerifyAndSignLogicalTime
error until you perform the
steps in Add the Initial Replica Set as a Shard.
You can also restart your application to stop it from receiving
CannotVerifyAndSignLogicalTime
errors.
Connect to the primary.
Use
mongosh
to connect to the primary:mongosh "mongodb://<username>@<host>:<port>" Step down the primary.
Run the following command:
rs.stepDown() Verify that the step down is complete.
Run
rs.status()
to confirm that the member you are connected to has stepped down and is now a secondary:rs.status() Shut down the former primary.
Run the following commands:
use admin db.shutdownServer() Wait for the shutdown to complete.
Edit the primary's configuration file.
In the primary's configuration file, set
sharding.clusterRole
toshardsvr
:security: keyFile: <PATH_TO_KEYFILE> replication: replSetName: rs0 sharding: clusterRole: shardsvr net: port: 27017 bindIp: localhost,<hostname(s)> Include additional options as appropriate for your deployment.
Restart the primary as a shard server.
Run the following command on the host containing the primary:
mongod --config <PATH_TO_CONFIG_FILE>
Connect to the primary.
Use
mongosh
to connect to one of the secondaries.If your deployment uses x.509 authentication, specify these
mongosh
options:If your deployment uses x.509 authentication, specify these
mongosh
options:mongosh "mongodb://<username>@<host>:<port>" --tls --tlsCAFile <CA_FILE> --tlsCertificateKeyFile <filename> Step down the primary.
Run the following command:
rs.stepDown() Verify that the step down is complete.
Run
rs.status()
to confirm that the member you are connected to has stepped down and is now a secondary:rs.status() Shut down the former primary.
Run the following commands:
use admin db.shutdownServer() Wait for the shutdown to complete.
Edit the primary's configuration file.
In the primary's configuration file, set
sharding.clusterRole
toshardsvr
:replication: replSetName: rs0 sharding: clusterRole: shardsvr security: clusterAuthMode: x509 net: port: 27017 tls: mode: requireTLS certificateKeyFile: <FILE_WITH_COMBINED_CERT_AND_KEY> CAFile: <CA_FILE> bindIp: localhost,<hostname(s)> Include additional options as appropriate for your deployment, such as
net.tls.certificateKeyFilePassword
if your TLS certificate-key file is encrypted with a password.Restart the primary as a shard server.
Run the following command on the host containing the primary:
mongod --config <PATH_TO_CONFIG_FILE>
Add the Initial Replica Set as a Shard
After you convert the initial replica set (rs0
) to a shard, add it
to the sharded cluster.
Connect to the
mongos
as your cluster's administrative user.The
mongos
instance is running on hostmongodb6.example.net
.To connect
mongosh
to themongos
, run the following command:mongosh "mongodb://admin01@mongodb6.example.net:27017" This command authenticates you as the
admin01
user you created on the sharded cluster. After you enter the command, enter your user's password.Add the shard.
To add a shard to the cluster, run the
sh.addShard()
method:sh.addShard( "rs0/mongodb0.example.net:27017,mongodb1.example.net:27017,mongodb2.example.net:27017" ) Warning
Once the new shard is active,
mongosh
and other clients must always connect to themongos
instance. Do not connect directly to themongod
instances. If your clients connect to shards directly, you may create data or metadata inconsistencies.
Update Your Application Connection String
After you add the first shard to your cluster, update the connection string used by your applications to the connection string for your sharded cluster. Then, restart your applications.
Deploy a Second Replica Set
Deploy a new replica set called rs1
. The members of replica set
rs1
are on the following hosts:
mongodb3.example.net
mongodb4.example.net
mongodb5.example.net
Start each member of the replica set.
For each
mongod
instance in the replica set, create a configuration file with these options:OptionValueThe path to the key file used for your initial replica set.rs1
shardsvr
localhost
, followed by any other hostnames on which themongod
should listen for client connections.security: keyFile: <PATH_TO_KEYFILE> replication: replSetName: rs1 sharding: clusterRole: shardsvr net: bindIp: localhost,<hostname(s)> Include additional options as appropriate for your deployment.
For each member, start a
mongod
with the following options:OptionValuers1
shardsvr
x509
requireTLS
The absolute path to the.pem
file that contains both the TLS certificate and key.The absolute path to the.pem
file that contains the root certificate chain from the Certificate Authority.localhost
, followed by any other hostnames on which themongod
should listen for client connections.replication: replSetName: rs1 sharding: clusterRole: shardsvr security: clusterAuthMode: x509 net: tls: mode: requireTLS certificateKeyFile: <FILE_WITH_COMBINED_CERT_AND_KEY> CAFile: <CA_FILE> bindIp: localhost,<hostname(s)> Deploy the
mongod
with your specified configuration:mongod --config <PATH_TO_CONFIG_FILE> Note
When you specify the
--shardsvr
option for amongod
instance, the instance runs on port27018
by default.Start each member of the replica set.
Connect to a replica set member.
Use
mongosh
to connect to one of the replica set members. For example:mongosh "mongodb://mongodb3.example.net:27018" mongosh "mongodb://mongodb3.example.net:27018" --tls --tlsCAFile <CA_FILE> --tlsCertificateKeyFile <filename> Initiate the replica set.
In
mongosh
, run thers.initiate()
method to initiate a replica set that contains the current member:rs.initiate( { _id : "rs1", members: [ { _id: 0, host: "mongodb3.example.net:27018" }, { _id: 1, host: "mongodb4.example.net:27018" }, { _id: 2, host: "mongodb5.example.net:27018" } ] } ) The preceding command requires the localhost exception to perform administrative actions without authentication.
Important
Run
rs.initiate()
on just one and only onemongod
instance for the replica set.Add an administrative user for the replica set.
After you deploy the replica set, use the localhost exception to create the replica set's first user.
Determine the replica set primary.
To determine the primary, run
rs.status()
:rs.status() In the command output, the
replSetGetStatus.members[n].stateStr
field indicates which member is the primary.Connect to the replica set primary.
Connect to the replica set primary with
mongosh
. For example, if the primary ismongodb4.example.net
, run this command:mongosh "mongodb://mongodb4.example.net:27018" Create an administrative user.
Run the following
db.createUser()
method to create a user namedrs1Admin
with theuserAdmin
role:use admin db.createUser( { user: "rs1Admin", pwd: passwordPrompt(), roles: [ { role: "userAdmin", db: "admin" } ] } ) After you run the command, the database prompts you to enter a password for the
rs1Admin
user.
Add the Second Replica Set to the Cluster as a Shard
Add the new replica set, rs1
, to the sharded cluster.
Connect
mongosh
to themongos
.Run the following command from the command line to connect to the
mongos
instance running on hostmongodb6.example.net
:mongosh "mongodb://admin01@mongodb6.example.net:27017/admin" mongosh "mongodb://admin01@mongodb6.example.net:27017/admin" --tls --tlsCAFile <CA_FILE> --tlsCertificateKeyFile <filename> This command authenticates you as the
admin01
user you created on the sharded cluster. After you enter the command, enter your user's password.Add the second shard.
After you connect to the
mongos
, add the replica setrs1
as a shard to the cluster with thesh.addShard()
method:sh.addShard( "rs1/mongodb3.example.net:27018,mongodb4.example.net:27018,mongodb5.example.net:27018" )
Shard a Collection
The final step of the procedure is to shard a collection in the sharded cluster.
Determine the shard key.
Determine the shard key for the collection. The shard key indicates how MongoDB distributes the documents between shards. Good shard keys:
Have values that are evenly distributed among all documents.
Group documents that are often accessed at the same time into contiguous chunks.
Allow for effective distribution of activity among shards.
For more information, see Choose a Shard Key.
This procedure uses the
number
field as the shard key for thetest_collection
collection.Create an index on the shard key.
Before you shard a non-empty collection, create an index on the shard key:
use test db.test_collection.createIndex( { "number" : 1 } ) Shard the collection.
In the
test
database, shard thetest_collection
. Specifynumber
as the shard key.sh.shardCollection( "test.test_collection", { "number" : 1 } ) The next time that the balancer runs, it redistributes chunks of documents between shards. As clients insert additional documents into this collection, the
mongos
routes the documents to the appropriate shard.When the balancer redistributes chunks, it may negatively impact your application's performance. To minimize performance impact, you can specify when the balancer runs so it does not run during peak hours. To learn more, see Schedule the Balancing Window.
Learn More
For more sharding tutorials and procedures, see these pages: