Skip to content

Commit fe3cf4a

Browse files
(DOCSP-44990) Updating High Availability Page (#95)
* Applied all changes from doc * Apply suggestions from code review Co-authored-by: Sarah Simpers <[email protected]> * Apply suggestions from code review Co-authored-by: Sarah Simpers <[email protected]> * Fixed title underline --------- Co-authored-by: Sarah Simpers <[email protected]>
1 parent 45080fe commit fe3cf4a

File tree

1 file changed

+28
-58
lines changed

1 file changed

+28
-58
lines changed

source/high-availability.txt

Lines changed: 28 additions & 58 deletions
Original file line numberDiff line numberDiff line change
@@ -14,7 +14,7 @@ High Availability
1414

1515
Consult this page to plan the appropriate cluster configuration that optimizes
1616
your availability and performance while aligning with your enterprise's
17-
billing and access needs.
17+
cost controls and access needs.
1818

1919
{+service+} Features and Recommendations for High Availability
2020
---------------------------------------------------------------
@@ -23,8 +23,8 @@ Features
2323
~~~~~~~~
2424

2525
When you launch a new cluster in |service|, {+service+} automatically configures a
26-
minimum three-node replica set and distributes it across availability zones. If a
27-
primary member experiences an outage, |service| automatically detects this failure
26+
minimum three-node replica set and distributes it in the region you deploy to. If a
27+
primary member experiences an outage, the MongoDB database automatically detects this failure
2828
and elects a secondary member as a replacement, and promotes this secondary member
2929
to become the new primary. |service| then restores or replaces the failing member
3030
to ensure that the cluster is returned to its target configuration as soon as possible.
@@ -56,67 +56,37 @@ ensure that {+service+} stores data in a particular geographical area.
5656
Recommended Deployment Topologies
5757
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
5858

59-
Single Region, 3 Node Replica Set / Shard (PSS)
60-
````````````````````````````````````````````````
59+
Single Region, 3 Node Replica Set / Shard (Primary Secondary Secondary)
60+
```````````````````````````````````````````````````````````````````````
6161
This topology is appropriate if low latency is required but high availability requirements
62-
are average. This topology can tolerate any 1 node failure, easily satify majority write
63-
concern with in-region secondaries, maintain a primary in the preferred region upon any node
64-
failure, is inexpensive, and the least complicated. The reads from secondaries in this topology
65-
encounter low latency in your preferred region.
62+
are limited to a single region. This topology can tolerate any 1 node failure and easily satisfy majority write
63+
concern within region secondaries. This will maintain a primary in the preferred region upon any node
64+
failure, limits cost, and is the least complicated from an application architecture perspective.
65+
The reads and writes from secondaries in this topology encounter low latency in your preferred region.
6666

6767
This topology, however, can't tolerate a regional outage.
6868

69-
2-Region, 3 Node Replica Set / Shard (PS - S)
70-
``````````````````````````````````````````````
71-
This topology is a highly available cluster configuration, with strong performance and availability.
72-
This topology can easily satisfy majority write concern with in-region secondary, tolerate any
73-
1 node failure, maintain a primary in the preferred region upon any node failure, is inexpensive,
74-
and can tolerate regional outage from the least preferred region. The reads from one secondary in
75-
this topology also encounter low latency in the preferred region.
76-
77-
This topology, however, can't tolerate a regional outage from a preferred region without manual
78-
intervention, such as adding additional nodes to the least-preferred region. Furthermore, the reads
79-
from secondaries in a specific region is limited to a single node, and there is potential data loss
80-
during regional outage.
81-
82-
3-Region, 3 Node Replica Set / Shard (P - S - S)
83-
``````````````````````````````````````````````````
84-
This topology is a typical multi-regional topology where high availability and data durability are
85-
more important than write latency performance. This topology can tolerate any 1 node failure, any
86-
1 region outage, and is inexpensive.
87-
88-
This topology, however, experiences higher latency because its majority write concern must replicate
89-
to a second region. Furthermore, a primary node failure shifts the primary out of the preferred region,
90-
and reads from secondaries always exist in a different region compared to the primary, and may exist
91-
outside of the preferred region.
92-
93-
3-Region, 5 Node Replica Set / Shard (PS - SS - S)
94-
````````````````````````````````````````````````````
95-
This topology is a typical multi-regional topology for mission-critical applications that require
96-
in-region availability. This topology can tolerate any 2 nodes' failure, tolerate primary node failure
69+
3-Region, 3 Node Replica Set / Shard (Primary - Secondary - Secondary)
70+
``````````````````````````````````````````````````````````````````````
71+
This topology is the standard multi-regional topology where high availability can be provided to tolerate a regional outage.
72+
This topology can tolerate any 1 node failure, any 1 region outage, and is the least expensive multi-region topology.
73+
74+
If the application requires high durability and the app server code has the write majority option set
75+
in the driver, this topology will experience higher latency for writes as they need to replicate to a second region.
76+
Additionally, any reads off the secondary always exist in a different region than the preferred primary region
77+
and will require a different app server architecture. Furthermore, any failover event will occur to a different region, and your
78+
application architecture must adjust as a result.
79+
80+
3-Region, 5 Node Replica Set / Shard (Primary Secondary - Secondary Secondary - Secondary)
81+
``````````````````````````````````````````````````````````````````````````````````````````
82+
This topology is the preferred topology that balances high availability, performance, and cost across multiple regions.
83+
This topology can tolerate any 2 nodes' failure, tolerate primary node failure
9784
while keeping the new primary in the preferred region, and tolerate any 1 region outage.
9885

99-
This topology, however, experiences higher latency because its majority write concern must replicate
100-
to a second region. Furthermore, reads from secondaries always exist in a different region compared
101-
to the primary, and may exist outside of the preferred region.
102-
103-
3-Region, 7 Node Replica Set / Shard (PSS - SSS - S)
104-
`````````````````````````````````````````````````````
105-
This topology is a non-typical multi-regional topology for mission-critical applications that require
106-
in-region availability. This topology can tolerate any 3 nodes' failure, tolerate primary node failure
107-
and a secondary node failure in the preferred region, and tolerate any 1 region outage.
108-
109-
This topology, however, experiences higher latency because its majority write concern must replicate
110-
to a second region.
111-
112-
5-Region, 9 Node Replica Set / Shard (PS - SS - SS - SS - S)
113-
`````````````````````````````````````````````````````````````
114-
This topology is a non-typical multi-regional topology for the most availability-demanding applications
115-
that require in-region availability. This topology can tolerate any 4 nodes' failure, tolerate primary
116-
node failure while keeping the new primary in the preferred region, and tolerate any 2 region outages.
117-
118-
This topology, however, experiences higher latency because its majority write concern must replicate
119-
to two additional regions (3 regions in total).
86+
If the application requires high durability and the app server code has the write majority option set
87+
in the driver, this topology will experience higher latency for writes as they need to replicate to a second region.
88+
In order to accommodate a regional outage, the application tier must be configured to also fail over and shift work to
89+
the elected primary.
12090

12191
Examples
12292
--------

0 commit comments

Comments
 (0)