-
Notifications
You must be signed in to change notification settings - Fork 5.6k
/
replsettest.js
2622 lines (2252 loc) · 106 KB
/
replsettest.js
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
/**
* Sets up a replica set. To make the set running, call {@link #startSet},
* followed by {@link #initiate} (and optionally,
* {@link #awaitSecondaryNodes} to block till the set is fully operational).
* Note that some of the replica start up parameters are not passed here,
* but to the #startSet method.
*
* @param {Object|string} opts If this value is a string, it specifies the connection string for
* a MongoD host to be used for recreating a ReplSetTest from. Otherwise, if it is an object,
* it must have the following contents:
*
* {
* name {string}: name of this replica set. Default: 'testReplSet'
* host {string}: name of the host machine. Hostname will be used
* if not specified.
* useHostName {boolean}: if true, use hostname of machine,
* otherwise use localhost
* nodes {number|Object|Array.<Object>}: number of replicas. Default: 0.
* Can also be an Object (or Array).
* Format for Object:
* {
* <any string>: replica member option Object. @see MongoRunner.runMongod
* <any string2>: and so on...
* }
* If object has a special "rsConfig" field then those options will be used for each
* replica set member config options when used to initialize the replica set, or
* building the config with getReplSetConfig()
*
* Format for Array:
* An array of replica member option Object. @see MongoRunner.runMongod
*
* Note: For both formats, a special boolean property 'arbiter' can be
* specified to denote a member is an arbiter.
*
* Note: A special "bridgeOptions" property can be specified in both the object and array
* formats to configure the options for the mongobridge corresponding to that node. These
* options are merged with the opts.bridgeOptions options, where the node-specific options
* take precedence.
*
* nodeOptions {Object}: Command-line options to apply to all nodes in the replica set.
* Format for Object:
* { cmdline-param-with-no-arg : "",
* param-with-arg : arg }
* This turns into "mongod --cmdline-param-with-no-arg --param-with-arg arg"
*
* causallyConsistent {boolean}: Specifies whether the connections to the replica set nodes
* should be created with the 'causal consistency' flag enabled, which means they will gossip
* the cluster time and add readConcern afterClusterTime where applicable. Defaults to false.
*
* oplogSize {number}: Default: 40
* useSeedList {boolean}: Use the connection string format of this set
* as the replica set name (overrides the name property). Default: false
* keyFile {string}
* protocolVersion {number}: protocol version of replset used by the replset initiation.
*
* useBridge {boolean}: If true, then a mongobridge process is started for each node in the
* replica set. Both the replica set configuration and the connections returned by startSet()
* will be references to the proxied connections. Defaults to false.
*
* bridgeOptions {Object}: Options to apply to all mongobridge processes. Defaults to {}.
*
* settings {object}: Setting used in the replica set config document.
* Example:
* settings: { chainingAllowed: false, ... }
* }
*
* Member variables:
* nodes {Array.<Mongo>} - connection to replica set members
*/
var ReplSetTest = function(opts) {
'use strict';
if (!(this instanceof ReplSetTest)) {
return new ReplSetTest(opts);
}
// Capture the 'this' reference
var self = this;
// Replica set health states
var Health = {UP: 1, DOWN: 0};
var _alldbpaths;
var _configSettings;
// mongobridge related variables. Only available if the bridge option is selected.
var _useBridge;
var _bridgeOptions;
var _unbridgedPorts;
var _unbridgedNodes;
var _causalConsistency;
// Some code still references kDefaultTimeoutMS as a (non-static) member variable, so make sure
// it's still accessible that way.
this.kDefaultTimeoutMS = ReplSetTest.kDefaultTimeoutMS;
var oplogName = 'oplog.rs';
// Publicly exposed variables
/**
* Returns the config document reported from the specified connection.
*/
function _replSetGetConfig(conn) {
return assert.commandWorked(conn.adminCommand({replSetGetConfig: 1})).config;
}
/**
* Invokes the 'ismaster' command on each individual node and returns the current primary, or
* false if none is found. Populates the following cached values:
* '_master': the current primary
* '_slaves': all nodes other than 'master' (note this includes arbiters)
* '_liveNodes': all currently reachable nodes
*/
function _callIsMaster() {
self._liveNodes = [];
self._master = null;
self._slaves = [];
var twoPrimaries = false;
self.nodes.forEach(function(node) {
try {
node.setSlaveOk();
var n = node.getDB('admin').runCommand({ismaster: 1});
self._liveNodes.push(node);
if (n.ismaster == true) {
if (self._master) {
twoPrimaries = true;
} else {
self._master = node;
}
} else {
self._slaves.push(node);
}
} catch (err) {
print("ReplSetTest Could not call ismaster on node " + node + ": " + tojson(err));
self._slaves.push(node);
}
});
if (twoPrimaries) {
return false;
}
return self._master || false;
}
/**
* Attempt to connect to all nodes and returns a list of slaves in which the connection was
* successful.
*/
function _determineLiveSlaves() {
_callIsMaster();
return self._slaves.filter(function(n) {
return self._liveNodes.indexOf(n) !== -1;
});
}
function asCluster(conn, fn, keyFileParam = self.keyFile) {
if (keyFileParam) {
return authutil.asCluster(conn, keyFileParam, fn);
} else {
return fn();
}
}
/**
* Returns 'true' if the "conn" has been configured to run without journaling enabled.
*/
function _isRunningWithoutJournaling(conn) {
var result = asCluster(conn, function() {
var serverStatus = assert.commandWorked(conn.adminCommand({serverStatus: 1}));
if (serverStatus.storageEngine.hasOwnProperty('persistent')) {
if (!serverStatus.storageEngine.persistent) {
return true;
}
} else if (serverStatus.storageEngine.name == 'inMemory' ||
serverStatus.storageEngine.name == 'ephemeralForTest') {
return true;
}
var cmdLineOpts = assert.commandWorked(conn.adminCommand({getCmdLineOpts: 1}));
var getWithDefault = function(dict, key, dflt) {
if (dict[key] === undefined)
return dflt;
return dict[key];
};
return !getWithDefault(
getWithDefault(getWithDefault(cmdLineOpts.parsed, "storage", {}), "journal", {}),
"enabled",
true);
});
return result;
}
/**
* Wait for a rs indicator to go to a particular state or states.
*
* @param node is a single node or list of nodes, by id or conn
* @param states is a single state or list of states
* @param ind is the indicator specified
* @param timeout how long to wait for the state to be reached
* @param reconnectNode indicates that we should reconnect to a node that stepped down
*/
function _waitForIndicator(node, states, ind, timeout, reconnectNode) {
timeout = timeout || self.kDefaultTimeoutMS;
if (reconnectNode === undefined) {
reconnectNode = true;
}
if (node.length) {
var nodes = node;
for (var i = 0; i < nodes.length; i++) {
if (states.length)
_waitForIndicator(nodes[i], states[i], ind, timeout, reconnectNode);
else
_waitForIndicator(nodes[i], states, ind, timeout, reconnectNode);
}
return;
}
if (!node.getDB) {
node = self.nodes[node];
}
if (!states.length) {
states = [states];
}
print("ReplSetTest waitForIndicator " + ind + " on " + node);
printjson(states);
print("ReplSetTest waitForIndicator from node " + node);
var lastTime = null;
var currTime = new Date().getTime();
var status;
let foundState;
assert.soon(function() {
try {
var conn = _callIsMaster();
if (!conn) {
conn = self._liveNodes[0];
}
// Try again to load connection
if (!conn)
return false;
asCluster(conn, function() {
status = conn.getDB('admin').runCommand({replSetGetStatus: 1});
});
} catch (ex) {
print("ReplSetTest waitForIndicator could not get status: " + tojson(ex));
return false;
}
var printStatus = false;
if (lastTime == null || (currTime = new Date().getTime()) - (1000 * 5) > lastTime) {
if (lastTime == null) {
print("ReplSetTest waitForIndicator Initial status (timeout : " + timeout +
") :");
}
printjson(status);
lastTime = new Date().getTime();
printStatus = true;
}
if (typeof status.members == 'undefined') {
return false;
}
for (var i = 0; i < status.members.length; i++) {
if (printStatus) {
print("Status for : " + status.members[i].name + ", checking " + node.host +
"/" + node.name);
}
if (status.members[i].name == node.host || status.members[i].name == node.name) {
for (var j = 0; j < states.length; j++) {
if (printStatus) {
print("Status -- " + " current state: " + status.members[i][ind] +
", target state : " + states[j]);
}
if (typeof(states[j]) != "number") {
throw new Error("State was not an number -- type:" + typeof(states[j]) +
", value:" + states[j]);
}
if (status.members[i][ind] == states[j]) {
foundState = states[j];
return true;
}
}
}
}
return false;
}, "waiting for state indicator " + ind + " for " + timeout + "ms", timeout);
// If we were waiting for the node to step down, wait until we can connect to it again,
// since primaries close external connections upon stepdown. This ensures that the
// connection to the node is usable after the function returns.
if (reconnectNode && foundState === ReplSetTest.State.SECONDARY) {
assert.soon(function() {
try {
node.getDB("foo").bar.stats();
return true;
} catch (e) {
return false;
}
}, "timed out waiting to reconnect to node " + node.name);
}
print("ReplSetTest waitForIndicator final status:");
printjson(status);
}
/**
* Wait for a health indicator to go to a particular state or states.
*
* @param node is a single node or list of nodes, by id or conn
* @param state is a single state or list of states. ReplSetTest.Health.DOWN can
* only be used in cases when there is a primary available or slave[0] can
* respond to the isMaster command.
*/
function _waitForHealth(node, state, timeout) {
_waitForIndicator(node, state, "health", timeout);
}
/**
* Returns true if the OpTime is empty, else false.
*
* Empty OpTime Formats:
* PV0: Timestamp(0,0)
* PV1: {ts: Timestamp(0,0), t: NumberLong(-1)}
*/
function _isEmptyOpTime(opTime) {
if (!opTime.hasOwnProperty("ts") || !opTime.hasOwnProperty("t")) {
return (opTime.getTime() == 0 && opTime.getInc() == 0);
}
return (opTime.ts.getTime() == 0 && opTime.ts.getInc() == 0 && opTime.t == -1);
}
/**
* Returns the OpTime for the specified host by issuing replSetGetStatus.
*/
function _getLastOpTime(conn) {
var replSetStatus =
assert.commandWorked(conn.getDB("admin").runCommand({replSetGetStatus: 1}));
var connStatus = replSetStatus.members.filter(m => m.self)[0];
var opTime = connStatus.optime;
if (_isEmptyOpTime(opTime)) {
throw new Error("last OpTime is empty -- connection: " + conn);
}
return opTime;
}
/**
* Returns the {readConcern: majority} OpTime for the host.
* This is the OpTime of the host's "majority committed" snapshot.
* This function may return an OpTime with Timestamp(0,0) and Term(0) if read concern majority
* is not enabled, or if there has not been a committed snapshot yet.
*/
function _getReadConcernMajorityOpTime(conn) {
var replSetStatus =
assert.commandWorked(conn.getDB("admin").runCommand({replSetGetStatus: 1}));
return (replSetStatus.OpTimes || replSetStatus.optimes).readConcernMajorityOpTime ||
{ts: Timestamp(0, 0), t: NumberLong(0)};
}
/**
* Returns the {readConcern: majority} OpTime for the host. Throws if not available.
*/
this.getReadConcernMajorityOpTimeOrThrow = function(conn) {
const majorityOpTime = _getReadConcernMajorityOpTime(conn);
if (friendlyEqual(majorityOpTime, {ts: Timestamp(0, 0), t: NumberLong(0)})) {
throw new Error("readConcern majority optime not available");
}
return majorityOpTime;
};
/**
* Returns the last durable OpTime for the host if running with journaling.
* Returns the last applied OpTime otherwise.
*/
function _getDurableOpTime(conn) {
var replSetStatus =
assert.commandWorked(conn.getDB("admin").runCommand({replSetGetStatus: 1}));
var opTimeType = "durableOpTime";
if (_isRunningWithoutJournaling(conn)) {
opTimeType = "appliedOpTime";
}
var opTime = replSetStatus.optimes[opTimeType];
if (_isEmptyOpTime(opTime)) {
throw new Error("last durable OpTime is empty -- connection: " + conn);
}
return opTime;
}
/*
* Compares Timestamp objects. Returns true if ts1 is 'earlier' than ts2, else false.
*/
function _isEarlierTimestamp(ts1, ts2) {
if (ts1.getTime() == ts2.getTime()) {
return ts1.getInc() < ts2.getInc();
}
return ts1.getTime() < ts2.getTime();
}
/*
* Returns true if the node can be elected primary of a replica set.
*/
function _isElectable(node) {
return !node.arbiterOnly && (node.priority === undefined || node.priority != 0);
}
/**
* Returns list of nodes as host:port strings.
*/
this.nodeList = function() {
var list = [];
for (var i = 0; i < this.ports.length; i++) {
list.push(this.host + ":" + this.ports[i]);
}
return list;
};
this.getNodeId = function(node) {
if (node.toFixed) {
return parseInt(node);
}
for (var i = 0; i < this.nodes.length; i++) {
if (this.nodes[i] == node) {
return i;
}
}
if (node instanceof ObjectId) {
for (i = 0; i < this.nodes.length; i++) {
if (this.nodes[i].runId == node) {
return i;
}
}
}
if (node.nodeId != null) {
return parseInt(node.nodeId);
}
return undefined;
};
this.getPort = function(n) {
var n = this.getNodeId(n);
return this.ports[n];
};
this._addPath = function(p) {
if (!_alldbpaths)
_alldbpaths = [p];
else
_alldbpaths.push(p);
return p;
};
this.getReplSetConfig = function() {
var cfg = {};
cfg._id = this.name;
cfg.protocolVersion = 1;
cfg.members = [];
for (var i = 0; i < this.ports.length; i++) {
var member = {};
member._id = i;
member.host = this.host;
if (!member.host.contains('/')) {
member.host += ":" + this.ports[i];
}
var nodeOpts = this.nodeOptions["n" + i];
if (nodeOpts) {
if (nodeOpts.arbiter) {
member.arbiterOnly = true;
}
if (nodeOpts.rsConfig) {
Object.extend(member, nodeOpts.rsConfig);
}
}
cfg.members.push(member);
}
if (_configSettings) {
cfg.settings = _configSettings;
}
return cfg;
};
this.getURL = function() {
var hosts = [];
for (var i = 0; i < this.ports.length; i++) {
hosts.push(this.host + ":" + this.ports[i]);
}
return this.name + "/" + hosts.join(",");
};
/**
* Starts each node in the replica set with the given options.
*
* @param options - The options passed to {@link MongoRunner.runMongod}
*/
this.startSet = function(options, restart) {
print("ReplSetTest starting set");
if (options && options.keyFile) {
self.keyFile = options.keyFile;
}
if (options) {
self.startOptions = options;
}
var nodes = [];
for (var n = 0; n < this.ports.length; n++) {
nodes.push(this.start(n, options, restart));
}
this.nodes = nodes;
return this.nodes;
};
/**
* Blocks until the secondary nodes have completed recovery and their roles are known. Blocks on
* all secondary nodes or just 'slaves', if specified.
*/
this.awaitSecondaryNodes = function(timeout, slaves) {
timeout = timeout || self.kDefaultTimeoutMS;
assert.soonNoExcept(function() {
// Reload who the current slaves are
self.getPrimary(timeout);
var slavesToCheck = slaves || self._slaves;
var len = slavesToCheck.length;
var ready = true;
for (var i = 0; i < len; i++) {
var isMaster = slavesToCheck[i].getDB("admin").runCommand({ismaster: 1});
var arbiter = (isMaster.arbiterOnly == undefined ? false : isMaster.arbiterOnly);
ready = ready && (isMaster.secondary || arbiter);
}
return ready;
}, "Awaiting secondaries", timeout);
};
/**
* Blocks until the specified node says it's syncing from the given upstream node.
*/
this.awaitSyncSource = function(node, upstreamNode, timeout) {
print("Waiting for node " + node.name + " to start syncing from " + upstreamNode.name);
var status = null;
assert.soonNoExcept(
function() {
status = node.getDB("admin").runCommand({replSetGetStatus: 1});
for (var j = 0; j < status.members.length; j++) {
if (status.members[j].self) {
return status.members[j].syncingTo === upstreamNode.host;
}
}
return false;
},
"Awaiting node " + node + " syncing from " + upstreamNode + ": " + tojson(status),
timeout);
};
/**
* Blocks until each node agrees that all other nodes have applied the most recent oplog entry.
*/
this.awaitNodesAgreeOnAppliedOpTime = function(timeout, nodes) {
timeout = timeout || self.kDefaultTimeoutMS;
nodes = nodes || self.nodes;
assert.soon(function() {
let appliedOpTimeConsensus = undefined;
for (let i = 0; i < nodes.length; i++) {
let replSetGetStatus;
try {
replSetGetStatus = nodes[i].adminCommand({replSetGetStatus: 1});
} catch (e) {
print("AwaitNodesAgreeOnAppliedOpTime: Retrying because node " + nodes[i].name +
" failed to execute replSetGetStatus: " + tojson(e));
return false;
}
assert.commandWorked(replSetGetStatus);
if (appliedOpTimeConsensus === undefined) {
if (replSetGetStatus.optimes) {
appliedOpTimeConsensus = replSetGetStatus.optimes.appliedOpTime;
} else {
// Older versions of mongod do not include an 'optimes' field in the
// replSetGetStatus response. We instead pull an optime from the first
// replica set member that includes one in its status. All we need here is
// any initial value that we can compare to all the other optimes.
let optimeMembers = replSetGetStatus.members.filter(m => m.optime);
assert(optimeMembers.length > 0,
"AwaitNodesAgreeOnAppliedOpTime: replSetGetStatus did not " +
"include optimes for any members: " + tojson(replSetGetStatus));
appliedOpTimeConsensus = optimeMembers[0].optime;
}
assert(appliedOpTimeConsensus,
"AwaitNodesAgreeOnAppliedOpTime: missing appliedOpTime in " +
"replSetGetStatus: " + tojson(replSetGetStatus));
}
if (replSetGetStatus.optimes &&
!friendlyEqual(replSetGetStatus.optimes.appliedOpTime,
appliedOpTimeConsensus)) {
print("AwaitNodesAgreeOnAppliedOpTime: Retrying because node " + nodes[i].name +
" has appliedOpTime " + tojson(replSetGetStatus.optimes.appliedOpTime) +
" that does not match the previously observed appliedOpTime " +
tojson(appliedOpTimeConsensus));
return false;
}
for (let j = 0; j < replSetGetStatus.members.length; j++) {
if (replSetGetStatus.members[j].state == ReplSetTest.State.ARBITER) {
// ARBITER nodes do not apply oplog entries and do not have an 'optime'
// field.
continue;
}
if (!friendlyEqual(replSetGetStatus.members[j].optime,
appliedOpTimeConsensus)) {
print("AwaitNodesAgreeOnAppliedOpTime: Retrying because node " +
nodes[i].name + " sees optime " +
tojson(replSetGetStatus.members[j].optime) + " on node " +
replSetGetStatus.members[j].name + " but expects to see optime " +
tojson(appliedOpTimeConsensus));
return false;
}
}
}
print(
"AwaitNodesAgreeOnAppliedOpTime: All nodes agree that all ops are applied up to " +
tojson(appliedOpTimeConsensus));
return true;
}, "Awaiting nodes to agree that all ops are applied across replica set", timeout);
};
this._findHighestPriorityNodes = function(config) {
let highestPriority = 0;
let highPriorityNodes = [];
for (let i = 0; i < config.members.length; i++) {
const member = config.members[i];
if (member.priority > highestPriority) {
highestPriority = member.priority;
highPriorityNodes = [this.nodes[i]];
} else if (member.priority === highestPriority) {
highPriorityNodes.push(this.nodes[i]);
}
}
return highPriorityNodes;
};
/**
* Blocks until the node with the highest priority is the primary. If there are multiple
* nodes tied for highest priority, waits until one of them is the primary.
*/
this.awaitHighestPriorityNodeIsPrimary = function(timeout) {
timeout = timeout || self.kDefaultTimeoutMS;
// First figure out the set of highest priority nodes.
const config = asCluster(this.nodes, () => self.getReplSetConfigFromNode());
const highPriorityNodes = this._findHighestPriorityNodes(config);
// Now wait for the primary to be one of the highest priority nodes.
assert.soon(
function() {
return highPriorityNodes.includes(self.getPrimary());
},
function() {
return "Expected primary to be one of: " + tojson(highPriorityNodes) +
", but found primary to be: " + tojson(self.getPrimary());
},
timeout);
// Finally wait for all nodes to agree on the primary.
this.awaitNodesAgreeOnPrimary(timeout);
const primary = this.getPrimary();
assert(highPriorityNodes.includes(primary),
"Primary switched away from highest priority node. Found primary: " +
tojson(primary) + ", but expected one of: " + tojson(highPriorityNodes));
};
/**
* Blocks until all nodes agree on who the primary is.
* If 'expectedPrimaryNodeId' is provided, ensure that every node is seeing this node as the
* primary. Otherwise, ensure that all the nodes in the set agree with the first node on the
* identity of the primary.
*/
this.awaitNodesAgreeOnPrimary = function(timeout, nodes, expectedPrimaryNodeId) {
timeout = timeout || self.kDefaultTimeoutMS;
nodes = nodes || self.nodes;
expectedPrimaryNodeId = expectedPrimaryNodeId || -1;
if (expectedPrimaryNodeId === -1) {
print("AwaitNodesAgreeOnPrimary: Waiting for nodes to agree on any primary.");
} else {
print("AwaitNodesAgreeOnPrimary: Waiting for nodes to agree on " +
nodes[expectedPrimaryNodeId].name + " as primary.");
}
assert.soonNoExcept(function() {
var primary = expectedPrimaryNodeId;
for (var i = 0; i < nodes.length; i++) {
var replSetGetStatus =
assert.commandWorked(nodes[i].getDB("admin").runCommand({replSetGetStatus: 1}));
var nodesPrimary = -1;
for (var j = 0; j < replSetGetStatus.members.length; j++) {
if (replSetGetStatus.members[j].state === ReplSetTest.State.PRIMARY) {
// Node sees two primaries.
if (nodesPrimary !== -1) {
print("AwaitNodesAgreeOnPrimary: Retrying because " + nodes[i].name +
" thinks both " + nodes[nodesPrimary].name + " and " +
nodes[j].name + " are primary.");
return false;
}
nodesPrimary = j;
}
}
// Node doesn't see a primary.
if (nodesPrimary < 0) {
print("AwaitNodesAgreeOnPrimary: Retrying because " + nodes[i].name +
" does not see a primary.");
return false;
}
if (primary < 0) {
// If we haven't seen a primary yet, set it to this.
primary = nodesPrimary;
} else if (primary !== nodesPrimary) {
print("AwaitNodesAgreeOnPrimary: Retrying because " + nodes[i].name +
" thinks the primary is " + nodes[nodesPrimary].name + " instead of " +
nodes[primary].name);
return false;
}
}
print("AwaitNodesAgreeOnPrimary: Nodes agreed on primary " + nodes[primary].name);
return true;
}, "Awaiting nodes to agree on primary", timeout);
};
/**
* Blocking call, which will wait for a primary to be elected and become master for some
* pre-defined timeout. If a primary is available it will return a connection to it.
* Otherwise throws an exception.
*/
this.getPrimary = function(timeout) {
timeout = timeout || self.kDefaultTimeoutMS;
var primary = null;
assert.soonNoExcept(function() {
primary = _callIsMaster();
return primary;
}, "Finding primary", timeout);
return primary;
};
this.awaitNoPrimary = function(msg, timeout) {
msg = msg || "Timed out waiting for there to be no primary in replset: " + this.name;
timeout = timeout || self.kDefaultTimeoutMS;
assert.soonNoExcept(function() {
return _callIsMaster() == false;
}, msg, timeout);
};
this.getSecondaries = function(timeout) {
var master = this.getPrimary(timeout);
var secs = [];
for (var i = 0; i < this.nodes.length; i++) {
if (this.nodes[i] != master) {
secs.push(this.nodes[i]);
}
}
return secs;
};
this.getSecondary = function(timeout) {
return this.getSecondaries(timeout)[0];
};
function isNodeArbiter(node) {
return node.getDB('admin').isMaster('admin').arbiterOnly;
}
this.getArbiters = function() {
let arbiters = [];
for (let i = 0; i < this.nodes.length; i++) {
const node = this.nodes[i];
let isArbiter = false;
assert.retryNoExcept(() => {
isArbiter = isNodeArbiter(node);
return true;
}, `Could not call 'isMaster' on ${node}.`, 3, 1000);
if (isArbiter) {
arbiters.push(node);
}
}
return arbiters;
};
this.getArbiter = function() {
return this.getArbiters()[0];
};
this.status = function(timeout) {
var master = _callIsMaster();
if (!master) {
master = this._liveNodes[0];
}
return master.getDB("admin").runCommand({replSetGetStatus: 1});
};
/**
* Adds a node to the replica set managed by this instance.
*/
this.add = function(config) {
var nextPort = allocatePort();
print("ReplSetTest Next port: " + nextPort);
this.ports.push(nextPort);
printjson(this.ports);
if (_useBridge) {
_unbridgedPorts.push(allocatePort());
}
var nextId = this.nodes.length;
printjson(this.nodes);
print("ReplSetTest nextId: " + nextId);
return this.start(nextId, config);
};
this.remove = function(nodeId) {
nodeId = this.getNodeId(nodeId);
this.nodes.splice(nodeId, 1);
this.ports.splice(nodeId, 1);
if (_useBridge) {
_unbridgedPorts.splice(nodeId, 1);
_unbridgedNodes.splice(nodeId, 1);
}
};
/*
* If journaling is disabled or we are using an ephemeral storage engine, set
* 'writeConcernMajorityJournalDefault' to false for the given 'config' object. If the
* 'writeConcernMajorityJournalDefault' field is already set, it does not override it,
* and returns the 'config' object unchanged. Does not affect 'config' when running CSRS.
*/
this._updateConfigIfNotDurable = function(config) {
// Get a replica set node (check for use of bridge).
var replNode = _useBridge ? _unbridgedNodes[0] : this.nodes[0];
// Don't update replset config for sharding config servers since config servers always
// require durable storage.
if (replNode.hasOwnProperty("fullOptions") &&
replNode.fullOptions.hasOwnProperty("configsvr")) {
return config;
}
// Don't override existing value.
var wcMajorityJournalField = "writeConcernMajorityJournalDefault";
if (config.hasOwnProperty(wcMajorityJournalField)) {
return config;
}
// Check journaling by sending commands through the bridge if it's used.
if (_isRunningWithoutJournaling(this.nodes[0])) {
config[wcMajorityJournalField] = false;
}
return config;
};
this._setDefaultConfigOptions = function(config) {
// Update config for non journaling test variants
this._updateConfigIfNotDurable(config);
// Add protocolVersion if missing
if (!config.hasOwnProperty('protocolVersion')) {
config['protocolVersion'] = 1;
}
};
/**
* Runs replSetInitiate on the first node of the replica set.
* Ensures that a primary is elected (not necessarily node 0).
* initiate() should be preferred instead of this, but this is useful when the connections
* aren't authorized to run replSetGetStatus.
* TODO(SERVER-14017): remove this in favor of using initiate() everywhere.
*/
this.initiateWithAnyNodeAsPrimary = function(cfg, initCmd, {
doNotWaitForStableRecoveryTimestamp: doNotWaitForStableRecoveryTimestamp = false
} = {}) {
var master = this.nodes[0].getDB("admin");
var config = cfg || this.getReplSetConfig();
var cmd = {};
var cmdKey = initCmd || 'replSetInitiate';
// Throw an exception if nodes[0] is unelectable in the given config.
if (!_isElectable(config.members[0])) {
throw Error("The node at index 0 must be electable");
}
// Start up a single node replica set then reconfigure to the correct size (if the config
// contains more than 1 node), so the primary is elected more quickly.
var originalMembers, originalSettings;
if (config.members && config.members.length > 1) {
originalMembers = config.members.slice();
config.members = config.members.slice(0, 1);
originalSettings = config.settings;
delete config.settings; // Clear settings to avoid tags referencing sliced nodes.
}
this._setDefaultConfigOptions(config);
cmd[cmdKey] = config;
printjson(cmd);
assert.commandWorked(master.runCommand(cmd), tojson(cmd));
this.getPrimary(); // Blocks until there is a primary.
// Reconfigure the set to contain the correct number of nodes (if necessary).
if (originalMembers) {
config.members = originalMembers;
if (originalSettings) {
config.settings = originalSettings;
}
config.version = 2;
// Nodes started with the --configsvr flag must have configsvr = true in their config.
if (this.nodes[0].hasOwnProperty("fullOptions") &&
this.nodes[0].fullOptions.hasOwnProperty("configsvr")) {
config.configsvr = true;
}
cmd = {replSetReconfig: config};
print("Reconfiguring replica set to add in other nodes");
printjson(cmd);
// replSetInitiate and replSetReconfig commands can fail with a NodeNotFound error
// if a heartbeat times out during the quorum check.
// They may also fail with NewReplicaSetConfigurationIncompatible on similar timeout
// during the config validation stage while deducing isSelf().
// This can fail with an InterruptedDueToStepDown error when interrupted.
// We retry three times to reduce the chance of failing this way.
assert.retry(() => {
var res;
try {
res = master.runCommand(cmd);
if (res.ok === 1) {
return true;
}
} catch (e) {
// reconfig can lead to a stepdown if the primary looks for a majority before
// a majority of nodes have successfully joined the set. If there is a stepdown
// then the reconfig request will be killed and respond with a network error.
if (isNetworkError(e)) {
return true;
}
throw e;
}
assert.commandFailedWithCode(res,