-
Notifications
You must be signed in to change notification settings - Fork 115
/
pcs.8.in
1677 lines (1548 loc) · 111 KB
/
pcs.8.in
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
.TH PCS "8" "@UTC_DATE@" "pcs @VERSION@" "System Administration Utilities"
.SH NAME
pcs \- pacemaker/corosync configuration system
.SH SYNOPSIS
.B pcs
[\fI\-f file\fR] [\fI\-h\fR] [\fIcommands\fR]...
.SH DESCRIPTION
Control and configure pacemaker and corosync.
.SH OPTIONS
.TP
\fB\-h\fR, \fB\-\-help\fR
Display usage and exit.
.TP
\fB\-f\fR file
Perform actions on file instead of active CIB.
.br
Commands supporting the option use the initial state of the specified file as their input and then overwrite the file with the state reflecting the requested operation(s).
.br
A few commands only use the specified file in read\-only mode since their effect is not a CIB modification.
.TP
\fB\-\-debug\fR
Print all network traffic and external commands run.
.TP
\fB\-\-version\fR
Print pcs version information. List pcs capabilities if \fB\-\-full\fR is specified.
.TP
\fB\-\-request\-timeout\fR=<timeout>
Timeout for each outgoing request to another node in seconds. Default is 60s.
.SS "Commands:"
.TP
cluster
Configure cluster options and nodes.
.TP
resource
Manage cluster resources.
.TP
stonith
Manage fence devices.
.TP
constraint
Manage resource constraints.
.TP
property
Manage pacemaker properties.
.TP
acl
Manage pacemaker access control lists.
.TP
qdevice
Manage quorum device provider on the local host.
.TP
quorum
Manage cluster quorum settings.
.TP
booth
Manage booth (cluster ticket manager).
.TP
status
View cluster status.
.TP
config
View and manage cluster configuration.
.TP
pcsd
Manage pcs daemon.
.TP
host
Manage hosts known to pcs/pcsd.
.TP
node
Manage cluster nodes.
.TP
alert
Manage pacemaker alerts.
.TP
client
Manage pcsd client configuration.
.TP
dr
Manage disaster recovery configuration.
.TP
tag
Manage pacemaker tags.
.SS "resource"
.TP
[status [<resource id | tag id>] [node=<node>] [\fB\-\-hide\-inactive\fR]]
Show status of all currently configured resources. If \fB\-\-hide\-inactive\fR is specified, only show active resources. If a resource or tag id is specified, only show status of the specified resource or resources in the specified tag. If node is specified, only show status of resources configured for the specified node.
.TP
config [<resource id>]...
Show options of all currently configured resources or if resource ids are specified show the options for the specified resource ids.
.TP
list [filter] [\fB\-\-nodesc\fR]
Show list of all available resource agents (if filter is provided then only resource agents matching the filter will be shown). If \fB\-\-nodesc\fR is used then descriptions of resource agents are not printed.
.TP
describe [<standard>:[<provider>:]]<type> [\fB\-\-full\fR]
Show options for the specified resource. If \fB\-\-full\fR is specified, all options including advanced and deprecated ones are shown.
.TP
create <resource id> [<standard>:[<provider>:]]<type> [resource options] [\fBop\fR <operation action> <operation options> [<operation action> <operation options>]...] [\fBmeta\fR <meta options>...] [\fBclone\fR [<clone id>] [<clone options>] | promotable [<clone id>] [<promotable options>] | \fB\-\-group\fR <group id> [\fB\-\-before\fR <resource id> | \fB\-\-after\fR <resource id>] | \fBbundle\fR <bundle id>] [\fB\-\-disabled\fR] [\fB\-\-no\-default\-ops] [\fB\-\-wait\fR[=n]]
Create specified resource. If \fBclone\fR is used a clone resource is created. If \fBpromotable\fR is used a promotable clone resource is created. If \fB\-\-group\fR is specified the resource is added to the group named. You can use \fB\-\-before\fR or \fB\-\-after\fR to specify the position of the added resource relatively to some resource already existing in the group. If \fBbundle\fR is specified, resource will be created inside of the specified bundle. If \fB\-\-disabled\fR is specified the resource is not started automatically. If \fB\-\-no\-default\-ops\fR is specified, only monitor operations are created for the resource and all other operations use default settings. If \fB\-\-wait\fR is specified, pcs will wait up to 'n' seconds for the resource to start and then return 0 if the resource is started, or 1 if the resource has not yet started. If 'n' is not specified it defaults to 60 minutes.
Example: Create a new resource called 'VirtualIP' with IP address 192.168.0.99, netmask of 32, monitored everything 30 seconds, on eth2: pcs resource create VirtualIP ocf:heartbeat:IPaddr2 ip=192.168.0.99 cidr_netmask=32 nic=eth2 op monitor interval=30s
.TP
delete <resource id|group id|bundle id|clone id>
Deletes the resource, group, bundle or clone (and all resources within the group/bundle/clone).
.TP
remove <resource id|group id|bundle id|clone id>
Deletes the resource, group, bundle or clone (and all resources within the group/bundle/clone).
.TP
enable <resource id | tag id>... [\fB\-\-wait\fR[=n]]
Allow the cluster to start the resources. Depending on the rest of the configuration (constraints, options, failures, etc), the resources may remain stopped. If \fB\-\-wait\fR is specified, pcs will wait up to 'n' seconds for the resources to start and then return 0 if the resources are started, or 1 if the resources have not yet started. If 'n' is not specified it defaults to 60 minutes.
.TP
disable <resource id | tag id>... [\fB\-\-safe\fR [\fB\-\-brief\fR] [\fB\-\-no\-strict\fR]] [\fB\-\-simulate\fR [\fB\-\-brief\fR]] [\fB\-\-wait\fR[=n]]
Attempt to stop the resources if they are running and forbid the cluster from starting them again. Depending on the rest of the configuration (constraints, options, failures, etc), the resources may remain started.
.br
If \fB\-\-safe\fR is specified, no changes to the cluster configuration will be made if other than specified resources would be affected in any way. If \fB\-\-brief\fR is also specified, only errors are printed.
.br
If \fB\-\-no\-strict\fR is specified, no changes to the cluster configuration will be made if other than specified resources would get stopped or demoted. Moving resources between nodes is allowed.
.br
If \fB\-\-simulate\fR is specified, no changes to the cluster configuration will be made and the effect of the changes will be printed instead. If \fB\-\-brief\fR is also specified, only a list of affected resources will be printed.
.br
If \fB\-\-wait\fR is specified, pcs will wait up to 'n' seconds for the resources to stop and then return 0 if the resources are stopped or 1 if the resources have not stopped. If 'n' is not specified it defaults to 60 minutes.
.TP
safe\-disable <resource id | tag id>... [\fB\-\-brief\fR] [\fB\-\-no\-strict\fR] [\fB\-\-simulate\fR [\fB\-\-brief\fR]] [\fB\-\-wait\fR[=n]] [\fB\-\-force\fR]
Attempt to stop the resources if they are running and forbid the cluster from starting them again. Depending on the rest of the configuration (constraints, options, failures, etc), the resources may remain started. No changes to the cluster configuration will be made if other than specified resources would be affected in any way.
.br
If \fB\-\-brief\fR is specified, only errors are printed.
.br
If \fB\-\-no\-strict\fR is specified, no changes to the cluster configuration will be made if other than specified resources would get stopped or demoted. Moving resources between nodes is allowed.
.br
If \fB\-\-simulate\fR is specified, no changes to the cluster configuration will be made and the effect of the changes will be printed instead. If \fB\-\-brief\fR is also specified, only a list of affected resources will be printed.
.br
If \fB\-\-wait\fR is specified, pcs will wait up to 'n' seconds for the resources to stop and then return 0 if the resources are stopped or 1 if the resources have not stopped. If 'n' is not specified it defaults to 60 minutes.
.br
If \fB\-\-force\fR is specified, checks for safe disable will be skipped.
.TP
restart <resource id> [node] [\fB\-\-wait\fR=n]
Restart the resource specified. If a node is specified and if the resource is a clone or bundle it will be restarted only on the node specified. If \fB\-\-wait\fR is specified, then we will wait up to 'n' seconds for the resource to be restarted and return 0 if the restart was successful or 1 if it was not.
.TP
debug\-start <resource id> [\fB\-\-full\fR]
This command will force the specified resource to start on this node ignoring the cluster recommendations and print the output from starting the resource. Using \fB\-\-full\fR will give more detailed output. This is mainly used for debugging resources that fail to start.
.TP
debug\-stop <resource id> [\fB\-\-full\fR]
This command will force the specified resource to stop on this node ignoring the cluster recommendations and print the output from stopping the resource. Using \fB\-\-full\fR will give more detailed output. This is mainly used for debugging resources that fail to stop.
.TP
debug\-promote <resource id> [\fB\-\-full\fR]
This command will force the specified resource to be promoted on this node ignoring the cluster recommendations and print the output from promoting the resource. Using \fB\-\-full\fR will give more detailed output. This is mainly used for debugging resources that fail to promote.
.TP
debug\-demote <resource id> [\fB\-\-full\fR]
This command will force the specified resource to be demoted on this node ignoring the cluster recommendations and print the output from demoting the resource. Using \fB\-\-full\fR will give more detailed output. This is mainly used for debugging resources that fail to demote.
.TP
debug\-monitor <resource id> [\fB\-\-full\fR]
This command will force the specified resource to be monitored on this node ignoring the cluster recommendations and print the output from monitoring the resource. Using \fB\-\-full\fR will give more detailed output. This is mainly used for debugging resources that fail to be monitored.
.TP
move <resource id> [destination node] [\fB\-\-promoted\fR] [\fB\-\-strict\fR] [\fB\-\-wait\fR[=n]]
Move the resource off the node it is currently running on. This is achieved by creating a \-INFINITY location constraint to ban the node. If destination node is specified the resource will be moved to that node by creating an INFINITY location constraint to prefer the destination node. The constraint needed for moving the resource will be automatically removed once the resource is running on it's new location. The command will fail in case it is not possible to verify that the resource will not be moved back after deleting the constraint.
If \fB\-\-strict\fR is specified, the command will also fail if other resources would be affected.
If \fB\-\-promoted\fR is used the scope of the command is limited to the Promoted role and promotable clone id must be used (instead of the resource id).
If \fB\-\-wait\fR is specified, pcs will wait up to 'n' seconds for the resource to move and then return 0 on success or 1 on error. If 'n' is not specified it defaults to 60 minutes.
NOTE: This command has been changed in pcs\-0.11. It is equivalent to command 'resource move <resource id> \fB\-\-autodelete\fR' from pcs\-0.10.9. Legacy functionality of the 'resource move' command is still available as 'resource move\-with\-constraint <resource id>'.
If you want the resource to preferably avoid running on some nodes but be able to failover to them use 'pcs constraint location avoids'.
.TP
move\-with\-constraint <resource id> [destination node] [lifetime=<lifetime>] [\fB\-\-promoted\fR] [\fB\-\-wait\fR[=n]]
Move the resource off the node it is currently running on by creating a \-INFINITY location constraint to ban the node. If destination node is specified the resource will be moved to that node by creating an INFINITY location constraint to prefer the destination node.
If lifetime is specified then the constraint will expire after that time, otherwise it defaults to infinity and the constraint can be cleared manually with 'pcs resource clear' or 'pcs constraint delete'. Lifetime is expected to be specified as ISO 8601 duration (see https://en.wikipedia.org/wiki/ISO_8601#Durations).
If \fB\-\-promoted\fR is used the scope of the command is limited to the Promoted role and promotable clone id must be used (instead of the resource id).
If \fB\-\-wait\fR is specified, pcs will wait up to 'n' seconds for the resource to move and then return 0 on success or 1 on error. If 'n' is not specified it defaults to 60 minutes.
If you want the resource to preferably avoid running on some nodes but be able to failover to them use 'pcs constraint location avoids'.
.TP
ban <resource id> [node] [\fB\-\-promoted\fR] [lifetime=<lifetime>] [\fB\-\-wait\fR[=n]]
Prevent the resource id specified from running on the node (or on the current node it is running on if no node is specified) by creating a \-INFINITY location constraint.
If \fB\-\-promoted\fR is used the scope of the command is limited to the Promoted role and promotable clone id must be used (instead of the resource id).
If lifetime is specified then the constraint will expire after that time, otherwise it defaults to infinity and the constraint can be cleared manually with 'pcs resource clear' or 'pcs constraint delete'. Lifetime is expected to be specified as ISO 8601 duration (see https://en.wikipedia.org/wiki/ISO_8601#Durations).
If \fB\-\-wait\fR is specified, pcs will wait up to 'n' seconds for the resource to move and then return 0 on success or 1 on error. If 'n' is not specified it defaults to 60 minutes.
If you want the resource to preferably avoid running on some nodes but be able to failover to them use 'pcs constraint location avoids'.
.TP
clear <resource id> [node] [\fB\-\-promoted\fR] [\fB\-\-expired\fR] [\fB\-\-wait\fR[=n]]
Remove constraints created by move and/or ban on the specified resource (and node if specified).
If \fB\-\-promoted\fR is used the scope of the command is limited to the Promoted role and promotable clone id must be used (instead of the resource id).
If \fB\-\-expired\fR is specified, only constraints with expired lifetimes will be removed.
If \fB\-\-wait\fR is specified, pcs will wait up to 'n' seconds for the operation to finish (including starting and/or moving resources if appropriate) and then return 0 on success or 1 on error. If 'n' is not specified it defaults to 60 minutes.
.TP
standards
List available resource agent standards supported by this installation (OCF, LSB, etc.).
.TP
providers
List available OCF resource agent providers.
.TP
agents [standard[:provider]]
List available agents optionally filtered by standard and provider.
.TP
update <resource id> [resource options] [op [<operation action> <operation options>]...] [meta <meta operations>...] [\fB\-\-wait\fR[=n]]
Add, remove or change options of specified resource, clone or multi\-state resource. Unspecified options will be kept unchanged. If you wish to remove an option, set it to empty value, i.e. 'option_name='.
If an operation (op) is specified it will update the first found operation with the same action on the specified resource. If no operation with that action exists then a new operation will be created. (WARNING: all existing options on the updated operation will be reset if not specified.) If you want to create multiple monitor operations you should use the 'op add' & 'op remove' commands.
If \fB\-\-wait\fR is specified, pcs will wait up to 'n' seconds for the changes to take effect and then return 0 if the changes have been processed or 1 otherwise. If 'n' is not specified it defaults to 60 minutes.
.TP
op add <resource id> <operation action> [operation properties]
Add operation for specified resource.
.TP
op delete <resource id> <operation action> [<operation properties>...]
Remove specified operation (note: you must specify the exact operation properties to properly remove an existing operation).
.TP
op delete <operation id>
Remove the specified operation id.
.TP
op remove <resource id> <operation action> [<operation properties>...]
Remove specified operation (note: you must specify the exact operation properties to properly remove an existing operation).
.TP
op remove <operation id>
Remove the specified operation id.
.TP
op defaults [config] [\fB\-\-all\fR] [\fB\-\-full\fR] [\fB\-\-no\-check\-expired\fR]
List currently configured default values for operations. If \fB\-\-all\fR is specified, also list expired sets of values. If \fB\-\-full\fR is specified, also list ids. If \fB\-\-no\-expire\-check\fR is specified, do not evaluate whether sets of values are expired.
.TP
op defaults <name>=<value>...
Set default values for operations.
.br
NOTE: Defaults do not apply to resources / stonith devices which override them with their own defined values.
.TP
op defaults set create [<set options>] [meta [<name>=<value>]...] [rule [<expression>]]
Create a new set of default values for resource / stonith device operations. You may specify a rule describing resources / stonith devices and / or operations to which the set applies.
Set options are: id, score
Expression looks like one of the following:
.br
op <operation name> [interval=<interval>]
.br
resource [<standard>]:[<provider>]:[<type>]
.br
defined|not_defined <node attribute>
.br
<node attribute> lt|gt|lte|gte|eq|ne [string|integer|number|version] <value>
.br
date gt|lt <date>
.br
date in_range [<date>] to <date>
.br
date in_range <date> to duration <duration options>
.br
date\-spec <date\-spec options>
.br
<expression> and|or <expression>
.br
(<expression>)
You may specify all or any of 'standard', 'provider' and 'type' in a resource expression. For example: 'resource ocf::' matches all resources of 'ocf' standard, while 'resource ::Dummy' matches all resources of 'Dummy' type regardless of their standard and provider.
Dates are expected to conform to ISO 8601 format.
Duration options are: hours, monthdays, weekdays, yearsdays, months, weeks, years, weekyears, moon. Value for these options is an integer.
Date\-spec options are: hours, monthdays, weekdays, yearsdays, months, weeks, years, weekyears, moon. Value for these options is an integer or a range written as integer\-integer.
NOTE: Defaults do not apply to resources / stonith devices which override them with their own defined values.
.TP
op defaults set delete [<set id>]...
Delete specified options sets.
.TP
op defaults set remove [<set id>]...
Delete specified options sets.
.TP
op defaults set update <set id> [meta [<name>=<value>]...]
Add, remove or change values in specified set of default values for resource / stonith device operations. Unspecified options will be kept unchanged. If you wish to remove an option, set it to empty value, i.e. 'option_name='.
NOTE: Defaults do not apply to resources / stonith devices which override them with their own defined values.
.TP
op defaults update <name>=<value>...
Add, remove or change default values for operations. This is a simplified command useful for cases when you only manage one set of default values. Unspecified options will be kept unchanged. If you wish to remove an option, set it to empty value, i.e. 'option_name='.
NOTE: Defaults do not apply to resources / stonith devices which override them with their own defined values.
.TP
meta <resource id | group id | clone id> <meta options> [\fB\-\-wait\fR[=n]]
Add specified options to the specified resource, group or clone. Meta options should be in the format of name=value, options may be removed by setting an option without a value. If \fB\-\-wait\fR is specified, pcs will wait up to 'n' seconds for the changes to take effect and then return 0 if the changes have been processed or 1 otherwise. If 'n' is not specified it defaults to 60 minutes.
.br
Example: pcs resource meta TestResource failure\-timeout=50 resource\-stickiness=
.TP
group list
Show all currently configured resource groups and their resources.
.TP
group add <group id> <resource id> [resource id] ... [resource id] [\fB\-\-before\fR <resource id> | \fB\-\-after\fR <resource id>] [\fB\-\-wait\fR[=n]]
Add the specified resource to the group, creating the group if it does not exist. If the resource is present in another group it is moved to the new group. If the group remains empty after move, it is deleted (for cloned groups, the clone is deleted as well). The delete operation may fail in case the group is referenced within the configuration, e.g. by constraints. In that case, use 'pcs resource ungroup' command prior to moving all resources out of the group.
You can use \fB\-\-before\fR or \fB\-\-after\fR to specify the position of the added resources relatively to some resource already existing in the group. By adding resources to a group they are already in and specifying \fB\-\-after\fR or \fB\-\-before\fR you can move the resources in the group.
If \fB\-\-wait\fR is specified, pcs will wait up to 'n' seconds for the operation to finish (including moving resources if appropriate) and then return 0 on success or 1 on error. If 'n' is not specified it defaults to 60 minutes.
.TP
group delete <group id> [resource id]... [\fB\-\-wait\fR[=n]]
Remove the group (note: this does not remove any resources from the cluster) or if resources are specified, remove the specified resources from the group. If \fB\-\-wait\fR is specified, pcs will wait up to 'n' seconds for the operation to finish (including moving resources if appropriate) and the return 0 on success or 1 on error. If 'n' is not specified it defaults to 60 minutes.
.TP
group remove <group id> [resource id]... [\fB\-\-wait\fR[=n]]
Remove the group (note: this does not remove any resources from the cluster) or if resources are specified, remove the specified resources from the group. If \fB\-\-wait\fR is specified, pcs will wait up to 'n' seconds for the operation to finish (including moving resources if appropriate) and the return 0 on success or 1 on error. If 'n' is not specified it defaults to 60 minutes.
.TP
ungroup <group id> [resource id]... [\fB\-\-wait\fR[=n]]
Remove the group (note: this does not remove any resources from the cluster) or if resources are specified, remove the specified resources from the group. If \fB\-\-wait\fR is specified, pcs will wait up to 'n' seconds for the operation to finish (including moving resources if appropriate) and the return 0 on success or 1 on error. If 'n' is not specified it defaults to 60 minutes.
.TP
clone <resource id | group id> [<clone id>] [clone options]... [\fB\-\-wait\fR[=n]]
Set up the specified resource or group as a clone. If \fB\-\-wait\fR is specified, pcs will wait up to 'n' seconds for the operation to finish (including starting clone instances if appropriate) and then return 0 on success or 1 on error. If 'n' is not specified it defaults to 60 minutes.
.TP
promotable <resource id | group id> [<clone id>] [clone options]... [\fB\-\-wait\fR[=n]]
Set up the specified resource or group as a promotable clone. This is an alias for 'pcs resource clone <resource id> promotable=true'.
.TP
unclone <clone id | resource id | group id> [\fB\-\-wait\fR[=n]]
Remove the specified clone or the clone which contains the specified group or resource (the resource or group will not be removed). If \fB\-\-wait\fR is specified, pcs will wait up to 'n' seconds for the operation to finish (including stopping clone instances if appropriate) and then return 0 on success or 1 on error. If 'n' is not specified it defaults to 60 minutes.
.TP
bundle create <bundle id> container <container type> [<container options>] [network <network options>] [port\-map <port options>]... [storage\-map <storage options>]... [meta <meta options>] [\fB\-\-disabled\fR] [\fB\-\-wait\fR[=n]]
Create a new bundle encapsulating no resources. The bundle can be used either as it is or a resource may be put into it at any time. If \fB\-\-disabled\fR is specified, the bundle is not started automatically. If \fB\-\-wait\fR is specified, pcs will wait up to 'n' seconds for the bundle to start and then return 0 on success or 1 on error. If 'n' is not specified it defaults to 60 minutes.
.TP
bundle reset <bundle id> [container <container options>] [network <network options>] [port\-map <port options>]... [storage\-map <storage options>]... [meta <meta options>] [\fB\-\-disabled\fR] [\fB\-\-wait\fR[=n]]
Configure specified bundle with given options. Unlike bundle update, this command resets the bundle according given options - no previous options are kept. Resources inside the bundle are kept as they are. If \fB\-\-disabled\fR is specified, the bundle is not started automatically. If \fB\-\-wait\fR is specified, pcs will wait up to 'n' seconds for the bundle to start and then return 0 on success or 1 on error. If 'n' is not specified it defaults to 60 minutes.
.TP
bundle update <bundle id> [container <container options>] [network <network options>] [port\-map (add <port options>) | (delete | remove <id>...)]... [storage\-map (add <storage options>) | (delete | remove <id>...)]... [meta <meta options>] [\fB\-\-wait\fR[=n]]
Add, remove or change options of specified bundle. Unspecified options will be kept unchanged. If you wish to remove an option, set it to empty value, i.e. 'option_name='.
If you wish to update a resource encapsulated in the bundle, use the 'pcs resource update' command instead and specify the resource id.
If \fB\-\-wait\fR is specified, pcs will wait up to 'n' seconds for the operation to finish (including moving resources if appropriate) and then return 0 on success or 1 on error. If 'n' is not specified it defaults to 60 minutes.
.TP
manage <resource id | tag id>... [\fB\-\-monitor\fR]
Set resources listed to managed mode (default). If \fB\-\-monitor\fR is specified, enable all monitor operations of the resources.
.TP
unmanage <resource id | tag id>... [\fB\-\-monitor\fR]
Set resources listed to unmanaged mode. When a resource is in unmanaged mode, the cluster is not allowed to start nor stop the resource. If \fB\-\-monitor\fR is specified, disable all monitor operations of the resources.
.TP
defaults [config] [\fB\-\-all\fR] [\fB\-\-full\fR] [\fB\-\-no\-check\-expired\fR]
List currently configured default values for resources / stonith devices. If \fB\-\-all\fR is specified, also list expired sets of values. If \fB\-\-full\fR is specified, also list ids. If \fB\-\-no\-expire\-check\fR is specified, do not evaluate whether sets of values are expired.
.TP
defaults <name>=<value>...
Set default values for resources / stonith devices.
.br
NOTE: Defaults do not apply to resources / stonith devices which override them with their own defined values.
.TP
defaults set create [<set options>] [meta [<name>=<value>]...] [rule [<expression>]]
Create a new set of default values for resources / stonith devices. You may specify a rule describing resources / stonith devices to which the set applies.
Set options are: id, score
Expression looks like one of the following:
.br
resource [<standard>]:[<provider>]:[<type>]
.br
date gt|lt <date>
.br
date in_range [<date>] to <date>
.br
date in_range <date> to duration <duration options>
.br
date\-spec <date\-spec options>
.br
<expression> and|or <expression>
.br
(<expression>)
You may specify all or any of 'standard', 'provider' and 'type' in a resource expression. For example: 'resource ocf::' matches all resources of 'ocf' standard, while 'resource ::Dummy' matches all resources of 'Dummy' type regardless of their standard and provider.
Dates are expected to conform to ISO 8601 format.
Duration options are: hours, monthdays, weekdays, yearsdays, months, weeks, years, weekyears, moon. Value for these options is an integer.
Date\-spec options are: hours, monthdays, weekdays, yearsdays, months, weeks, years, weekyears, moon. Value for these options is an integer or a range written as integer\-integer.
NOTE: Defaults do not apply to resources / stonith devices which override them with their own defined values.
.TP
defaults set delete [<set id>]...
Delete specified options sets.
.TP
defaults set remove [<set id>]...
Delete specified options sets.
.TP
defaults set update <set id> [meta [<name>=<value>]...]
Add, remove or change values in specified set of default values for resources / stonith devices. Unspecified options will be kept unchanged. If you wish to remove an option, set it to empty value, i.e. 'option_name='.
NOTE: Defaults do not apply to resources / stonith devices which override them with their own defined values.
.TP
defaults update <name>=<value>...
Add, remove or change default values for resources / stonith devices. This is a simplified command useful for cases when you only manage one set of default values. Unspecified options will be kept unchanged. If you wish to remove an option, set it to empty value, i.e. 'option_name='.
NOTE: Defaults do not apply to resources / stonith devices which override them with their own defined values.
.TP
cleanup [<resource id | stonith id>] [node=<node>] [operation=<operation> [interval=<interval>]] [\fB\-\-strict\fR]
Make the cluster forget failed operations from history of the resource / stonith device and re\-detect its current state. This can be useful to purge knowledge of past failures that have since been resolved.
If the named resource is part of a group, or one numbered instance of a clone or bundled resource, the clean\-up applies to the whole collective resource unless \fB\-\-strict\fR is given.
If a resource id / stonith id is not specified then all resources / stonith devices will be cleaned up.
If a node is not specified then resources / stonith devices on all nodes will be cleaned up.
.TP
refresh [<resource id | stonith id>] [node=<node>] [\fB\-\-strict\fR]
Make the cluster forget the complete operation history (including failures) of the resource / stonith device and re\-detect its current state. If you are interested in forgetting failed operations only, use the 'pcs resource cleanup' command.
If the named resource is part of a group, or one numbered instance of a clone or bundled resource, the refresh applies to the whole collective resource unless \fB\-\-strict\fR is given.
If a resource id / stonith id is not specified then all resources / stonith devices will be refreshed.
If a node is not specified then resources / stonith devices on all nodes will be refreshed.
.TP
failcount [show [<resource id | stonith id>] [node=<node>] [operation=<operation> [interval=<interval>]]] [\fB\-\-full\fR]
Show current failcount for resources and stonith devices, optionally filtered by a resource / stonith device, node, operation and its interval. If \fB\-\-full\fR is specified do not sum failcounts per resource / stonith device and node. Use 'pcs resource cleanup' or 'pcs resource refresh' to reset failcounts.
.TP
relocate dry\-run [resource1] [resource2] ...
The same as 'relocate run' but has no effect on the cluster.
.TP
relocate run [resource1] [resource2] ...
Relocate specified resources to their preferred nodes. If no resources are specified, relocate all resources. This command calculates the preferred node for each resource while ignoring resource stickiness. Then it creates location constraints which will cause the resources to move to their preferred nodes. Once the resources have been moved the constraints are deleted automatically. Note that the preferred node is calculated based on current cluster status, constraints, location of resources and other settings and thus it might change over time.
.TP
relocate show
Display current status of resources and their optimal node ignoring resource stickiness.
.TP
relocate clear
Remove all constraints created by the 'relocate run' command.
.TP
utilization [<resource id> [<name>=<value> ...]]
Add specified utilization options to specified resource. If resource is not specified, shows utilization of all resources. If utilization options are not specified, shows utilization of specified resource. Utilization option should be in format name=value, value has to be integer. Options may be removed by setting an option without a value. Example: pcs resource utilization TestResource cpu= ram=20
.TP
relations <resource id> [\fB\-\-full\fR]
Display relations of a resource specified by its id with other resources in a tree structure. Supported types of resource relations are: ordering constraints, ordering set constraints, relations defined by resource hierarchy (clones, groups, bundles). If \fB\-\-full\fR is used, more verbose output will be printed.
.SS "cluster"
.TP
setup <cluster name> (<node name> [addr=<node address>]...)... [transport knet|udp|udpu [<transport options>] [link <link options>]... [compression <compression options>] [crypto <crypto options>]] [totem <totem options>] [quorum <quorum options>] [\fB\-\-no\-cluster\-uuid\fR] ([\fB\-\-enable\fR] [\fB\-\-start\fR [\fB\-\-wait\fR[=<n>]]] [\fB\-\-no\-keys\-sync\fR]) | [\fB\-\-corosync_conf\fR <path>]
Create a cluster from the listed nodes and synchronize cluster configuration files to them. If \fB\-\-corosync_conf\fR is specified, do not connect to other nodes and save corosync.conf to the specified path; see 'Local only mode' below for details.
Nodes are specified by their names and optionally their addresses. If no addresses are specified for a node, pcs will configure corosync to communicate with that node using an address provided in 'pcs host auth' command. Otherwise, pcs will configure corosync to communicate with the node using the specified addresses.
Transport knet:
.br
This is the default transport. It allows configuring traffic encryption and compression as well as using multiple addresses (links) for nodes.
.br
Transport options are: ip_version, knet_pmtud_interval, link_mode
.br
Link options are: link_priority, linknumber, mcastport, ping_interval, ping_precision, ping_timeout, pong_count, transport (udp or sctp)
.br
Each 'link' followed by options sets options for one link in the order the links are defined by nodes' addresses. You can set link options for a subset of links using a linknumber. See examples below.
.br
Compression options are: level, model, threshold
.br
Crypto options are: cipher, hash, model
.br
By default, encryption is enabled with cipher=aes256 and hash=sha256. To disable encryption, set cipher=none and hash=none.
Transports udp and udpu:
.br
These transports are limited to one address per node. They do not support traffic encryption nor compression.
.br
Transport options are: ip_version, netmtu
.br
Link options are: bindnetaddr, broadcast, mcastaddr, mcastport, ttl
Totem and quorum can be configured regardless of used transport.
.br
Totem options are: block_unlisted_ips, consensus, downcheck, fail_recv_const, heartbeat_failures_allowed, hold, join, max_messages, max_network_delay, merge, miss_count_const, send_join, seqno_unchanged_const, token, token_coefficient, token_retransmit, token_retransmits_before_loss_const, window_size
.br
Quorum options are: auto_tie_breaker, last_man_standing, last_man_standing_window, wait_for_all
Transports and their options, link, compression, crypto and totem options are all documented in corosync.conf(5) man page; knet link options are prefixed 'knet_' there, compression options are prefixed 'knet_compression_' and crypto options are prefixed 'crypto_'. Quorum options are documented in votequorum(5) man page.
\fB\-\-no\-cluster\-uuid\fR will not generate a unique ID for the cluster. \fB\-\-enable\fR will configure the cluster to start on nodes boot. \fB\-\-start\fR will start the cluster right after creating it. \fB\-\-wait\fR will wait up to 'n' seconds for the cluster to start. \fB\-\-no\-keys\-sync\fR will skip creating and distributing pcsd SSL certificate and key and corosync and pacemaker authkey files. Use this if you provide your own certificates and keys.
Local only mode:
.br
By default, pcs connects to all specified nodes to verify they can be used in the new cluster and to send cluster configuration files to them. If this is not what you want, specify \fB\-\-corosync_conf\fR option followed by a file path. Pcs will save corosync.conf to the specified file and will not connect to cluster nodes. These are the tasks that pcs skips in that case:
.br
* make sure the nodes are not running or configured to run a cluster already
.br
* make sure cluster packages are installed on all nodes and their versions are compatible
.br
* make sure there are no cluster configuration files on any node (run 'pcs cluster destroy' and remove pcs_settings.conf file on all nodes)
.br
* synchronize corosync and pacemaker authkeys, /etc/corosync/authkey and /etc/pacemaker/authkey respectively, and the corosync.conf file
.br
* authenticate the cluster nodes against each other ('pcs cluster auth' or 'pcs host auth' command)
.br
* synchronize pcsd certificates (so that pcs web UI can be used in an HA mode)
Examples:
.br
Create a cluster with default settings:
pcs cluster setup newcluster node1 node2
.br
Create a cluster using two links:
pcs cluster setup newcluster node1 addr=10.0.1.11 addr=10.0.2.11 node2 addr=10.0.1.12 addr=10.0.2.12
.br
Set link options for all links. Link options are matched to the links in order. The first link (link 0) has sctp transport, the second link (link 1) has mcastport 55405:
pcs cluster setup newcluster node1 addr=10.0.1.11 addr=10.0.2.11 node2 addr=10.0.1.12 addr=10.0.2.12 transport knet link transport=sctp link mcastport=55405
.br
Set link options for the second and fourth links only. Link options are matched to the links based on the linknumber option (the first link is link 0):
pcs cluster setup newcluster node1 addr=10.0.1.11 addr=10.0.2.11 addr=10.0.3.11 addr=10.0.4.11 node2 addr=10.0.1.12 addr=10.0.2.12 addr=10.0.3.12 addr=10.0.4.12 transport knet link linknumber=3 mcastport=55405 link linknumber=1 transport=sctp
.br
Create a cluster using udp transport with a non-default port:
pcs cluster setup newcluster node1 node2 transport udp link mcastport=55405
.TP
config [show] [\fB\-\-output\-format\fR <cmd|json|text>] [\fB\-\-corosync_conf\fR <path>]
Show cluster configuration. There are 3 formats of output available: 'cmd', 'json' and 'text', default is 'text'. Format 'text' is a human friendly output. Format 'cmd' prints a cluster setup command which recreates a cluster with the same configuration. Format 'json' is a machine oriented output with cluster configuration. If \fB\-\-corosync_conf\fR is specified, configuration file specified by <path> is used instead of the current cluster configuration.
.TP
config update [transport <transport options>] [compression <compression options>] [crypto <crypto options>] [totem <totem options>] [\fB\-\-corosync_conf\fR <path>]
Update cluster configuration. Unspecified options will be kept unchanged. If you wish to remove an option, set it to empty value, i.e. 'option_name='.
If \fB\-\-corosync_conf\fR is specified, update cluster configuration in a file specified by <path>.
All options are documented in corosync.conf(5) man page. There are different transport options for transport types. Compression and crypto options are only available for knet transport. Totem options can be set regardless of the transport type.
.br
Transport options for knet transport are: ip_version, knet_pmtud_interval, link_mode
.br
Transport options for udp and updu transports are: ip_version, netmtu
.br
Compression options are: level, model, threshold
.br
Crypto options are: cipher, hash, model
.br
Totem options are: block_unlisted_ips, consensus, downcheck, fail_recv_const, heartbeat_failures_allowed, hold, join, max_messages, max_network_delay, merge, miss_count_const, send_join, seqno_unchanged_const, token, token_coefficient, token_retransmit, token_retransmits_before_loss_const, window_size
.TP
config uuid generate [\fB\-\-corosync_conf\fR <path>] [\fB\-\-force\fR]
Generate a new cluster UUID and distribute it to all cluster nodes. Cluster UUID is not used by the cluster stack in any way, it is provided to easily distinguish between multiple clusters in a multi-cluster environment since the cluster name does not have to be unique.
If \fB\-\-corosync_conf\fR is specified, update cluster configuration in file specified by <path>.
If \fB\-\-force\fR is specified, existing UUID will be overwritten.
.TP
authkey corosync [<path>]
Generate a new corosync authkey and distribute it to all cluster nodes. If <path> is specified, do not generate a key and use key from the file.
.TP
start [\fB\-\-all\fR | <node>... ] [\fB\-\-wait\fR[=<n>]] [\fB\-\-request\-timeout\fR=<seconds>]
Start a cluster on specified node(s). If no nodes are specified then start a cluster on the local node. If \fB\-\-all\fR is specified then start a cluster on all nodes. If the cluster has many nodes then the start request may time out. In that case you should consider setting \fB\-\-request\-timeout\fR to a suitable value. If \fB\-\-wait\fR is specified, pcs waits up to 'n' seconds for the cluster to get ready to provide services after the cluster has successfully started.
.TP
stop [\fB\-\-all\fR | <node>... ] [\fB\-\-request\-timeout\fR=<seconds>]
Stop a cluster on specified node(s). If no nodes are specified then stop a cluster on the local node. If \fB\-\-all\fR is specified then stop a cluster on all nodes. If the cluster is running resources which take long time to stop then the stop request may time out before the cluster actually stops. In that case you should consider setting \fB\-\-request\-timeout\fR to a suitable value.
.TP
kill
Force corosync and pacemaker daemons to stop on the local node (performs kill \-9). Note that init system (e.g. systemd) can detect that cluster is not running and start it again. If you want to stop cluster on a node, run pcs cluster stop on that node.
.TP
enable [\fB\-\-all\fR | <node>... ]
Configure cluster to run on node boot on specified node(s). If node is not specified then cluster is enabled on the local node. If \fB\-\-all\fR is specified then cluster is enabled on all nodes.
.TP
disable [\fB\-\-all\fR | <node>... ]
Configure cluster to not run on node boot on specified node(s). If node is not specified then cluster is disabled on the local node. If \fB\-\-all\fR is specified then cluster is disabled on all nodes.
.TP
auth [\fB\-u\fR <username>] [\fB\-p\fR <password>]
Authenticate pcs/pcsd to pcsd on nodes configured in the local cluster.
.TP
status
View current cluster status (an alias of 'pcs status cluster').
.TP
sync
Sync cluster configuration (files which are supported by all subcommands of this command) to all cluster nodes.
.TP
sync corosync
Sync corosync configuration to all nodes found from current corosync.conf file.
.TP
cib [filename] [scope=<scope> | \fB\-\-config\fR]
Get the raw xml from the CIB (Cluster Information Base). If a filename is provided, we save the CIB to that file, otherwise the CIB is printed. Specify scope to get a specific section of the CIB. Valid values of the scope are: acls, alerts, configuration, constraints, crm_config, fencing-topology, nodes, op_defaults, resources, rsc_defaults, tags. \fB\-\-config\fR is the same as scope=configuration. Do not specify a scope if you want to edit the saved CIB using pcs (pcs \fB\-f\fR <command>).
.TP
cib\-push <filename> [\fB\-\-wait\fR[=<n>]] [diff\-against=<filename_original> | scope=<scope> | \fB\-\-config\fR]
Push the raw xml from <filename> to the CIB (Cluster Information Base). You can obtain the CIB by running the 'pcs cluster cib' command, which is recommended first step when you want to perform desired modifications (pcs \fB\-f\fR <command>) for the one\-off push.
.br
If diff\-against is specified, pcs diffs contents of filename against contents of filename_original and pushes the result to the CIB.
.br
Specify scope to push a specific section of the CIB. Valid values of the scope are: acls, alerts, configuration, constraints, crm_config, fencing-topology, nodes, op_defaults, resources, rsc_defaults, tags. \fB\-\-config\fR is the same as scope=configuration. Use of \fB\-\-config\fR is recommended. Do not specify a scope if you need to push the whole CIB or be warned in the case of outdated CIB.
.br
If \fB\-\-wait\fR is specified wait up to 'n' seconds for changes to be applied.
.br
\fBWARNING:\fR the selected scope of the CIB will be overwritten by the current content of the specified file.
Example:
pcs cluster cib > original.xml
cp original.xml new.xml
pcs \-f new.xml constraint location apache prefers node2
pcs cluster cib\-push new.xml diff\-against=original.xml
.TP
cib\-upgrade
Upgrade the CIB to conform to the latest version of the document schema.
.TP
edit [scope=<scope> | \fB\-\-config\fR]
Edit the cib in the editor specified by the $EDITOR environment variable and push out any changes upon saving. Specify scope to edit a specific section of the CIB. Valid values of the scope are: acls, alerts, configuration, constraints, crm_config, fencing-topology, nodes, op_defaults, resources, rsc_defaults, tags. \fB\-\-config\fR is the same as scope=configuration. Use of \fB\-\-config\fR is recommended. Do not specify a scope if you need to edit the whole CIB or be warned in the case of outdated CIB.
.TP
node add <node name> [addr=<node address>]... [watchdog=<watchdog path>] [device=<SBD device path>]... [\fB\-\-start\fR [\fB\-\-wait\fR[=<n>]]] [\fB\-\-enable\fB] [\fB\-\-no\-watchdog\-validation\fR]
Add the node to the cluster and synchronize all relevant configuration files to the new node. This command can only be run on an existing cluster node.
The new node is specified by its name and optionally its addresses. If no addresses are specified for the node, pcs will configure corosync to communicate with the node using an address provided in 'pcs host auth' command. Otherwise, pcs will configure corosync to communicate with the node using the specified addresses.
Use 'watchdog' to specify a path to a watchdog on the new node, when SBD is enabled in the cluster. If SBD is configured with shared storage, use 'device' to specify path to shared device(s) on the new node.
If \fB\-\-start\fR is specified also start cluster on the new node, if \fB\-\-wait\fR is specified wait up to 'n' seconds for the new node to start. If \fB\-\-enable\fR is specified configure cluster to start on the new node on boot. If \fB\-\-no\-watchdog\-validation\fR is specified, validation of watchdog will be skipped.
.B WARNING: By default, it is tested whether the specified watchdog is supported. This may cause a restart of the system when a watchdog with no\-way\-out\-feature enabled is present. Use \-\-no\-watchdog\-validation to skip watchdog validation.
.TP
node delete <node name> [<node name>]...
Shutdown specified nodes and remove them from the cluster.
.TP
node remove <node name> [<node name>]...
Shutdown specified nodes and remove them from the cluster.
.TP
node add\-remote <node name> [<node address>] [options] [op <operation action> <operation options> [<operation action> <operation options>]...] [meta <meta options>...] [\fB\-\-wait\fR[=<n>]]
Add the node to the cluster as a remote node. Sync all relevant configuration files to the new node. Start the node and configure it to start the cluster on boot. Options are port and reconnect_interval. Operations and meta belong to an underlying connection resource (ocf:pacemaker:remote). If node address is not specified for the node, pcs will configure pacemaker to communicate with the node using an address provided in 'pcs host auth' command. Otherwise, pcs will configure pacemaker to communicate with the node using the specified addresses. If \fB\-\-wait\fR is specified, wait up to 'n' seconds for the node to start.
.TP
node delete\-remote <node identifier>
Shutdown specified remote node and remove it from the cluster. The node\-identifier can be the name of the node or the address of the node.
.TP
node remove\-remote <node identifier>
Shutdown specified remote node and remove it from the cluster. The node\-identifier can be the name of the node or the address of the node.
.TP
node add\-guest <node name> <resource id> [options] [\fB\-\-wait\fR[=<n>]]
Make the specified resource a guest node resource. Sync all relevant configuration files to the new node. Start the node and configure it to start the cluster on boot. Options are remote\-addr, remote\-port and remote\-connect\-timeout. If remote\-addr is not specified for the node, pcs will configure pacemaker to communicate with the node using an address provided in 'pcs host auth' command. Otherwise, pcs will configure pacemaker to communicate with the node using the specified addresses. If \fB\-\-wait\fR is specified, wait up to 'n' seconds for the node to start.
.TP
node delete\-guest <node identifier>
Shutdown specified guest node and remove it from the cluster. The node\-identifier can be the name of the node or the address of the node or id of the resource that is used as the guest node.
.TP
node remove\-guest <node identifier>
Shutdown specified guest node and remove it from the cluster. The node\-identifier can be the name of the node or the address of the node or id of the resource that is used as the guest node.
.TP
node clear <node name>
Remove specified node from various cluster caches. Use this if a removed node is still considered by the cluster to be a member of the cluster.
.TP
link add <node_name>=<node_address>... [options <link options>]
Add a corosync link. One address must be specified for each cluster node. If no linknumber is specified, pcs will use the lowest available linknumber.
.br
Link options (documented in corosync.conf(5) man page) are: link_priority, linknumber, mcastport, ping_interval, ping_precision, ping_timeout, pong_count, transport (udp or sctp)
.TP
link delete <linknumber> [<linknumber>]...
Remove specified corosync links.
.TP
link remove <linknumber> [<linknumber>]...
Remove specified corosync links.
.TP
link update <linknumber> [<node_name>=<node_address>...] [options <link options>]
Add, remove or change node addresses / link options of an existing corosync link. Use this if you cannot add / remove links which is the preferred way. Unspecified options will be kept unchanged. If you wish to remove an option, set it to empty value, i.e. 'option_name='.
.br
Link options (documented in corosync.conf(5) man page) are:
.br
for knet transport: link_priority, mcastport, ping_interval, ping_precision, ping_timeout, pong_count, transport (udp or sctp)
.br
for udp and udpu transports: bindnetaddr, broadcast, mcastaddr, mcastport, ttl
.TP
uidgid
List the current configured uids and gids of users allowed to connect to corosync.
.TP
uidgid add [uid=<uid>] [gid=<gid>]
Add the specified uid and/or gid to the list of users/groups allowed to connect to corosync.
.TP
uidgid delete [uid=<uid>] [gid=<gid>]
Remove the specified uid and/or gid from the list of users/groups allowed to connect to corosync.
.TP
uidgid remove [uid=<uid>] [gid=<gid>]
Remove the specified uid and/or gid from the list of users/groups allowed to connect to corosync.
.TP
corosync [node]
Get the corosync.conf from the specified node or from the current node if node not specified.
.TP
reload corosync
Reload the corosync configuration on the current node.
.TP
destroy [\fB\-\-all\fR] [\fB\-\-force\fR]
Permanently destroy the cluster on the current node, killing all cluster processes and removing all cluster configuration files. Using \fB\-\-all\fR will attempt to destroy the cluster on all nodes in the local cluster.
\fBWARNING:\fR This command permanently removes any cluster configuration that has been created. It is recommended to run 'pcs cluster stop' before destroying the cluster. To prevent accidental running of this command, \fB\-\-force\fR or interactive user response is required in order to proceed.
.TP
verify [\fB\-\-full\fR] [\fB\-f\fR <filename>]
Checks the pacemaker configuration (CIB) for syntax and common conceptual errors. If no filename is specified the check is performed on the currently running cluster. If \fB\-\-full\fR is used more verbose output will be printed.
.TP
report [\fB\-\-from\fR "YYYY\-M\-D H:M:S" [\fB\-\-to\fR "YYYY\-M\-D H:M:S"]] <dest>
Create a tarball containing everything needed when reporting cluster problems. If \fB\-\-from\fR and \fB\-\-to\fR are not used, the report will include the past 24 hours.
.SS "stonith"
.TP
[status [<resource id | tag id>] [node=<node>] [\fB\-\-hide\-inactive\fR]]
Show status of all currently configured stonith devices. If \fB\-\-hide\-inactive\fR is specified, only show active stonith devices. If a resource or tag id is specified, only show status of the specified resource or resources in the specified tag. If node is specified, only show status of resources configured for the specified node.
.TP
config [<stonith id>]...
Show options of all currently configured stonith devices or if stonith ids are specified show the options for the specified stonith device ids.
.TP
list [filter] [\fB\-\-nodesc\fR]
Show list of all available stonith agents (if filter is provided then only stonith agents matching the filter will be shown). If \fB\-\-nodesc\fR is used then descriptions of stonith agents are not printed.
.TP
describe <stonith agent> [\fB\-\-full\fR]
Show options for specified stonith agent. If \fB\-\-full\fR is specified, all options including advanced and deprecated ones are shown.
.TP
create <stonith id> <stonith device type> [stonith device options] [op <operation action> <operation options> [<operation action> <operation options>]...] [meta <meta options>...] [\fB\-\-group\fR <group id> [\fB\-\-before\fR <stonith id> | \fB\-\-after\fR <stonith id>]] [\fB\-\-disabled\fR] [\fB\-\-wait\fR[=n]]
Create stonith device with specified type and options. If \fB\-\-group\fR is specified the stonith device is added to the group named. You can use \fB\-\-before\fR or \fB\-\-after\fR to specify the position of the added stonith device relatively to some stonith device already existing in the group. If\fB\-\-disabled\fR is specified the stonith device is not used. If \fB\-\-wait\fR is specified, pcs will wait up to 'n' seconds for the stonith device to start and then return 0 if the stonith device is started, or 1 if the stonith device has not yet started. If 'n' is not specified it defaults to 60 minutes.
Example: Create a device for nodes node1 and node2
.br
pcs stonith create MyFence fence_virt pcmk_host_list=node1,node2
.br
Example: Use port p1 for node n1 and ports p2 and p3 for node n2
.br
pcs stonith create MyFence fence_virt 'pcmk_host_map=n1:p1;n2:p2,p3'
.TP
update <stonith id> [stonith options] [op [<operation action> <operation options>]...] [meta <meta operations>...] [\fB\-\-wait\fR[=n]]
Add, remove or change options of specified stonith device. Unspecified options will be kept unchanged. If you wish to remove an option, set it to empty value, i.e. 'option_name='.
If an operation (op) is specified it will update the first found operation with the same action on the specified stonith device. If no operation with that action exists then a new operation will be created. (WARNING: all existing options on the updated operation will be reset if not specified.) If you want to create multiple monitor operations you should use the 'op add' & 'op remove' commands.
If \fB\-\-wait\fR is specified, pcs will wait up to 'n' seconds for the changes to take effect and then return 0 if the changes have been processed or 1 otherwise. If 'n' is not specified it defaults to 60 minutes.
.TP
update\-scsi\-devices <stonith id> (set <device\-path> [<device\-path>...]) | (add <device\-path> [<device\-path>...] delete|remove <device\-path> [<device\-path>...] )
Update scsi fencing devices without affecting other resources. You must specify either list of set devices or at least one device for add or delete/remove devices. Stonith resource must be running on one cluster node. Each device will be unfenced on each cluster node running cluster. Supported fence agents: fence_scsi, fence_mpath.
.TP
delete <stonith id>
Remove stonith id from configuration.
.TP
remove <stonith id>
Remove stonith id from configuration.
.TP
op add <stonith id> <operation action> [operation properties]
Add operation for specified stonith device.
.TP
op delete <stonith id> <operation action> [<operation properties>...]
Remove specified operation (note: you must specify the exact operation properties to properly remove an existing operation).
.TP
op delete <operation id>
Remove the specified operation id.
.TP
op remove <stonith id> <operation action> [<operation properties>...]
Remove specified operation (note: you must specify the exact operation properties to properly remove an existing operation).
.TP
op remove <operation id>
Remove the specified operation id.
.TP
op defaults [config] [\fB\-\-all\fR] [\fB\-\-full\fR] [\fB\-\-no\-check\-expired\fR]
This command is an alias of 'resource op defaults [config]' command.
List currently configured default values for operations. If \fB\-\-all\fR is specified, also list expired sets of values. If \fB\-\-full\fR is specified, also list ids. If \fB\-\-no\-expire\-check\fR is specified, do not evaluate whether sets of values are expired.
.TP
op defaults <name>=<value>...
This command is an alias of 'resource op defaults' command.
Set default values for operations.
.br
NOTE: Defaults do not apply to resources / stonith devices which override them with their own defined values.
.TP
op defaults set create [<set options>] [meta [<name>=<value>]...] [rule [<expression>]]
This command is an alias of 'resource op defaults set create' command.
Create a new set of default values for resource / stonith device operations. You may specify a rule describing resources / stonith devices and / or operations to which the set applies.
Set options are: id, score
Expression looks like one of the following:
.br
op <operation name> [interval=<interval>]
.br
resource [<standard>]:[<provider>]:[<type>]
.br
defined|not_defined <node attribute>
.br
<node attribute> lt|gt|lte|gte|eq|ne [string|integer|number|version] <value>
.br
date gt|lt <date>
.br
date in_range [<date>] to <date>
.br
date in_range <date> to duration <duration options>
.br
date\-spec <date\-spec options>
.br
<expression> and|or <expression>
.br
(<expression>)
You may specify all or any of 'standard', 'provider' and 'type' in a resource expression. For example: 'resource ocf::' matches all resources of 'ocf' standard, while 'resource ::Dummy' matches all resources of 'Dummy' type regardless of their standard and provider.
Dates are expected to conform to ISO 8601 format.
Duration options are: hours, monthdays, weekdays, yearsdays, months, weeks, years, weekyears, moon. Value for these options is an integer.
Date\-spec options are: hours, monthdays, weekdays, yearsdays, months, weeks, years, weekyears, moon. Value for these options is an integer or a range written as integer\-integer.
NOTE: Defaults do not apply to resources / stonith devices which override them with their own defined values.
.TP
op defaults set delete [<set id>]...
This command is an alias of 'resource op defaults set delete' command.
Delete specified options sets.
.TP
op defaults set remove [<set id>]...
This command is an alias of 'resource op defaults set delete' command.
Delete specified options sets.
.TP
op defaults set update <set id> [meta [<name>=<value>]...]
This command is an alias of 'resource op defaults set update' command.
Add, remove or change values in specified set of default values for resource / stonith device operations. Unspecified options will be kept unchanged. If you wish to remove an option, set it to empty value, i.e. 'option_name='.
NOTE: Defaults do not apply to resources / stonith devices which override them with their own defined values.
.TP
op defaults update <name>=<value>...
This command is an alias of 'resource op defaults update' command.
Add, remove or change default values for operations. This is a simplified command useful for cases when you only manage one set of default values. Unspecified options will be kept unchanged. If you wish to remove an option, set it to empty value, i.e. 'option_name='.
NOTE: Defaults do not apply to resources / stonith devices which override them with their own defined values.
.TP
meta <stonith id> <meta options> [\fB\-\-wait\fR[=n]]
Add specified options to the specified stonith device. Meta options should be in the format of name=value, options may be removed by setting an option without a value. If \fB\-\-wait\fR is specified, pcs will wait up to 'n' seconds for the changes to take effect and then return 0 if the changes have been processed or 1 otherwise. If 'n' is not specified it defaults to 60 minutes.
Example: pcs stonith meta test_stonith failure\-timeout=50 resource\-stickiness=
.TP
defaults [config] [\fB\-\-all\fR] [\fB\-\-full\fR] [\fB\-\-no\-check\-expired\fR]
This command is an alias of 'resource defaults [config]' command.
List currently configured default values for resources / stonith devices. If \fB\-\-all\fR is specified, also list expired sets of values. If \fB\-\-full\fR is specified, also list ids. If \fB\-\-no\-expire\-check\fR is specified, do not evaluate whether sets of values are expired.
.TP
defaults <name>=<value>...
This command is an alias of 'resource defaults' command.
Set default values for resources / stonith devices.
.br
NOTE: Defaults do not apply to resources / stonith devices which override them with their own defined values.
.TP
defaults set create [<set options>] [meta [<name>=<value>]...] [rule [<expression>]]
This command is an alias of 'resource defaults set create' command.
Create a new set of default values for resources / stonith devices. You may specify a rule describing resources / stonith devices to which the set applies.
Set options are: id, score
Expression looks like one of the following:
.br
resource [<standard>]:[<provider>]:[<type>]
.br
date gt|lt <date>
.br
date in_range [<date>] to <date>
.br
date in_range <date> to duration <duration options>
.br
date\-spec <date\-spec options>
.br
<expression> and|or <expression>
.br
(<expression>)
You may specify all or any of 'standard', 'provider' and 'type' in a resource expression. For example: 'resource ocf::' matches all resources of 'ocf' standard, while 'resource ::Dummy' matches all resources of 'Dummy' type regardless of their standard and provider.
Dates are expected to conform to ISO 8601 format.
Duration options are: hours, monthdays, weekdays, yearsdays, months, weeks, years, weekyears, moon. Value for these options is an integer.
Date\-spec options are: hours, monthdays, weekdays, yearsdays, months, weeks, years, weekyears, moon. Value for these options is an integer or a range written as integer\-integer.
NOTE: Defaults do not apply to resources / stonith devices which override them with their own defined values.
.TP
defaults set delete [<set id>]...
This command is an alias of 'resource defaults set delete' command.
Delete specified options sets.
.TP
defaults set remove [<set id>]...
This command is an alias of 'resource defaults set delete' command.
Delete specified options sets.
.TP
defaults set update <set id> [meta [<name>=<value>]...]
This command is an alias of 'resource defaults set update' command.
Add, remove or change values in specified set of default values for resources / stonith devices. Unspecified options will be kept unchanged. If you wish to remove an option, set it to empty value, i.e. 'option_name='.
NOTE: Defaults do not apply to resources / stonith devices which override them with their own defined values.
.TP
defaults update <name>=<value>...
This command is an alias of 'resource defaults update' command.
Add, remove or change default values for resources / stonith devices. This is a simplified command useful for cases when you only manage one set of default values. Unspecified options will be kept unchanged. If you wish to remove an option, set it to empty value, i.e. 'option_name='.
NOTE: Defaults do not apply to resources / stonith devices which override them with their own defined values.
.TP
cleanup [<resource id | stonith id>] [node=<node>] [operation=<operation> [interval=<interval>]] [\fB\-\-strict\fR]
This command is an alias of 'resource cleanup' command.
Make the cluster forget failed operations from history of the resource / stonith device and re\-detect its current state. This can be useful to purge knowledge of past failures that have since been resolved.
If the named resource is part of a group, or one numbered instance of a clone or bundled resource, the clean\-up applies to the whole collective resource unless \fB\-\-strict\fR is given.
If a resource id / stonith id is not specified then all resources / stonith devices will be cleaned up.
If a node is not specified then resources / stonith devices on all nodes will be cleaned up.
.TP
refresh [<resource id | stonith id>] [node=<node>] [\fB\-\-strict\fR]
This command is an alias of 'resource refresh' command.
Make the cluster forget the complete operation history (including failures) of the resource / stonith device and re\-detect its current state. If you are interested in forgetting failed operations only, use the 'pcs resource cleanup' command.
If the named resource is part of a group, or one numbered instance of a clone or bundled resource, the refresh applies to the whole collective resource unless \fB\-\-strict\fR is given.
If a resource id / stonith id is not specified then all resources / stonith devices will be refreshed.
If a node is not specified then resources / stonith devices on all nodes will be refreshed.
.TP
failcount [show [<resource id | stonith id>] [node=<node>] [operation=<operation> [interval=<interval>]]] [\fB\-\-full\fR]
This command is an alias of 'resource failcount show' command.
Show current failcount for resources and stonith devices, optionally filtered by a resource / stonith device, node, operation and its interval. If \fB\-\-full\fR is specified do not sum failcounts per resource / stonith device and node. Use 'pcs resource cleanup' or 'pcs resource refresh' to reset failcounts.
.TP
enable <stonith id>... [\fB\-\-wait[=n]\fR]
Allow the cluster to use the stonith devices. If \fB\-\-wait\fR is specified, pcs will wait up to 'n' seconds for the stonith devices to start and then return 0 if the stonith devices are started, or 1 if the stonith devices have not yet started. If 'n' is not specified it defaults to 60 minutes.
.TP
disable <stonith id>... [\fB\-\-wait[=n]\fR]
Attempt to stop the stonith devices if they are running and disallow the cluster to use them. If \fB\-\-wait\fR is specified, pcs will wait up to 'n' seconds for the stonith devices to stop and then return 0 if the stonith devices are stopped or 1 if the stonith devices have not stopped. If 'n' is not specified it defaults to 60 minutes.
.TP
level [config]
Lists all of the fencing levels currently configured.
.TP
level add <level> <target> <stonith id> [stonith id]...
Add the fencing level for the specified target with the list of stonith devices to attempt for that target at that level. Fence levels are attempted in numerical order (starting with 1). If a level succeeds (meaning all devices are successfully fenced in that level) then no other levels are tried, and the target is considered fenced. Target may be a node name <node_name> or %<node_name> or node%<node_name>, a node name regular expression regexp%<node_pattern> or a node attribute value attrib%<name>=<value>.
.TP
level delete <level> [target <target>] [stonith <stonith id>...]
Removes the fence level for the level, target and/or devices specified. If no target or devices are specified then the fence level is removed. Target may be a node name <node_name> or %<node_name> or node%<node_name>, a node name regular expression regexp%<node_pattern> or a node attribute value attrib%<name>=<value>.
.TP
level remove <level> [target <target>] [stonith <stonith id>...]
Removes the fence level for the level, target and/or devices specified. If no target or devices are specified then the fence level is removed. Target may be a node name <node_name> or %<node_name> or node%<node_name>, a node name regular expression regexp%<node_pattern> or a node attribute value attrib%<name>=<value>.
.TP
level clear [target <target> | stonith <stonith id>...]
Clears the fence levels on the target (or stonith id) specified or clears all fence levels if a target/stonith id is not specified. Target may be a node name <node_name> or %<node_name> or node%<node_name>, a node name regular expression regexp%<node_pattern> or a node attribute value attrib%<name>=<value>. Example: pcs stonith level clear stonith dev_a dev_b
.TP
level verify
Verifies all fence devices and nodes specified in fence levels exist.
.TP
fence <node> [\fB\-\-off\fR]
Fence the node specified (if \fB\-\-off\fR is specified, use the 'off' API call to stonith which will turn the node off instead of rebooting it).
.TP
confirm <node> [\fB\-\-force\fR]
Confirm to the cluster that the specified node is powered off. This allows the cluster to recover from a situation where no stonith device is able to fence the node. This command should \fBONLY\fR be used after manually ensuring that the node is powered off and has no access to shared resources.
.B WARNING: If this node is not actually powered off or it does have access to shared resources, data corruption/cluster failure can occur. To prevent accidental running of this command, \-\-force or interactive user response is required in order to proceed.
NOTE: It is not checked if the specified node exists in the cluster in order to be able to work with nodes not visible from the local cluster partition.
.TP
history [show [<node>]]
Show fencing history for the specified node or all nodes if no node specified.
.TP
history cleanup [<node>]
Cleanup fence history of the specified node or all nodes if no node specified.
.TP
history update
Update fence history from all nodes.
.TP
sbd enable [watchdog=<path>[@<node>]]... [device=<path>[@<node>]]... [<SBD_OPTION>=<value>]... [\fB\-\-no\-watchdog\-validation\fR]
Enable SBD in cluster. Default path for watchdog device is /dev/watchdog. Allowed SBD options: SBD_WATCHDOG_TIMEOUT (default: 5), SBD_DELAY_START (default: no), SBD_STARTMODE (default: always) and SBD_TIMEOUT_ACTION. SBD options are documented in sbd(8) man page. It is possible to specify up to 3 devices per node. If \fB\-\-no\-watchdog\-validation\fR is specified, validation of watchdogs will be skipped.
.B WARNING: Cluster has to be restarted in order to apply these changes.
.B WARNING: By default, it is tested whether the specified watchdog is supported. This may cause a restart of the system when a watchdog with no\-way\-out\-feature enabled is present. Use \-\-no\-watchdog\-validation to skip watchdog validation.
Example of enabling SBD in cluster with watchdogs on node1 will be /dev/watchdog2, on node2 /dev/watchdog1, /dev/watchdog0 on all other nodes, device /dev/sdb on node1, device /dev/sda on all other nodes and watchdog timeout will bet set to 10 seconds:
pcs stonith sbd enable watchdog=/dev/watchdog2@node1 watchdog=/dev/watchdog1@node2 watchdog=/dev/watchdog0 device=/dev/sdb@node1 device=/dev/sda SBD_WATCHDOG_TIMEOUT=10
.TP
sbd disable
Disable SBD in cluster.
.B WARNING: Cluster has to be restarted in order to apply these changes.
.TP
sbd device setup device=<path> [device=<path>]... [watchdog\-timeout=<integer>] [allocate\-timeout=<integer>] [loop\-timeout=<integer>] [msgwait\-timeout=<integer>]
Initialize SBD structures on device(s) with specified timeouts.
.B WARNING: All content on device(s) will be overwritten.
.TP
sbd device message <device\-path> <node> <message\-type>
Manually set a message of the specified type on the device for the node. Possible message types (they are documented in sbd(8) man page): test, reset, off, crashdump, exit, clear
.TP
sbd status [\fB\-\-full\fR]
Show status of SBD services in cluster and local device(s) configured. If \fB\-\-full\fR is specified, also dump of SBD headers on device(s) will be shown.
.TP
sbd config
Show SBD configuration in cluster.
.TP
sbd watchdog list
Show all available watchdog devices on the local node.
.B WARNING: Listing available watchdogs may cause a restart of the system when a watchdog with no\-way\-out\-feature enabled is present.
.TP
sbd watchdog test [<watchdog\-path>]
This operation is expected to force\-reboot the local system without following any shutdown procedures using a watchdog. If no watchdog is specified, available watchdog will be used if only one watchdog device is available on the local system.
.SS "acl"
.TP
[config]
List all current access control lists.
.TP
enable
Enable access control lists.