Cannot seem to start pcs cluster (NFS Cluster) disk_fencing trouble
1
vote
0
answers
603
views
For the life of me, I can't find a clear answer on how to start my NFS active / passive cluster. I have two nodes, node1 and node2 and followed the guide here: https://www.linuxtechi.com/configure-nfs-server-clustering-pacemaker-centos-7-rhel-7/
Here are my logs:
May 25 10:35:59 node1 stonith-ng: notice: Couldn't find anyone to fence (on) node1 with any device
May 25 10:35:59 node1 stonith-ng: error: Operation on of node1 by for crmd.3928@node1.97f683f8: No route to host
May 25 10:35:59 node1 crmd: notice: Stonith operation 142/2:72:0:f3e078bf-24f5-4160-95c1-0eeeea0e5e12: No route to host (-113)
May 25 10:35:59 node1 crmd: notice: Stonith operation 142 for node1 failed (No route to host): aborting transition.
May 25 10:35:59 node1 crmd: warning: Too many failures (71) to fence node1, giving up
May 25 10:35:59 node1 crmd: notice: Transition aborted: Stonith failed
May 25 10:35:59 node1 crmd: error: Unfencing of node1 by failed: No route to host (-113)
May 25 10:35:59 node1 stonith-ng: notice: Couldn't find anyone to fence (on) node2 with any device
May 25 10:35:59 node1 stonith-ng: error: Operation on of node2 by for crmd.3928@node1.2680795a: No route to host
May 25 10:35:59 node1 crmd: notice: Stonith operation 143/1:72:0:f3e078bf-24f5-4160-95c1-0eeeea0e5e12: No route to host (-113)
May 25 10:35:59 node1 crmd: notice: Stonith operation 143 for node2 failed (No route to host): aborting transition.
May 25 10:35:59 node1 crmd: warning: Too many failures (71) to fence node2, giving up
May 25 10:35:59 node1 crmd: error: Unfencing of node2 by failed: No route to host (-113)
Here is the status:
[root@node1 ~]# pcs status
Cluster name: nfs_cluster
Stack: corosync
Current DC: node1 (version 1.1.20-5.amzn2.0.2-3c4c782f70) - partition with quorum
Last updated: Mon May 25 10:45:56 2020
Last change: Sun May 24 21:04:55 2020 by root via cibadmin on node1
2 nodes configured
5 resources configured
Online: [ node1 node2 ]
Full list of resources:
disk_fencing (stonith:fence_scsi): Stopped
Resource Group: nfsgrp
nfsshare (ocf::heartbeat:Filesystem): Stopped
nfsd (ocf::heartbeat:nfsserver): Stopped
nfsroot (ocf::heartbeat:exportfs): Stopped
nfsip (ocf::heartbeat:IPaddr2): Stopped
Failed Fencing Actions:
* unfencing of node2 failed: delegate=, client=crmd.3928, origin=node1,
last-failed='Mon May 25 10:35:59 2020'
* unfencing of node1 failed: delegate=, client=crmd.3928, origin=node1,
last-failed='Mon May 25 10:35:59 2020'
Daemon Status:
corosync: active/enabled
pacemaker: active/enabled
pcsd: active/enabled
[root@node1 ~]#
The disk_fencing is set to scsi, but not sure if that is the best options for two AWS ec2 instances. Perhaps I can't get disk_fencing to work so it can't start? I can ping node1 from node 2 and vice versa. Open to ideas...
Asked by jasontt33
(11 rep)
May 25, 2020, 10:49 AM