Sample Header Ad - 728x90

Cannot seem to start pcs cluster (NFS Cluster) disk_fencing trouble

1 vote
0 answers
603 views
For the life of me, I can't find a clear answer on how to start my NFS active / passive cluster. I have two nodes, node1 and node2 and followed the guide here: https://www.linuxtechi.com/configure-nfs-server-clustering-pacemaker-centos-7-rhel-7/ Here are my logs: May 25 10:35:59 node1 stonith-ng: notice: Couldn't find anyone to fence (on) node1 with any device May 25 10:35:59 node1 stonith-ng: error: Operation on of node1 by for crmd.3928@node1.97f683f8: No route to host May 25 10:35:59 node1 crmd: notice: Stonith operation 142/2:72:0:f3e078bf-24f5-4160-95c1-0eeeea0e5e12: No route to host (-113) May 25 10:35:59 node1 crmd: notice: Stonith operation 142 for node1 failed (No route to host): aborting transition. May 25 10:35:59 node1 crmd: warning: Too many failures (71) to fence node1, giving up May 25 10:35:59 node1 crmd: notice: Transition aborted: Stonith failed May 25 10:35:59 node1 crmd: error: Unfencing of node1 by failed: No route to host (-113) May 25 10:35:59 node1 stonith-ng: notice: Couldn't find anyone to fence (on) node2 with any device May 25 10:35:59 node1 stonith-ng: error: Operation on of node2 by for crmd.3928@node1.2680795a: No route to host May 25 10:35:59 node1 crmd: notice: Stonith operation 143/1:72:0:f3e078bf-24f5-4160-95c1-0eeeea0e5e12: No route to host (-113) May 25 10:35:59 node1 crmd: notice: Stonith operation 143 for node2 failed (No route to host): aborting transition. May 25 10:35:59 node1 crmd: warning: Too many failures (71) to fence node2, giving up May 25 10:35:59 node1 crmd: error: Unfencing of node2 by failed: No route to host (-113) Here is the status: [root@node1 ~]# pcs status Cluster name: nfs_cluster Stack: corosync Current DC: node1 (version 1.1.20-5.amzn2.0.2-3c4c782f70) - partition with quorum Last updated: Mon May 25 10:45:56 2020 Last change: Sun May 24 21:04:55 2020 by root via cibadmin on node1 2 nodes configured 5 resources configured Online: [ node1 node2 ] Full list of resources: disk_fencing (stonith:fence_scsi): Stopped Resource Group: nfsgrp nfsshare (ocf::heartbeat:Filesystem): Stopped nfsd (ocf::heartbeat:nfsserver): Stopped nfsroot (ocf::heartbeat:exportfs): Stopped nfsip (ocf::heartbeat:IPaddr2): Stopped Failed Fencing Actions: * unfencing of node2 failed: delegate=, client=crmd.3928, origin=node1, last-failed='Mon May 25 10:35:59 2020' * unfencing of node1 failed: delegate=, client=crmd.3928, origin=node1, last-failed='Mon May 25 10:35:59 2020' Daemon Status: corosync: active/enabled pacemaker: active/enabled pcsd: active/enabled [root@node1 ~]# The disk_fencing is set to scsi, but not sure if that is the best options for two AWS ec2 instances. Perhaps I can't get disk_fencing to work so it can't start? I can ping node1 from node 2 and vice versa. Open to ideas...
Asked by jasontt33 (11 rep)
May 25, 2020, 10:49 AM