Running into a bit of a problem here , I set up two servers ( Centos 6 ) with Glusterfs and a shared directory between them, I have moved the nfs directory to the shared Gluster folder and have created a symlink on both boxes. The machines can talk to themselves via hostnames and Gluster replication is handled via another ethernet card between the servers.
The problem I am having, is that even though the resources fail over correctly ( though it seems to come up and down a few times while failing over ) , I get stale nfs handles on the client. Below is my crm config; what am I doing wrong?
The nfs mount on the client is as plain as possible.
node GlusterFS01
node GlusterFS02
primitive ClusterIP ocf:heartbeat:IPaddr2 \
params ip="10.10.10.167" cidr_netmask="24" clusterip_hash="sourceip" \
op monitor interval="5s"
primitive exportfs ocf:heartbeat:exportfs \
params fsid="0" directory="/GlusterFS/Files" \
options="rw,sync,no_subtree_check,no_root_squash" \
clientspec="10.10.10.0/24" \
wait_for_leasetime_on_stop="false" \
op monitor interval="5s" \
op start interval="0s" timeout="240s" \
op stop interval="0s" timeout="100s" \
meta is-managed="true" target-role="Started"
primitive nfs lsb:nfs \
meta target-role="Started" \
op monitor interval="5s" timeout="5s"
colocation sitewithnfs inf: ClusterIP exportfs nfs
order nfsorder inf: exportfs ClusterIP nfs
property $id="cib-bootstrap-options" \
dc-version="1.1.10-14.el6_5.2-368c726" \
cluster-infrastructure="classic openais (with plugin)" \
expected-quorum-votes="2" \
stonith-enabled="false" \
no-quorum-policy="ignore" \
last-lrm-refresh="1395246465" \
default-resource-stickiness="100"
rsc_defaults $id="rsc-options" \
resource-stickiness="100"
Thank you for your time.
Update1: I have decided that I was overcomplicating everything. After a call with Florian, he convinced me to simplify. I am sharing nfs directly from Gluster, and I just have the ip resource being handled by corosync/pacemaker. Much simpler solution and it fits my needs.
I will say however that Dok was completely correct in his assessment and suggestions, even though I was not able to get it up and running 100% on the productions environment ( even thought is worked in testing ).