Sven Moeller
2015-01-20 16:43:44 UTC
Hi,
i have a Cluster running on SLES 11 SP3. A resource group is defined. So every entry in that resource group is a hard dependency for the folliowing resource in that group. If just only one resource fails on both nodes, the resource res_MyApplication won't be started. I want to change that behavior. I want the cluster to start the resources on just one host. The local resources should be started on just the same host, regardless of the status of the CIFS Shares. If the CIFS Shares are available they have to be mounted on the same node running the application. I would try to accomplish this by deleting the resource group grp_application, creating colocation constrains for local fs, service ip and application. Additionally I would create co-location constrains for each CIFS Share to mount it on the node running the application. Any hints/thoughts on that? Is that the "right way" to achieve what I want?
Here's the currently running config:
node mynodea\
attributes standby="off"
node mynodeb \
attributes standby="off"
primitive res_Service-IP ocf:heartbeat:IPaddr2 \
params ip="192.168.10.120" cidr_netmask="24" \
op monitor interval="10s" timeout="20s" depth="0" \
meta target-role="started"
primitive res_MyApplication lsb:myapp \
operations $id="res_MyApplication-operations" \
op start interval="0" timeout="180" \
op stop interval="0" timeout="180" \
op status interval="0" timeout="600" \
op monitor interval="15" timeout="15" start-delay="120" \
op meta-data interval="0" timeout="600" \
meta target-role="Started"
primitive res_mount-CIFSshareData ocf:heartbeat:Filesystem \
params device="//CIFS/Share1" directory="/Datafstype="cifs" options="uid=30666,gid=30666" \
op monitor interval="20" timeout="40" depth="0" \
meta target-role="started"
primitive res_mount-CIFSshareData2 ocf:heartbeat:Filesystem \
params device="//CIFS/Share2" directory="/Data2" fstype="cifs" options="uid=30666,gid=30666" \
op monitor interval="20" timeout="40" depth="0" \
meta target-role="started"
primitive res_mount-CIFSshareData3 ocf:heartbeat:Filesystem \
params device="//CIFS/Share3" directory="/Data3" fstype="cifs" options="uid=30666,gid=30666" \
op monitor interval="20" timeout="40" depth="0" \
meta target-role="started"
primitive res_mount-application ocf:heartbeat:Filesystem \
params device="/dev/disk/by-id/scsi-36000d56e0000324561d9dcae19034e90-part1" directory="/MyApp" fstype="ext3" \
op monitor interval="20" timeout="40" depth="0" \
meta target-role="started"
primitive stonith-sbd stonith:external/sbd \
params sbd_device="/dev/disk/by-id/scsi-36000d56e0000a31a36d8dcaebaf5a439" \
meta target-role="Started"
group grp_application res_mount-application res_Service-IP res_mount-CIFSshareData res_mount-CIFSshareData2 res_mount-CIFSshareData3 res_MyApplication
location prefer-application grp_application 50: nodea
property $id="cib-bootstrap-options" \
stonith-enabled="true" \
no-quorum-policy="ignore" \
placement-strategy="balanced" \
dc-version="1.1.9-2db99f1" \
cluster-infrastructure="classic openais (with plugin)" \
expected-quorum-votes="2" \
last-lrm-refresh="1412758622" \
stonith-action="poweroff" \
stonith-timeout="216s" \
maintenance-mode="true"
rsc_defaults $id="rsc-options" \
resource-stickiness="100" \
migration-threshold="3"
op_defaults $id="op-options" \
timeout="600" \
record-pending="true"
Thanks and kind regards
Sven
_______________________________________________
Pacemaker mailing list: ***@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker
Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org
i have a Cluster running on SLES 11 SP3. A resource group is defined. So every entry in that resource group is a hard dependency for the folliowing resource in that group. If just only one resource fails on both nodes, the resource res_MyApplication won't be started. I want to change that behavior. I want the cluster to start the resources on just one host. The local resources should be started on just the same host, regardless of the status of the CIFS Shares. If the CIFS Shares are available they have to be mounted on the same node running the application. I would try to accomplish this by deleting the resource group grp_application, creating colocation constrains for local fs, service ip and application. Additionally I would create co-location constrains for each CIFS Share to mount it on the node running the application. Any hints/thoughts on that? Is that the "right way" to achieve what I want?
Here's the currently running config:
node mynodea\
attributes standby="off"
node mynodeb \
attributes standby="off"
primitive res_Service-IP ocf:heartbeat:IPaddr2 \
params ip="192.168.10.120" cidr_netmask="24" \
op monitor interval="10s" timeout="20s" depth="0" \
meta target-role="started"
primitive res_MyApplication lsb:myapp \
operations $id="res_MyApplication-operations" \
op start interval="0" timeout="180" \
op stop interval="0" timeout="180" \
op status interval="0" timeout="600" \
op monitor interval="15" timeout="15" start-delay="120" \
op meta-data interval="0" timeout="600" \
meta target-role="Started"
primitive res_mount-CIFSshareData ocf:heartbeat:Filesystem \
params device="//CIFS/Share1" directory="/Datafstype="cifs" options="uid=30666,gid=30666" \
op monitor interval="20" timeout="40" depth="0" \
meta target-role="started"
primitive res_mount-CIFSshareData2 ocf:heartbeat:Filesystem \
params device="//CIFS/Share2" directory="/Data2" fstype="cifs" options="uid=30666,gid=30666" \
op monitor interval="20" timeout="40" depth="0" \
meta target-role="started"
primitive res_mount-CIFSshareData3 ocf:heartbeat:Filesystem \
params device="//CIFS/Share3" directory="/Data3" fstype="cifs" options="uid=30666,gid=30666" \
op monitor interval="20" timeout="40" depth="0" \
meta target-role="started"
primitive res_mount-application ocf:heartbeat:Filesystem \
params device="/dev/disk/by-id/scsi-36000d56e0000324561d9dcae19034e90-part1" directory="/MyApp" fstype="ext3" \
op monitor interval="20" timeout="40" depth="0" \
meta target-role="started"
primitive stonith-sbd stonith:external/sbd \
params sbd_device="/dev/disk/by-id/scsi-36000d56e0000a31a36d8dcaebaf5a439" \
meta target-role="Started"
group grp_application res_mount-application res_Service-IP res_mount-CIFSshareData res_mount-CIFSshareData2 res_mount-CIFSshareData3 res_MyApplication
location prefer-application grp_application 50: nodea
property $id="cib-bootstrap-options" \
stonith-enabled="true" \
no-quorum-policy="ignore" \
placement-strategy="balanced" \
dc-version="1.1.9-2db99f1" \
cluster-infrastructure="classic openais (with plugin)" \
expected-quorum-votes="2" \
last-lrm-refresh="1412758622" \
stonith-action="poweroff" \
stonith-timeout="216s" \
maintenance-mode="true"
rsc_defaults $id="rsc-options" \
resource-stickiness="100" \
migration-threshold="3"
op_defaults $id="op-options" \
timeout="600" \
record-pending="true"
Thanks and kind regards
Sven
_______________________________________________
Pacemaker mailing list: ***@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker
Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org