[CentOS] mounting gfs partition hangs; Solution

Wed Mar 31 08:57:09 UTC 2010
sandra-llistes <sandra-llistes at fib.upc.edu>

It seems that all the problem was caused by selinux.
I booted all the nodes with the kernel option selinux=0 and now I can
mount GFS partitions without problems.
Thanks,

Sandra

sandra-llistes wrote:
> Hi,
> 
> I'm not a GFS expert either :-)
> In fact, I erased all the configuration of the cluster and try to do it
> again from scratch with luci/ricci configuration tools (perhaps I did
> something wrong last time)
> 
> mount -vv gives the following information:
> [root at node1 gfs]# mount -t gfs -vv /dev/home2/home2 /home2
> /sbin/mount.gfs: mount /dev/mapper/home2-home2 /home2
> /sbin/mount.gfs: parse_opts: opts = "rw"
> /sbin/mount.gfs:   clear flag 1 for "rw", flags = 0
> /sbin/mount.gfs: parse_opts: flags = 0
> /sbin/mount.gfs: parse_opts: extra = ""
> /sbin/mount.gfs: parse_opts: hostdata = ""
> /sbin/mount.gfs: parse_opts: lockproto = ""
> /sbin/mount.gfs: parse_opts: locktable = ""
> /sbin/mount.gfs: message to gfs_controld: asking to join mountgroup:
> /sbin/mount.gfs: write "join /home2 gfs lock_dlm gfs-test:gfs-data rw
> /dev/mapper/home2-home2"
> ...
> 
> And hangs at that point (On the other node it happens the same)
> 
> I tried it turning off the local firewalls on the nodes and they reached
> each other without problem with pings. Also, there are no more firewalls
> between them.
> 
> The new configuration is more simple:
> [root at node1 gfs]# more /etc/cluster/cluster.conf
> <?xml version="1.0"?>
> <cluster alias="gfs-test" config_version="6" name="gfs-test">
>         <fence_daemon clean_start="0" post_fail_delay="0"
> post_join_delay="3"/>
>         <clusternodes>
>                 <clusternode name="node1.fib.upc.es" nodeid="1" votes="1">
>                         <fence>
>                                 <method name="1">
>                                         <device name="test"
> nodename="node1.fib.upc.es"/>
>                                 </method>
>                         </fence>
>                 </clusternode>
>                 <clusternode name="node2.fib.upc.es" nodeid="2" votes="1">
>                         <fence>
>                                 <method name="1">
>                                         <device name="test"
> nodename="node2.fib.upc.es"/>
>                                         <device name="test"
> nodename="node2.fib.upc.es"/>
>                                 </method>
>                         </fence>
>                 </clusternode>
>         </clusternodes>
>         <cman expected_votes="1" two_node="1"/>
>         <fencedevices>
>                 <fencedevice agent="fence_manual" name="test"/>
>         </fencedevices>
>         <rm>
>                 <failoverdomains/>
>                 <resources>
>                         <clusterfs device="/dev/home2/home2"
> force_unmount="0" fsid="3280" fstype="gfs" mountpoint="/home2"
> name="home" self_fence="0"/>
>                 </resources>
>         </rm>
> </cluster>
> 
> Finally, I reformatted /dev/home2/home2 with the following command that
> gave no errors but it doesn't affect the final result:
> gfs_mkfs -O -j 3 -p lock_dlm -t gfs-test:gfs-data /dev/home2/home2
> 
> Thanks,
> 
> Sandra
> 
> PD: I append an strace but I can't see anything useful.