[CentOS] Recover botched drdb gfs2 setup .

Mon Feb 28 00:54:32 UTC 2011
Machin, Greg <Greg.Machin at openpolytechnic.ac.nz>

Hi.

No worries it was a firewall issue. Not quite as bad as I though J .

 

 

Greg Machin
Systems Administrator - Linux
Infrastructure Group, Information Services

 

From: centos-bounces at centos.org [mailto:centos-bounces at centos.org] On
Behalf Of Machin, Greg
Sent: Monday, 28 February 2011 9:33 a.m.
To: CentOS mailing list
Subject: [CentOS] Recover botched drdb gfs2 setup .

 

Hi.

The short story... Rush job, never done clustered file systems before,
vlan didn't support multicast. Thus I ended up with drbd working ok
between the two servers but cman / gfs2 not working, resulting in what
was meant to be a drbd primary/primary cluster being a primary/secondary
cluster until the vlan could be fixed with gfs only mounted on the one
server. I got the single server working and left to for the contractor
to do there bit. Two months down the line and a few other hiccups in the
mix I have a server that wont mount the gfs partition.. assuming that
drbd hasn't gotten confused and lost the data on the drive..

 

If I can how do I fix this.

 

Drbd is currently as follows: 

[root at mcvpsam01 init.d]# drbd-overview

  1:r0  WFConnection Primary/Unknown UpToDate/DUnknown C r----

 

Cman:

 

[root at mcvpsam01 init.d]# /etc/init.d/cman status

groupd is stopped  

 

gfs2 mount

 

[root at mcvpsam01 init.d]# ./gfsmount.sh start

Mounting gfs2 partition

/sbin/mount.gfs2: can't connect to gfs_controld: Connection refused

/sbin/mount.gfs2: can't connect to gfs_controld: Connection refused

/sbin/mount.gfs2: can't connect to gfs_controld: Connection refused

/sbin/mount.gfs2: can't connect to gfs_controld: Connection refused

/sbin/mount.gfs2: can't connect to gfs_controld: Connection refused

/sbin/mount.gfs2: can't connect to gfs_controld: Connection refused

/sbin/mount.gfs2: can't connect to gfs_controld: Connection refused

/sbin/mount.gfs2: can't connect to gfs_controld: Connection refused

/sbin/mount.gfs2: can't connect to gfs_controld: Connection refused

/sbin/mount.gfs2: can't connect to gfs_controld: Connection refused

/sbin/mount.gfs2: gfs_controld not running

/sbin/mount.gfs2: error mounting lockproto lock_dlm

[root at mcvpsam01 init.d]#

 

And log/messages

 

Feb 28 09:20:39 mcvpsam01 openais[3328]: [TOTEM] The consensus timeout
expired.

Feb 28 09:20:39 mcvpsam01 openais[3328]: [TOTEM] entering GATHER state
from 3.

Feb 28 09:20:54 mcvpsam01 openais[3328]: [TOTEM] The consensus timeout
expired.

Feb 28 09:20:54 mcvpsam01 openais[3328]: [TOTEM] entering GATHER state
from 3.

Feb 28 09:21:09 mcvpsam01 openais[3328]: [TOTEM] The consensus timeout
expired.

Feb 28 09:21:09 mcvpsam01 openais[3328]: [TOTEM] entering GATHER state
from 3.

 

cluster.conf

 

[root at mcvpsam01 init.d]# cat /etc/cluster/cluster.conf

<?xml version="1.0"?>

<cluster alias="cluster-setup" config_version="1" name="cluster-setup">

  <rm log_level="4"/>

  <fence_daemon clean_start="1" post_fail_delay="0"
post_join_delay="3"/>

  <clusternodes>

    <clusternode name="mcvpsam01" nodeid="1" votes="1">

      <fence>

        <method name="2">

          <device name="LastResortNode01"/>

        </method>

      </fence>

    </clusternode>

    <clusternode name="drvpsam01" nodeid="2" votes="1">

      <fence>

        <method name="2">

          <device name="LastResortNode02"/>

        </method>

      </fence>

    </clusternode>

  </clusternodes>

  <cman expected_votes="1" two_node="1"/>

  <fencedevices>

    <fencedevice agent="fence_manual" name="LastResortNode01"
nodename="mcvpsam01"/>

    <fencedevice agent="fence_manual" name="LastResortNode02"
nodename="drvpsam01"/>

  </fencedevices>

  <rm/>

  <totem consensus="4800" join="60" token="10000"
token_retransmits_before_loss_const="20"/>

</cluster>

[root at mcvpsam01 init.d]#

 

Drbd.conf

 

[root at mcvpsam01 init.d]# cat /etc/drbd.conf

resource r0 {

protocol C;

 

syncer { rate 1000M; }

 

 

startup {

    wfc-timeout 120;                    # wait 2min for other peers

    degr-wfc-timeout 120;               # wait 2min if peer was already

                                        # down before this node was
rebooted

    become-primary-on both;

  }

 

net {

    allow-two-primaries;

#    cram-hmac-alg "sha1";              # algo to enable peer
authentication

#    shared-secret "123456";

 

    # handle split-brain situations

    after-sb-0pri discard-least-changes;# if no primary auto sync from
the

                                        # node that touched more blocks
during

                                        # the split brain situation.

    after-sb-1pri discard-secondary;    # if one primary

    after-sb-2pri disconnect;           # if two primaries

 

    # solve the cases when the outcome

    # of the resync decision is incompatible

    # with the current role assignment in

    # the cluster

    rr-conflict disconnect;             # no automatic resynchronization

                                        # simply disconnect

}

 

 

disk {

    on-io-error detach;                 # detach the device from its

                                        # backing storage if the driver
of

                                        # the lower_device reports an
error

                                        # to DRBD

#    fencing resource-and-stonith;

  }

 

 

 on mcvpsam01 {

        device /dev/drbd1;

        disk /dev/sdb1;

        address 202.37.1.133:7789;

        meta-disk internal;

        }

 

 on drvpsam01 {

        device /dev/drbd1;

        disk /dev/sdb1;

        address 202.37.1.134:7789;

        meta-disk internal;

        }

}

[root at mcvpsam01 init.d]#

 

 

[root at mcvpsam01 init.d]# cat /etc/drbd.d/global_common.conf

global {

        usage-count yes;

        # minor-count dialog-refresh disable-ip-verification

}

 

common {

        protocol C;

 

        handlers {

                pri-on-incon-degr
"/usr/lib/drbd/notify-pri-on-incon-degr.sh;
/usr/lib/drbd/notify-emergency-reboot.sh; echo b > /proc/sysrq-trigger ;
reboot -f";

                pri-lost-after-sb
"/usr/lib/drbd/notify-pri-lost-after-sb.sh;
/usr/lib/drbd/notify-emergency-reboot.sh; echo b > /proc/sysrq-trigger ;
reboot -f";

                local-io-error "/usr/lib/drbd/notify-io-error.sh;
/usr/lib/drbd/notify-emergency-shutdown.sh; echo o > /proc/sysrq-trigger
; halt -f";

                # fence-peer "/usr/lib/drbd/crm-fence-peer.sh";

                # split-brain "/usr/lib/drbd/notify-split-brain.sh
root";

                # out-of-sync "/usr/lib/drbd/notify-out-of-sync.sh
root";

                # before-resync-target
"/usr/lib/drbd/snapshot-resync-target-lvm.sh -p 15 -- -c 16k";

                # after-resync-target
/usr/lib/drbd/unsnapshot-resync-target-lvm.sh;

        }

 

        startup {

                # wfc-timeout degr-wfc-timeout outdated-wfc-timeout
wait-after-sb

        }

 

        disk {

                # on-io-error fencing use-bmbv no-disk-barrier
no-disk-flushes

                # no-disk-drain no-md-flushes max-bio-bvecs

        }

 

        net {

                # sndbuf-size rcvbuf-size timeout connect-int ping-int
ping-timeout max-buffers

                # max-epoch-size ko-count allow-two-primaries
cram-hmac-alg shared-secret

                # after-sb-0pri after-sb-1pri after-sb-2pri
data-integrity-alg no-tcp-cork

        }

 

        syncer {

                # rate after al-extents use-rle cpu-mask verify-alg
csums-alg

        }

}

[root at mcvpsam01 init.d]#

 

 

Any ideas how I can get the file system mounted to recover the data.

 

 

 

Thanks

 

 

 

 

Greg Machin
Systems Administrator - Linux
Infrastructure Group, Information Services

 

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.centos.org/pipermail/centos/attachments/20110228/7ae8b5cb/attachment-0003.html>