On 06/24/2012 12:04 PM Steve Clark wrote:
On 06/24/2012 11:21 AM, ken wrote:
On 06/24/2012 09:41 AM Benjamin Franz wrote:
On 06/24/2012 12:05 AM, Gene Heskett wrote:
And what do you do when this LVM goes corrupt in about a month? I've had it self destruct on me twice. I hate it when that happens.
I would look for some other issue like bad hardware. Over the last several years I've routinely used LVM for pretty much everything and have never had it go corrupt on me except when there was a hardware failure involved. My standard buildouts use LVM over RAID.
Gene,
Yeah, the problem is more than likely in your hardware. I've used it on hundreds of machines and since 1999 and never had a problem traceable to LVM. On the other hand, I've seen a lot of disks go bad.
CentOS mailing list CentOS@centos.org http://lists.centos.org/mailman/listinfo/centos
What I don't like about LVM. at least on a personal system, is it obfuscates where things are if you have multiple underlying drives. You can't just do a df -h and see what the physical layout really is. I guess there are some pvdisplay and lvdisplay commands that can show this - but I always have to look them up and when things go kaflooey and your system isn't working then what - bring out the rescue cd and hope you can figure it out.
-- Stephen Clark *NetWolves* Director of Technology Phone: 813-579-3200 Fax: 813-882-0209 Email: steve.clark@netwolves.com http://www.netwolves.com
It helps during their creation, rather than just accepting the defaults, to give the LVs meaningful names. But even if you don't:
# df -H Filesystem Size Used Avail Use% Mounted on /dev/mapper/VolGroup00-lvroot 31G 12G 18G 39% / /dev/mapper/VolGroup00-lvtmp 195M 55M 131M 30% /tmp /dev/mapper/VolGroup00-lvvar 21G 1.2G 19G 6% /var /dev/mapper/VolGroup00-lvhome 185G 40G 136G 23% /home /dev/hda3 518M 46M 446M 10% /boot
Where's the difficulty?
On Sun, 24 Jun 2012, ken wrote:
It helps during their creation, rather than just accepting the defaults, to give the LVs meaningful names. But even if you don't:
# df -H Filesystem Size Used Avail Use% Mounted on /dev/mapper/VolGroup00-lvroot 31G 12G 18G 39% / /dev/mapper/VolGroup00-lvtmp 195M 55M 131M 30% /tmp /dev/mapper/VolGroup00-lvvar 21G 1.2G 19G 6% /var /dev/mapper/VolGroup00-lvhome 185G 40G 136G 23% /home /dev/hda3 518M 46M 446M 10% /boot
Where's the difficulty?
OK. But what about a drive that is already partitioned with live data on it. Is it easy to make that work with LVM, or does it mean I have to do a fresh installation to use LVM?
Keith
----------------------------------------------------------- Websites: http://www.karsites.net http://www.php-debuggers.net http://www.raised-from-the-dead.org.uk
All email addresses are challenge-response protected with TMDA [http://tmda.net] -----------------------------------------------------------
On 06/24/2012 12:42 PM Keith Roberts wrote:
....
OK. But what about a drive that is already partitioned with live data on it. Is it easy to make that work with LVM, or does it mean I have to do a fresh installation to use LVM?
Keith,
AFAIK, you'd need to have spare disk space (which might mean, e.g., adding a disk, plugging in a thumb drive), temporarily copying the partition there, then creating the physical volume, volume group, and logical volume, then copying the data back into the LV. To ensure data integrity, you'd most likely want to bring the system down to single-user mode when doing all of this. If the partition you're working on is read-only, and given you do the unmounting and mounting at the right times, it would be possible to do this live.
hth, ken
On 06/24/2012 12:24 PM, ken wrote:
On 06/24/2012 12:04 PM Steve Clark wrote:
On 06/24/2012 11:21 AM, ken wrote:
On 06/24/2012 09:41 AM Benjamin Franz wrote:
On 06/24/2012 12:05 AM, Gene Heskett wrote:
And what do you do when this LVM goes corrupt in about a month? I've had it self destruct on me twice. I hate it when that happens.
I would look for some other issue like bad hardware. Over the last several years I've routinely used LVM for pretty much everything and have never had it go corrupt on me except when there was a hardware failure involved. My standard buildouts use LVM over RAID.
Gene,
Yeah, the problem is more than likely in your hardware. I've used it on hundreds of machines and since 1999 and never had a problem traceable to LVM. On the other hand, I've seen a lot of disks go bad.
CentOS mailing list CentOS@centos.org http://lists.centos.org/mailman/listinfo/centos
What I don't like about LVM. at least on a personal system, is it obfuscates where things are if you have multiple underlying drives. You can't just do a df -h and see what the physical layout really is. I guess there are some pvdisplay and lvdisplay commands that can show this - but I always have to look them up and when things go kaflooey and your system isn't working then what - bring out the rescue cd and hope you can figure it out.
-- Stephen Clark *NetWolves* Director of Technology Phone: 813-579-3200 Fax: 813-882-0209 Email: steve.clark@netwolves.com http://www.netwolves.com
It helps during their creation, rather than just accepting the defaults, to give the LVs meaningful names. But even if you don't:
# df -H Filesystem Size Used Avail Use% Mounted on /dev/mapper/VolGroup00-lvroot 31G 12G 18G 39% / /dev/mapper/VolGroup00-lvtmp 195M 55M 131M 30% /tmp /dev/mapper/VolGroup00-lvvar 21G 1.2G 19G 6% /var /dev/mapper/VolGroup00-lvhome 185G 40G 136G 23% /home /dev/hda3 518M 46M 446M 10% /boot
Where's the difficulty?
What are the underlying actual physical devices?
On 24.6.2012 20:58, Steve Clark wrote:
On 06/24/2012 12:24 PM, ken wrote:
# df -H Filesystem Size Used Avail Use% Mounted on /dev/mapper/VolGroup00-lvroot 31G 12G 18G 39% / /dev/mapper/VolGroup00-lvtmp 195M 55M 131M 30% /tmp /dev/mapper/VolGroup00-lvvar 21G 1.2G 19G 6% /var /dev/mapper/VolGroup00-lvhome 185G 40G 136G 23% /home /dev/hda3 518M 46M 446M 10% /boot
Where's the difficulty?
What are the underlying actual physical devices?
$ lvdisplay --maps