[CentOS] mkfs.ext3 on a 9TB volume

Joshua Baker-LePain jlb17 at duke.edu
Mon Sep 12 15:49:39 UTC 2005


On Mon, 12 Sep 2005 at 8:42am, Francois Caen wrote

> On 9/12/05, Joshua Baker-LePain <jlb17 at duke.edu> wrote:
> > As I mentioned, I'm running centos-4, which, as we all know, is based off
> > RHEL 4.  If you go to <http://www.redhat.com/software/rhel/features/>,
> > they explicitly state that they support ext3 FSs up to 8TB.
> 
> Wow! Odd! RH says 8TB but ext3 FAQ says 4TB.

I wouldn't call it that odd.  RH patches their kernels to a fair extent, 
both for stability and features.

> >From my personal testing on CentOS 4.1, you can't go over 4TB without kludging.
>  
> > I then did a software RAIDO across them, and finally:
> > 
> > mke2fs -b 4096 -j -m 0 -R stride=1024 -T largefile4 /dev/md0
> 
> Joshua, thanks for the reply on this.
> There's something kludgy about having to do softraid across 2
> partitions before formatting. It adds a layer of complexity and
> reduces reliability. Is that the trick RH recommended to go up to 8TB?

Err, it's not a kludge and it's not a trick.  Those 2 "disks" are hardware 
RAID5 arrays from 2 12 port 3ware 9500 cards.  I like 3ware's hardware 
RAID, and those are the biggest (in terms of ports) cards 3ware makes.  
So, I hook 12 disks up to each card, and the OS sees those as 2 SCSI 
disks.  I then do the software RAID to get 1) speed and 2) one partition 
to present to the users.  Folks (myself included) have been doing this for 
years.

The one gotcha in this setup (other than not being able to boot from the 
big RAID5 arrays, since each is >2TiB) is that the version of mdadm 
shipped with RHEL4 does not support array members bigger than 2TiB.  I had 
to upgrade to an upstream release to get that support.

-- 
Joshua Baker-LePain
Department of Biomedical Engineering
Duke University



More information about the CentOS mailing list