[CentOS] really large file systems with centos

Thu Jul 14 13:53:11 UTC 2011
Pasi Kärkkäinen <pasik at iki.fi>

On Wed, Jul 13, 2011 at 11:32:14PM -0700, John R Pierce wrote:
> I've been asked for ideas on building a rather large archival storage 
> system for inhouse use, on the order of 100-400TB. Probably using CentOS 
> 6.    The existing system this would replace is using Solaris 10 and 
> ZFS, but I want to explore using Linux instead.
> 
> We have our own tomcat based archiving software that would run on this 
> storage server, along with NFS client and server.   Its a write once, 
> read almost never kind of application, storing compressed batches of 
> archive files for a year or two.   400TB written over 2 years translates 
> to about 200TB/year or about 7MB/second average write speed.   The very 
> rare and occasional read accesses are done by batches where a client 
> makes a webservice call to get a specific set of files, then they are 
> pushed as a batch to staging storage where the user can then browse 
> them, this can take minutes without any problems.
> 
> My general idea is a 2U server with 1-4 SAS cards connected to strings 
> of about 48 SATA disks (4 x 12 or 3 x 16), all configured as JBOD, so 
> there would potentially be 48 or 96 or 192 drives on this one server.    
> I'm thinking they should be laid as as 4 or 8 or 16 seperate RAID6 sets 
> of 10 disks each, then use LVM to put those into a larger volume.   
> About 10% of the disks would be reserved as global hot spares.
> 
> So, my questions...
> 
> D) anything important I've neglected?
> 

Remember Solaris ZFS does checksumming for all data, so with weekly/monthly ZFS scrubbing it can detect silent data/disk corruption automatically and fix it. With a lot of data, that might get pretty important..

-- Pasi