[CentOS] really large file systems with centos
pasik at iki.fi
Thu Jul 14 13:53:11 UTC 2011
On Wed, Jul 13, 2011 at 11:32:14PM -0700, John R Pierce wrote:
> I've been asked for ideas on building a rather large archival storage
> system for inhouse use, on the order of 100-400TB. Probably using CentOS
> 6. The existing system this would replace is using Solaris 10 and
> ZFS, but I want to explore using Linux instead.
> We have our own tomcat based archiving software that would run on this
> storage server, along with NFS client and server. Its a write once,
> read almost never kind of application, storing compressed batches of
> archive files for a year or two. 400TB written over 2 years translates
> to about 200TB/year or about 7MB/second average write speed. The very
> rare and occasional read accesses are done by batches where a client
> makes a webservice call to get a specific set of files, then they are
> pushed as a batch to staging storage where the user can then browse
> them, this can take minutes without any problems.
> My general idea is a 2U server with 1-4 SAS cards connected to strings
> of about 48 SATA disks (4 x 12 or 3 x 16), all configured as JBOD, so
> there would potentially be 48 or 96 or 192 drives on this one server.
> I'm thinking they should be laid as as 4 or 8 or 16 seperate RAID6 sets
> of 10 disks each, then use LVM to put those into a larger volume.
> About 10% of the disks would be reserved as global hot spares.
> So, my questions...
> D) anything important I've neglected?
Remember Solaris ZFS does checksumming for all data, so with weekly/monthly ZFS scrubbing it can detect silent data/disk corruption automatically and fix it. With a lot of data, that might get pretty important..
More information about the CentOS