[OpenAFS] Which storage technology to use for terabytes of storage with AFS?

Rob Banz rob@nofocus.org
Fri, 30 Nov 2007 14:54:02 -0500


Run solaris x86! The site I used to work at has been using it  
exclusively on its afs servers since sol 10 came out.

-rob


On Nov 30, 2007, at 13:51, Stephen Joyce <stephen@physics.unc.edu>  
wrote:

> I don't have money for FC or a SAN, so I've stuck with DAS. I've had  
> good experience with building many smallish servers rather than one  
> big or expensive one.
>
> I'm currently using cheap Dell PowerEdge servers running linux. I  
> think we got them for about $800/ea, and they support console  
> redirection (critical when you have lots of physical servers). We  
> added a 2-port 3ware raid1 for the OS and a 4-port 3ware for the  
> data (raid1 or raid5 depending on the requirements). Right now I'm  
> keeping the servers to around 1TB each, but they're capable of  
> hosting 2-4TB each (depending on raid level) with the largest  
> current drives.
>
> If money were no object, I'd have opted for hot-swappable drives,  
> but with under 1TB of data on each, any time I've needed to replace  
> a drive I've just moved the volumes to another server.
>
> These systems are cheap enough (under about $1.5K each for  
> everything) that I keep a spare of everything just in case (spare  
> fully configured and running server plus spare raid cards and drives  
> on the shelf).
>
> I _strongly_ advise raid. Raid1 for the OS and raid1, 5, or 6 for  
> the data, depending on your requirements. I know some people have  
> reported impressive results with linux software raid, but I swear by  
> 3ware hardware raid controllers; they "just work." Just avoid  
> "fakeraid" controller cards (promise, low-end adaptec, etc) like the  
> plague. They're far more trouble than they're worth.
>
> I really like solaris, but this setup is MUCH cheaper and faster  
> than our old solaris setup.
>
> On Fri, 30 Nov 2007, Jason Edgecombe wrote:
>
>> Hi everyone,
>>
>> Traditionally, we have used direct-attached scsi disk packs on Sun  
>> Sparc
>> servers running Solaria 9 for OpenAFS. This has given us the most  
>> bang
>> for the buck. We forgo RAID because we have the backup capabilities  
>> of AFS.
>>
>> What types of storage technologies are other AFS sites using for  
>> their
>> AFS vicep partitions? We need to figure our future direction for the
>> next couple of years. Fibre channel seems all the rage, but it's  
>> quite
>> expensive. I'm open to any and all feedback. What works? What  
>> doesn't?
>> What offers the best bang for the buck on an OpenAFS server?
>>
>> This is for an academic environment that fills both academic and
>> research needs. Researchers are asking for lots of AFS space (200GB 
>> +).
>> Of course this needs to be backed up as well.
>>
>> Thanks,
>> Jason
>
> Cheers, Stephen
> --
> Stephen Joyce
> Systems Administrator                                            P A  
> N I C
> Physics & Astronomy Department                         Physics &  
> Astronomy
> University of North Carolina at Chapel Hill         Network  
> Infrastructure
> voice: (919) 962-7214                                        and  
> Computing
> fax: (919) 962-0480                               http://www.panic.unc.edu
>
> Don't judge a book by its movie.
> _______________________________________________
> OpenAFS-info mailing list
> OpenAFS-info@openafs.org
> https://lists.openafs.org/mailman/listinfo/openafs-info