[OpenAFS] AFS namei file servers, SAN, any issues elsewhere? We've had some. Can AFS _cause_ SAN issues?

Kim Kimball dhk@ccre.com
Thu, 20 Mar 2008 13:40:07 -0600


Thanks, Jason.

Is the hardware the same as what you tested last year?

Kim


Jason Edgecombe wrote:
> Is this what you need?
>
>   PKGINST:  SUNWsan
>      NAME:  SAN Foundation Kit
>  CATEGORY:  system
>      ARCH:  sparc
>   VERSION:  1.0
>   BASEDIR:  /
>    VENDOR:  Sun Microsystems, Inc.
>      DESC:  This package provides a support for the SAN Foundation Kit.
>    PSTAMP:  sanserve-a20031029172438
>  INSTDATE:  Jan 15 2008 10:37
>   HOTLINE:  Please contact your local service provider
>    STATUS:  completely installed
>     FILES:       22 installed pathnames
>                   4 shared pathnames
>                   1 linked files
>                  11 directories
>                   2 executables
>                 239 blocks used (approx)
>
>
> Running Solaris 9 09/05HW Sparc with Sun SAN foundation.
>
> Jason
>
> Kim Kimball wrote:
>> Hi Jason,
>>
>> Thanks!
>>
>> Can you tell me which flavor of SAN you're using?
>>
>> Kim
>>
>>
>> Jason Edgecombe wrote:
>>> Robert Banz wrote:
>>>>
>>>>
>>>> AFS can't really cause "san issues" in that it's just another 
>>>> application using your filesystem.  In some cases, it can be quite 
>>>> a heavy user of such, but since its only interacting through the 
>>>> fs, its not going to know anything about your underlying storage 
>>>> fabric, or have any way of targeting it for any more badness than 
>>>> any other filesystem user.
>>>>
>>>> One of the big differences that would effect the filesystem IO load 
>>>> that occurred between 1.4.1 & 1.4.6 was the removal functions that 
>>>> made copious fsync operations.  These operations were called in 
>>>> fileserver/volserver functions that modified various in-volume 
>>>> structures, specifically file creations and deletions, and would 
>>>> lead to rather underwhelming performance when doing vos restores, 
>>>> deleting, or copying large file trees.  In many configurations, 
>>>> this causes the OS to pass on a call to the underlying storage to 
>>>> verify that all changes written have been written to *disk*, 
>>>> causing the storage controller to flush its write cache.  Since 
>>>> this defeats many of the benefits (wrt I/O scheduling) on your 
>>>> storage hardware of having a cache, this could lead to overloaded 
>>>> storage.
>>>>
>>>> Some storage devices have the option to ignore these calls from 
>>>> devices, assuming your write cache is reliable.
>>>>
>>>> Under UFS, I would suggest that you'd be running in 'logging' mode 
>>>> when using the namei fileserver on Solaris, as yes, fsck is rather 
>>>> horrible to run.  Performance on reasonably recent versions of ZFS 
>>>> were quite acceptable as well.
>>>
>>> I can confirm Robert's observations. I recently tested openafs 1.4.1 
>>> inode vs 1.4.6 namei on solaris 9 sparc with a Sun Storedge 3511 
>>> Expansion tray fibre channel device. The difference is stagerring 
>>> with vos move and such. We have been using the 1.4.6 namei config on 
>>> a SAN for a few months now with no issues.
>>>
>>> Jason
>>> _______________________________________________
>>> OpenAFS-info mailing list
>>> OpenAFS-info@openafs.org
>>> https://lists.openafs.org/mailman/listinfo/openafs-info
>>>
>>>
>>>
>> _______________________________________________
>> OpenAFS-info mailing list
>> OpenAFS-info@openafs.org
>> https://lists.openafs.org/mailman/listinfo/openafs-info
>>
>
> _______________________________________________
> OpenAFS-info mailing list
> OpenAFS-info@openafs.org
> https://lists.openafs.org/mailman/listinfo/openafs-info
>
>
>