[OpenAFS] AFS namei file servers, SAN, any issues elsewhere? We've had some. Can AFS _cause_ SAN issues?

Jason Edgecombe jason@rampaginggeek.com
Wed, 19 Mar 2008 20:56:32 -0400


Is this what you need?

   PKGINST:  SUNWsan
      NAME:  SAN Foundation Kit
  CATEGORY:  system
      ARCH:  sparc
   VERSION:  1.0
   BASEDIR:  /
    VENDOR:  Sun Microsystems, Inc.
      DESC:  This package provides a support for the SAN Foundation Kit.
    PSTAMP:  sanserve-a20031029172438
  INSTDATE:  Jan 15 2008 10:37
   HOTLINE:  Please contact your local service provider
    STATUS:  completely installed
     FILES:       22 installed pathnames
                   4 shared pathnames
                   1 linked files
                  11 directories
                   2 executables
                 239 blocks used (approx)


Running Solaris 9 09/05HW Sparc with Sun SAN foundation.

Jason

Kim Kimball wrote:
> Hi Jason,
>
> Thanks!
>
> Can you tell me which flavor of SAN you're using?
>
> Kim
>
>
> Jason Edgecombe wrote:
>> Robert Banz wrote:
>>>
>>>
>>> AFS can't really cause "san issues" in that it's just another 
>>> application using your filesystem.  In some cases, it can be quite a 
>>> heavy user of such, but since its only interacting through the fs, 
>>> its not going to know anything about your underlying storage fabric, 
>>> or have any way of targeting it for any more badness than any other 
>>> filesystem user.
>>>
>>> One of the big differences that would effect the filesystem IO load 
>>> that occurred between 1.4.1 & 1.4.6 was the removal functions that 
>>> made copious fsync operations.  These operations were called in 
>>> fileserver/volserver functions that modified various in-volume 
>>> structures, specifically file creations and deletions, and would 
>>> lead to rather underwhelming performance when doing vos restores, 
>>> deleting, or copying large file trees.  In many configurations, this 
>>> causes the OS to pass on a call to the underlying storage to verify 
>>> that all changes written have been written to *disk*, causing the 
>>> storage controller to flush its write cache.  Since this defeats 
>>> many of the benefits (wrt I/O scheduling) on your storage hardware 
>>> of having a cache, this could lead to overloaded storage.
>>>
>>> Some storage devices have the option to ignore these calls from 
>>> devices, assuming your write cache is reliable.
>>>
>>> Under UFS, I would suggest that you'd be running in 'logging' mode 
>>> when using the namei fileserver on Solaris, as yes, fsck is rather 
>>> horrible to run.  Performance on reasonably recent versions of ZFS 
>>> were quite acceptable as well.
>>
>> I can confirm Robert's observations. I recently tested openafs 1.4.1 
>> inode vs 1.4.6 namei on solaris 9 sparc with a Sun Storedge 3511 
>> Expansion tray fibre channel device. The difference is stagerring 
>> with vos move and such. We have been using the 1.4.6 namei config on 
>> a SAN for a few months now with no issues.
>>
>> Jason
>> _______________________________________________
>> OpenAFS-info mailing list
>> OpenAFS-info@openafs.org
>> https://lists.openafs.org/mailman/listinfo/openafs-info
>>
>>
>>
> _______________________________________________
> OpenAFS-info mailing list
> OpenAFS-info@openafs.org
> https://lists.openafs.org/mailman/listinfo/openafs-info
>