[OpenAFS] Re: Need volume state / fileserver / salvage knowledge

Jeff Blaine jblaine@kickflop.net
Fri, 28 Jan 2011 13:17:31 -0500


Examples from FileLog.old:

Fri Jan 28 10:02:48 2011 VAttachVolume: volume /vicepf/V2023864046.vol 
needs to be salvaged; not attached.

Fri Jan 28 10:02:49 2011 VAttachVolume: volume salvage flag is ON for 
/vicepa//V2023886583.vol; volume needs salvage

Examples from SalvageLog old pretty much run the gamut (it's
a 4MB file...).

01/28/2011 10:30:50 Found 13 orphaned files and directories (approx. 26 KB)

01/28/2011 10:30:52 Volume uniquifier is too low; fixed

01/28/2011 10:31:11 Vnode 34: version < inode version; fixed (old status)

01/28/2011 12:54:15 Volume 536872710 (src.local) mount point ./flex/011 
to '#src.flex.011#' invalid, converted to symbolic link

01/28/2011 12:27:30 dir vnode 15: special old unlink-while-referenced 
file .__afs9803 is deleted (vnode 2248)

01/28/2011 12:28:22 dir vnode 1075: ./.gconfd/lock/ior (vnode 4272): 
unique changed from 54370 to 57920

01/28/2011 12:28:22 dir vnode 1077: ./.gconf/%gconf-xml-backend.lock/ior 
already claimed by directory vnode 1 (vnode 4278, unique 54373) -- deleted

01/28/2011 12:28:28 dir vnode 607: invalid entry: ./.gconfd/lock/ior 
(vnode 1114, unique 132811)

01/28/2011 12:37:28 dir vnode 1: invalid entry deleted: 
./.ab_library.lock (vnode 50816, unique 25535)

On 1/28/2011 12:33 PM, Andrew Deason wrote:
> On Fri, 28 Jan 2011 12:10:38 -0500
> Jeff Blaine<jblaine@kickflop.net>  wrote:
>
>> The last time we brought our fileservers down (cleanly, according to
>> "shutdown" info via bos status), it struck me as odd that salvages
>> were needed once it came up.  I sort of brushed it off.
>
> As in, it salvaged everything automatically when it came back up, or
> volumes were not attached when it came back up, and you needed to
> salvage to bring them online?
>
>> We've done it again, and the same situation is presenting itself,
>> and I'm really confused as to how that is and what is happening
>> incorrectly.  One of the three cleanly shutdown fileservers came
>> up with hundreds of unattachable volumes, and is salvaging now
>> by our hand.
>
> Well, why are they not attaching? FileLog should tell you. And the
> salvage logs should say what they fixed, if anything, to bring them back
> online.
>
> Also, salvaging an entire partition at once may be quite a bit faster
> than salvaging volumes individually, depending on how many volumes you
> have. The fileserver needs to be shutdown for that to happen, though.
>