[OpenAFS] Re: Need volume state / fileserver / salvage knowledge
Jeff Blaine
jblaine@kickflop.net
Fri, 28 Jan 2011 13:17:31 -0500
Examples from FileLog.old:
Fri Jan 28 10:02:48 2011 VAttachVolume: volume /vicepf/V2023864046.vol
needs to be salvaged; not attached.
Fri Jan 28 10:02:49 2011 VAttachVolume: volume salvage flag is ON for
/vicepa//V2023886583.vol; volume needs salvage
Examples from SalvageLog old pretty much run the gamut (it's
a 4MB file...).
01/28/2011 10:30:50 Found 13 orphaned files and directories (approx. 26 KB)
01/28/2011 10:30:52 Volume uniquifier is too low; fixed
01/28/2011 10:31:11 Vnode 34: version < inode version; fixed (old status)
01/28/2011 12:54:15 Volume 536872710 (src.local) mount point ./flex/011
to '#src.flex.011#' invalid, converted to symbolic link
01/28/2011 12:27:30 dir vnode 15: special old unlink-while-referenced
file .__afs9803 is deleted (vnode 2248)
01/28/2011 12:28:22 dir vnode 1075: ./.gconfd/lock/ior (vnode 4272):
unique changed from 54370 to 57920
01/28/2011 12:28:22 dir vnode 1077: ./.gconf/%gconf-xml-backend.lock/ior
already claimed by directory vnode 1 (vnode 4278, unique 54373) -- deleted
01/28/2011 12:28:28 dir vnode 607: invalid entry: ./.gconfd/lock/ior
(vnode 1114, unique 132811)
01/28/2011 12:37:28 dir vnode 1: invalid entry deleted:
./.ab_library.lock (vnode 50816, unique 25535)
On 1/28/2011 12:33 PM, Andrew Deason wrote:
> On Fri, 28 Jan 2011 12:10:38 -0500
> Jeff Blaine<jblaine@kickflop.net> wrote:
>
>> The last time we brought our fileservers down (cleanly, according to
>> "shutdown" info via bos status), it struck me as odd that salvages
>> were needed once it came up. I sort of brushed it off.
>
> As in, it salvaged everything automatically when it came back up, or
> volumes were not attached when it came back up, and you needed to
> salvage to bring them online?
>
>> We've done it again, and the same situation is presenting itself,
>> and I'm really confused as to how that is and what is happening
>> incorrectly. One of the three cleanly shutdown fileservers came
>> up with hundreds of unattachable volumes, and is salvaging now
>> by our hand.
>
> Well, why are they not attaching? FileLog should tell you. And the
> salvage logs should say what they fixed, if anything, to bring them back
> online.
>
> Also, salvaging an entire partition at once may be quite a bit faster
> than salvaging volumes individually, depending on how many volumes you
> have. The fileserver needs to be shutdown for that to happen, though.
>