[OpenAFS] Replication, Fail-over, Disconnected Operation and Caching
openafs@bobich.net
openafs@bobich.net
Fri, 25 Jan 2008 17:20:02 +0000 (GMT)
Hi,
I've looked through the documentation, but couldn't find any specifics on
this, so I'd be grateful if somebody could point me at the page I've
missed.
1) How do OpenAFS clients pick a server to access a volume from if the
volume is replicated on multiple servers?
2) From the documentation, it looks like the replication mechanism is
single-master / multiple-slaves, i.e. one read-write server, multiple
read-only servers. Is that correct? If so, do clients transparently handle
this? Are writes transparently routed to the read-write server while still
allowing reads to come from a more local, faster, read-only server.
3) Can the root volume be replicated? What I am really looking to do is
have 2 servers, one as master and the other with all the volumes
replicated. Is that possible?
4) If the read-write server fails, how does OpenAFS handle failing over to
the replicated backup? When the original master comes back up, how
transparently / gracefully does this happen?
5) Is disconnected operation supported via local caching (as per Coda)? If
so, are there limits on sane cache sizes? Is it reasonable to expect to
have tens of GB of cached content available on the client nodes?
I am currently using GFS in reliable environments, and Coda on a small
scale in environments that have to tollerate disconnections, but I have
concerns about Coda's stability (perpetual betaware, or so it seems) in
larger and harsher environments (terabytes of storage, hundreds of
clients, thousands of users), hence why I am looking at OpenAFS as a
possible more stable alternative.
Thanks in advance.
Gordan