WARNING: This is the _old_ Lustre wiki, and it is in the process of being retired. The information found here is all likely to be out of date. Please search the new wiki for more up to date information.

Architecture - Feature FS Replication

From Obsolete Lustre Wiki
Revision as of 11:40, 21 January 2010 by Docadmin (talk | contribs) (→‎Flash Cache Sync demo)
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
Jump to navigationJump to search

Summary

This article describes a feature to facilitate efficient replication of large Lustre filesystems. Target filesystems may be Lustre or any other. This article does not address replication in the presence of clustered metadata.


Requirements

  1. The solution must scale to large file systems and will avoid full file system scans
  2. The algorithm must be exact - no modified files will be missed
  3. If the file system is static and the replication is performed the target will equal the source, barring errors arising during the synchronization
  4. The algorithms is safe when run repeatedly or run after an aborted attempt, will lead to understandable results when applied to a file system that is being modified during synchronization
  5. The solutions will use a list of modified files for synchronization
  6. The solution should have a suitable architecture to synchronize flash caches or object replicas
  7. The architecture of this solution will be suitable for (i) ldiskfs (ii) Lustre 1.8 (ldiskfs with MDS&OSS) (iii) Lustre 1.10 (new MDS with OSD based fids)
  8. The solution will address future changes of log record formats (since these will contain rollback & audit information also in due course)
  9. The solution may initially only work on file system without hard links (regular files with link count > 1).
  10. The synchronization mechanism has a facility to switch the role of source and target to perform failover and failback of services.
  11. The solution must be able to deal with different (future) record formats
  12. The solution must provide for reverse replication for the recovery case

Critical use cases

identifier attribute summary
restart availability Metadata synchronization is aborted and restarted. The outcome needs to be correct.
file system full correctness Replication logs can cause the file system to be full. Correct messaging to user space is included in the solution.
MDT/OST log sync correctness OST records new file creation in log, but the event took place after the last MDT log sync
reverse replication correctness A master filesystem that is restarted after a failover to a backup filesystem must be made consistent with the backup

restart

  1. Namespace changes may not be repeated (e.g. rm a, mv b a)
  2. Rename operations may be half-finished on target (e.g. rename, but haven't updated mtime of parent dir yet when power fails, so mtime is 'now' on target, but should be 'rename time' instead)

MDT/OST log sync

Correctly deal with the following cases: after a namespace log sync (epoch 1):

  1. New file is created
    Ignore OST updates to files that were created after the epoch. The creation will be noted in the next MDT epoch, at which point the entire file (data and md) must be copied.
    • Sync namespace epoch 1
    • (data record a) Modify file foo, ino=1
    • (namespace record b) mv foo bar
    • (namespace record c) Create file foo, ino=2
    • (data record d) Modify file foo, ino=2
    • Sync data
      • (record a) lookup destname of ino=1: it is foo, so copy ino=1 to dest:/foo
      • (record d) lookup destname of ino=2: when computing destname, we determine it did not exist at epoch 1 (we see create record in the active namespace log); return a special code and don't sync this file. Old "foo" on target is not modified.
    • Sync namespace epoch 2
      • (record b) move dest:/foo to dest:/bar
      • (record c) create dest:/foo; copy ino=2 to dest:/foo
  2. File is deleted
    Leave the old version on dest alone. File will be deleted on dest in epoch 2; any recent data changes made on the source before the delete will be lost.
    • See Solution Limitation A
  3. File moved off of OST (archive, space balance, ?)
    • for file-level replication, this is a non-event
    • for object-level replication, ?? leave special redirect record stub file

Reverse Replication

filesystem A is "failed over" to a backup system B that was current as of some epoch. When failing back, updates from B must be replicated on A such that A is consistent with B.

FIXME We need to decide something here Options: Changes on A after the epoch should be:

  1. reverted
    1. deleted files can be resurrected from B
    2. created files should be deleted
    3. namespace moves are undone
    4. ? files with mtimes after the epoch are recopied from B
    5. this assumes we have complete up-to-date changelogs
  2. kept, except in the case of conflicts
    1. conflict policy: new master always wins
    2. conflict policy: latest update wins
    3. conflict policy: ? something else
    4. changes since the epoch must be replicated back to B as well

Definitions

Source File System
The file system that is being changed and that we want to replicate.
Target File System
The file system that we wish to update to be identical to the source file system.
Parent Tracker
a subsystem responsible for recording the parent inode in disk file systems inodes, for pathname reconstruction.
Changelog Generator
a subsystem used by the MDS and OSS systems to generate changelog entries transactionally when the file system changes.
Changelog Notifier
subsystem responsible for notifying that a certain number of entries have been made, OR a certain amount of time has elapsed since changelog entries were made available, OR a large amount of changelog entries remains unprocessed.
Changelog Consumer
subsystem reading changelog entries for further processing.
File Data & Attribute Synchronizer
subsystem responsible for opening the correct source and target file and synchronizing the data/attributes in the file.
Namespace Synchronizer
subsystem responsible for synchronizing the namespace. This subsystem executes creation, deletion and rename operations.
Layout Manager
changes MDS object layout information associated with the replicated file data.
Namespace Operation
creation (open with create, mknod, symlink, mkdir, link), deletion (unlink, rmdir) or change (rename) of a name
Replication Epoch
A sequence of namespace operations, a set of inodes with attribute and file data changes, bracketed by an initial and final time and record number.
Active log
The changelog to which the filesystem is currently appending change records.
Staged log
One of zero or more 'closed' logs, no longer active.


Components

600px-Components.png 600px-Deployments.png

Parent Tracker

Uses relationships

Used for
current full path lookup given an inode or fid
Requires
the disk file system extended attribute interfaces to record primary parent inode. An indexed file to record pairs of inodes and secondary parents.

Logic

  • When a new inode is created (creat, mknod, mkdir, open(O_CREAT), symlink, (link)), the parent directory inode number (inum) is automatically added as an EA. On Lustre, since OST inodes are precreated, we will modify the EA in filter_update_fidea. This means that an EA update is now required for these operations, which may change the transaction size by one block.
  • Upon rename the parent inode is changed.
  • The initial implementation may ignore multiple parents (hardlinks); for replication purposes synchronizing any one of the hardlinks is sufficient.
modifications
obdfilter (filter_update_fidea), ldiskfs (ldiskfs_new_inode)

Changelog Generator

Uses relationships

Used for
The disk file system on MDS or OSS to record changes in the file system.
Requires
llog or another transactional efficient logging mechanism to record changes, the file system or VFS api's to create logs

Exported Interfaces

  1. changelog accessor


Changelog Accessor

Changelogs are stored in a hidden directory at the filesystem root (/.changelog/). They are (mode 440 user "audit) files generated automatically by the filesystem when a special mount option is supplied (-o changelog). Only completed ("staged") log files are visible in the directory. Files are named with their changelog file sequence number. The current sequence number is stored on disk in a separate file (/.changelog/sequence).

modifications
ldiskfs; mdt/ost to allow serve special non-MDT-managed OST files to Lustre clients, mdt to serve special files directly to lustre clients without using OSTs

Interactions

500px-Changelog.png

Features and Limitations

  • Log records are placed into a special changelog file.
  • The records contain fids of existing objects and names of new objects
  • The records contain the parent fid
  • Changelog records are recorded within the filesystem transaction
  • Modified files are noted in the changelog log only once (not at every transaction)
  • Every instance of a namespace change operation (rename, create, delete) is recorded in the log

Theorem: Given a changelog with the properties described above a correct synchronization can be performed. Proof:

Logic

  1. Global changelog_active_start_time is set to the active changelog's crtime at startup and every time a new active changelog is created.
  2. When an inode is first read from disk (or created), save the old mtime, so we can easily detect when it changes
  3. Every time an inode is dirtied, check old_mtime.
    1. If old_mtime is after changelog_active_start_time, then inode has already been added to changelog; don't add it again.
    2. If old_mtime is earlier, then add the change record to the active changelog
    3. Update old_mtime to current mtime. If mtime is not later than changelog_active_start_time, we can use ctime, dtime, or current time instead of mtime; it just needs to be a time after changelog_active_start_time.
  4. Record every namespace change operation in the log for path reconstruction of a live filesystem. Record every instance, not just once per log. To simplify the log format, move/rename operations may be broken up into two log records (the first with the old name/parent, the second with the new).
    1. Include full path information in these records
modifications
ldiskfs (ldiskfs_mark_inode_dirty, ldiskfs_new_inode, ldiskfs_link, ldiskfs_unlink, ldiskfs_rmdir, ldiskfs_rename, ldiskfs_ioctl, ldiskfs_truncate, ldiskfs_xattr_set, ldiskfs_setattr)

Pathname Lookup

Uses relationships

Used for
Data Replicator
Requires
Parent Tracker, Changelog Accessor, ability to temporarily lock filesystem against namespace changes

Exported Interfaces

source_path(ino, *sourcepath, *end_rec)
ino : inode or fid from data changelog record
sourcepath : current path on source, as of end_rec
end_rec : latest namespace record number
  • may return ENOENT: file has been deleted from source
target_path(sourcepath, end_rec, namespace_log, *targetpath)
namespace_log : the name of the namespace log that we have last synced on the target
targetpath : the path for this inode on the target system, as of the end of namespace_log
  • may return ENOENT: file does not exist (yet) on target

Features and Limitations

  • Lives on MDT, exported interface to Lustre clients

Logic

source_path

  1. lock filesystem against renames
  2. lookup current full path name on MDT from ino/fid
    • open inode. Return ENOENT if doesn't exist
    • open parent ino's as stored in EA
    • likely that path elements will not be cached on the MDT during this lookup
  3. return full path in sourcepath
  4. return last record number in the active namespace changelog in end_rec
  5. unlock filesystem against renames

target_path

  1. generate a list of parent path elements from sourcepath (names or inodes)
  2. search backward through the namespace logs (active and staged) from end_rec, replacing any renamed path elements with their old (previous) version, until either:
    1. we reach the first record after end of the given namespace_log
    2. a create record for this inode is encountered; return ENOENT
  3. return targetpath
Hardlinks
source_path() will always return a single path. The path need not be consistent. target_path() will always return a single valid path on the target to one of the hardlinks (renames along this "chosen" path will be undone.) Renames along any other hardlinked paths may be ignored: target_path is used to update file data and attributes, which are shared between all the hardlinked files on the target. The renames of the other hardlinked paths themselves are synchronized by the Namespace Synchronizer.

Changelog Notifier

Uses relationships

Used for
The disk file system, to notify the Changelog Consumer of available data
Requires
Changelog Generator

Exported Interfaces

  • notification of new or excessive changelog
    • updates sequence file

Features and Limitations

  • When an active changelog has reached a preset limit of size, record count, or active time, the notifier:
    1. increases the sequence number (changelog identifier)
    2. creates a new active changelog with the new identifier
    3. marks the old changelog as staged (unhidden).
  • A Changelog Consumer could poll the sequence file for mtime changes, signalling a new staged file.

Logic

Multiple logs are kept on each disk file system volume; zero or more non-active ("finished") logs and exactly 1 active log. Finished logs will be stored in a special directory (i.e. /.changelog). Staged logs are named with a monotonically increasing sequence number. The sequence number of the latest staged log is stored on disk in a separate file (/.changelog/sequence).

  1. Begin recording into the active log. We append to this file until some trigger criteria is met (either time or log size). The active log is not visible to filesystem users.
  2. When the trigger criteria is met
    1. stop writing into the active log
    2. create a new active log
    3. thenceforth record all subsequent changes into the new active log
    4. indicate that the old log is now staged by marking it as visible and updating the sequence file
  3. When a user process has finished processing the files referred to in the staged log (including synchronizing the files remotely), it signals its completion by deleting the staged log (which may act as part of a new trigger criteria.)
  4. The cycle begins again from step 2.
  5. If the number of staged logs exceeds some threshold, the Notifier records a warning in the syslog (D_WARNING)

A user process is signaled that a new changelog is ready by polling the sequence file mtime.

If the user process dies, upon restart it re-reads the staged log, perhaps repeating old actions (sync). If the server dies, upon restart it continues recording into the active log.


modifications
ldiskfs

Changelog Consumer

500px-Sync.png

Uses relationships

Used for
Managing filesystem replication
Requires
Namespace Synchronizer, File Data & Attribute Synchronizer, Changelog Notifier

Exported Interfaces

replicate(lustre_root_path, target_root_path, [implicit target sync file])
lustre_root_path : path to root of Lustre source fs
target_root_path : path to root of target (must be locally mounted!)

The last namespace log record processed is stored in a sync file on the target. This file is read at the beginning of replicate() to determine the next start record.

Features and Limitations

The consumer uses the changelogs to coordinate filesystem replication using the Namespace Synchronizer and the File Data Synchronizer described below.

  • Multiple replica filesystems may exist at different sync points.

Logic

A synchronization cycle consists of synchronizing the namespace and file data for each changed file/dir. Special care is needed because a file/dir may be renamed at any time during the synchronization cycle, changing the path name resolution.

Synchronization for Lustre requires both OST and MDT logs, with path lookup on the MDT for any files changed on the OSTs.

  1. Synchronize the namespace change operations from a staged log using the Namespace Synchronizer (MDT)
    • Now destination namespace mirrors source, but file contents/md may not match
  2. For all other ops (md and data) in the staged log(s on the MDT and all OSTs), call the Data/MD Synchronizer
  3. Data/MD synchronization should be carried out in parallel for each OST and the MDT.
    • Now destination filesystem namespace matches, but data/md may be newer than a theoretical snapshot of the source taken at Namespace Sync time.
modifications
new userspace utility

File Data & Attribute Synchronizer

Uses relationships

Used for
Changelog Consumer - data replication
Requires
Changelog Accessor, Path Lookup, VFS api to open file by ino/fid, remote filesystem access

Features and Limitations

  • Multiple data change logs exist and these can be synchronized in parallel.
  • Individual changes are NOT recorded; only the fact that an inode is dirty. The data and metadata of the file on the target filesystem will match the source at the time of the copy, not as of the namespace sync. In other words, the target is not an exact snapshot of the source filesystem at a single point.
  • Open-by-fid requires root privileges

Logic

Find pathnames in the destination
Data records are recorded with inode/fids on the source. To transform this into an operation that can be applied to the destination file system we find the target file name using:
  1. read data record
  2. source_path(ino, *sourcepath, *end_rec)
    • if ENOENT: file has been deleted from source; don't synchronize
  3. target_path(sourcepath, end_rec, namespace_change_log, *targetpath)
    • if ENOENT: file does not exist (yet) on target; don't synchronize
Synchronize file data and attributes
  1. iopen(ino) on source for reading
    • requires root privileges on the source
  2. open(targetpath) on target for writing
  3. copy file data and attributes from source to target
    • may require temporary access permission changes on the target
    • ownership changes may require root privileges on the target
    • data copy may be predicated on mtime change, checksum, etc. May use rsync as a primitive here.
    • data changes should be made before attribute changes (mtime), in case of power failure
Parallel file processing
if multiple OSTs (and/or the MDT) all note that a file needs to be synced, they may race. Normal locking should insure no inconsistent results, but copying the same file multiple times should be avoided for efficiency. Simple rules like "don't copy if target mtime is later than source mtime" will help, but the open(target)'s may still race after such checks. Possible solutions include using a lockfile on the target (difficult to determine if lockfile is still valid after a crash bec. possible distributed lockers), or interprocess communications on the sources, or perhaps a special "sync_in_progress" bit on the MDT inode set in the open-by-fid path. FIXME nonfatal, performance
modifications
new userspace component (data syncer), ldiskfs/mdt (ioctl open by inode)

Namespace Synchronizer

Uses relationships

Used for
Changelog Consumer - namespace replicator
Requires
Changelog Accessor, remote filesystem access


Logic

Path lookup at namespace change time
Records are recorded with inode/fids and full path (up to the filesystem root) on the source at the time of the change. At this point the path is likely to be cached so this should be cheap. The old (mvFrom) and the new (mvTo) paths are recorded
Implementing a record
The metadata synchronizer processes all namespace change records in the changelog such that target filesystem namespace matches the source namespace as of the close of this changelog, including the mtime of the affected directories.
Finding the first valid record to implement
The synchronizer performs an ordered write on the target filesystem. This records the record number (transno) of the namespace record that it is currently working on in a special file on the target, as well as a full copy of the transaction record, before it makes the namespace changes to the target. This leads to a re-do operation that is understandable:
  • a deletion is repeated if the name is still there. The mtime is updated if the name is not there and the mtime does not match the post-mtime in the record.
  • a creation is repeated if the name is not yet there.
  • a rename is repeated if the names or parent mtimes are not what they should be.
  • the full transaction record on the replica is used for recovery in case of master-to-replica failover.


modifications
new userspace utility

Solution Limitations

Notable limitations to the proposed solution:

A. A replicated file system is not a snapshot taken at a single point in time on a live filesystem. It is a namespace snapshot, combined with data and metadata contents that may be later than the snapshot time. In the case of deleted files, we may miss some data/md updates.

B. Data and attribute synchronization may require temporary access permission changes on the target.

C. Open-by-inode may require root permission on source

D. A new record is not created for every data/md change; at most one per file per epoch is recorded. (However, every namespace change is recorded). This may have implications for potential audit, undo future features.

States and Transitions

Source File System

  • A directory for changelog files (per MDT/OST)
  • Staged and active changelogs
    • header: start record, start mtime
    • records: transno, ino, parent ino, mtime (see Changelog Format above)
    • footer: final record, final mtime, (future) ZFS snapshot name associated with the final state
  • Sequence file
    • active changelog filename

Changelog Format

Changelogs are packed binary data files. The first record is a header; subsequent records use the following structure:

  • magic (_u32)
  • flags (_u32)
  • record_num (__u64) (transno)
  • record_type(_u16)
  • ino(long)
  • parent_ino(long)
  • namespace changes only:
    • old mtime (struct timespec)
      note: post-op mtime is in the next namespace change rec, or current mtime
    • strlen(_u16)
    • pathname(string)
      note: for renames, "from" is in the mvFrom rec, "to" is in the mvTo rec
record_types
Create, Link, Unlink, Rmdir, mvFrom, mvTo, Ioctl, Trunc, Xattr, Setattr, *unknown

The header record:

  • magic (_u32)
  • flags (_u32)
  • start_record_num (__u64)
  • changelog_file_seqno(long)
  • log start time (struct timespec)
  • strlen(_u16)
  • note(string)

Tail record:

  • magic (_u32)
  • flags (_u32)
  • end_record_num (__u64)
  • next_changelog_seqno(long)
  • log end time (struct timespec)
  • strlen(_u16)
  • snapshot name(string) (future)

Target file system

  • Last attempted namespace record (entire record)
  • Last attempted data / metadata records for each replicator
    optional, allows for faster recovery
    • OST index
    • replication start
    • replication last completed

State Changes in the Source File System

  • Some file system transactions result in changelog records recorded in the active changelog
    • file data or metadata modifications in files that have not already been added to the active changelog
    • any namespace change operations
  • Active changelogs are occasionally staged. Multiple staged logs may exist.
  • The sequence file is updated to reflect the latest staged log.
  • "Completed" changelogs may be deleted at will (after synchronization)

File:Source state.png (image missing in original Arch wiki article)

State Changes in the Target File System

  • Each time a namespace operation is attempted the last attempted record is recorded. This may also be done for attribute and file synchronization operations, for efficient restarts (note that this would require a separate file per replication 'chunk').
  • Namespace operations are recorded on the target filesystem using user-level filesystem commands (mv, rm, mknod, touch) or the posix file API (open, creat, unlink, ioctl) to adjust file, directory, and parent directory names and metadata.
  • Data / metadata operations are recorded on the target filesystem using user-level filesystem commands (cp, rsync, touch) or the posix file API (read, write, ioctl) to synchronize file content and metadata


Related Applications

The mechanisms to use a list containing fids of modified files for synchronization hinges on opening files by fid or correctly computing the pathname on a previously synced replica. This mechanism has applications elsewhere in Lustre which we discuss here.

Flash Cache Sync demo

700px-Flash-cache-sync-demo.png

Fast Incremental Backup

700px-Incremental-backup.png

Space Balancing Migration

700px-Space-balancing-migration.png

Hadoop File System style Server Network Stripes

700px-Hdfs-sns.png

Alternatives

mtime on MDT

A always-current mtime on the MDT allows us to record changed files on the MDT only (no OST logging). Bug 11063 has an update-mtime-on-close patch, but this is unsafe in case of client crash. We could use full SOM-style recovery to guarantee mtime (backport from HEAD), but this may be very complicated given MDT changes. Arguably the changelog we are creating on the OSS nodes is similar to the recovery mechanisms used by SOM.

NOTE: on HEAD (with SOM), we need only record file (metadata) changes on the MDT, since every OST change will result in an updated mtime on the MDT inode. This eliminates the logging requirement for replication on the OSTs, removing possible racing OST syncs. The synchronizer on the MDT would be responsible for distributing chunks of the data synchronization to multiple clients and insuring their completion.

ZFS snapshots

By taking a snapshot on the MDT at the end of each epoch, the target pathname reconstruction step can be avoided. "Current" pathname of the inode is looked up on the snapshot; this gives the target pathname directly.

Potentially, if cluster-wide synchronized snapshots were available, then true snapshot backups could be made.

References

bug 14169


to be removed: HLD reqs

Entry Criteria

You need to have on hand

   1. Architecture document
   2. Quality Attribute Scenarios & Use cases
   3. LOGT, LOGD


1. External Functional specifications

     Following the architecture, define prototypes for all externally
     visible interfaces (library functions, methods etc) of all
     modules changed by the program.  Be sufficiently detailed in the
     specification to cover:
     a. Layering of API's
     b. How memory for variables / parameters is allocated
     c. In what context the functions run
   

2. High level logic

     Use high level pseudocode to indicate how all elements of the
     program will be implemented.

3. Use case scenarios

   a. Write use cases for all normal and abnormal uses of externally
       visible functions.
   b. Write use cases demonstrating interoperability between the
       software with and without this module
   c. Write use cases demonstrating the scalability use cases
      mentioned in the architecture.
   d. Include use case scenarios for all locking situations and 
      describe how likely they are.

4. State machine design

   With great care describe state machines included, used or affected
   by the module.   Describe the transitions between states.   Be
   alert to the fact that any function called by the module can
   change a state machine in the environment and this may interact
   with state machines in your model.
   Pay particular attention to:
   a. locking (ordering, lock/unlock)
   b. cache usage
   c. recovery (connect, disconnect, export/import/request state
      machines)
   d. disk state changes

5. High Level Logic Design

   a What synchronization primitives (lock types etc) need to be chosen 
        to handle locking use cases most efficiently.

6. Test plan

  Give a high level design of all tests and describe test cases to
  cover/verify all critical use cases and quality attribute scenarios


7. Plan review

  Review the estimates in the TSP cycle plan based on the information
  obtained during the work on the HLD.  Bring wrong estimates to the
  attention of the planning manager in the weekly meeting.

8. EXIT CRITERIA

  A well formed design with prototypes, state machines, use cases,
  logic.