DistOS 2015W Session 7

From Soma-notes
Revision as of 01:07, 24 February 2015 by Apoorv (talk | contribs) (→‎Ceph)
Jump to navigation Jump to search

Ceph

  • Key advantage is that it is a general purpose distributed file system.
  • System is composed of three units:

*Client, *Cluster of Object Storage device (OSDs): It is basically stores data and metadata and clients communicate directly with it to perform IO operations. *MetaData Server (MDS): It is used to manage the file and directories.Client basically interacts with it to perform metadata operations like open, rename. It manages the capabilities of a client.

  • system has three key features:
    * decoupled data and metadata: 
    * Dynamic Distributed Metadata Management: It distribute the metadata among multiple metadata servers using dynamic subtree partitioning to increase the performance and avoid metadata access hot spots.
    * Object based storage: Using cluster of OSDs to form a Reliable Autonomic Distributed Object-Store(RADOS) for ceph failure detection and recovery.   
  • CRUSH (Controlled, Replicated, Under Scalable, Hashing) is the hashing algorithm used to calculate the location of object instead of looking for them. The CRUSH paper on Ceph’s website can be downloaded from here http://ceph.com/papers/weil-crush-sc06.pdf.
  • RADOS (Reliable Autonomic Distributed Object-Store) is the object store for Ceph.

Chubby

  • Is a consensus algorithm among a set of servers to agree on who is the master that is in charge of the metadata.
  • Can be considered a distributed file system for small size files only “256 KB” with very low scalability “5 servers”.
  • Is defined in the paper as “A lock service used within a loosely-coupled distributed system consisting of moderately large number of small machines connected by a high speed network”.