Difference between revisions of "DistOS 2014W Lecture 18"

From Soma-notes
Jump to navigation Jump to search
(Added info about DHT's and tapestry)
Line 1: Line 1:
==Distributed Hash Tables (March 18)==
* [http://en.wikipedia.org/wiki/Distributed_hash_table Wikipedia's article on Distributed Hash Tables]
* [http://pdos.csail.mit.edu/~strib/docs/tapestry/tapestry_jsac03.pdf Zhao et al, "Tapestry: A Resilient Global-Scale Overlay for Service Deployment" (JSAC 2003)]
== Distributed Hash Table Overview ==
== Distributed Hash Table Overview ==


Line 4: Line 10:
distributed across many nodes in a network.  Keys are hashed to generate the  
distributed across many nodes in a network.  Keys are hashed to generate the  
index at which the value can be found.  Depending on the nature of the hash  
index at which the value can be found.  Depending on the nature of the hash  
function, only exact queries may be returned.  Each node has a partial view of  
function, typically, only exact queries may be returned.   
the hash table, as opposed to a full replica.  This has given rise to a number  
 
of different search techniques:
Usually, each node has a partial view of  
* A centralized server may maintain a list of all keys and associated nodes at which the value can be found.  This method involves a single point of failure.
the hash table, as opposed to a full replica. They don't know exactly which other node is responsible for a given key.  This has given rise to a number  
* Each node may query all connected nodes.  This method has performance and scalability shortcomings  
of different routing techniques:
* The keyspace can be partitioned such that nodes will maintain the values for keys that hash to similar indices (e.g., within a certain hamming distance). This method is not suitable for all applications, as it involves bandwidth-intensive migration of data stored in the DHT.
* A centralized server may maintain a list of all keys and associated nodes at which the value can be found.  This method involves a single point of failure.  
** eg. Napster
* Flooding: Each node may query all connected nodes.  This method has performance and scalability shortcomings but had the benefit of being decentralized.
** eg. Gnutella
* [http://en.wikipedia.org/wiki/Consistent_hashing Consistent Hashing] The keyspace can be partitioned such that nodes will maintain the values for keys that hash to similar indices (e.g., within a certain hamming distance). Given a query, nodes do not know specifically on which node a key is located, but they do know a few nodes (a proper subset of the network) located "closer" to the key. The query then continues onto the closest node.
** eg. Tapestry


===Tapestry:===
==Tapestry:==
Tapestry is an overlay network which makes use of a DHT to provide routing for
Tapestry is an overlay network which makes use of a DHT to provide routing for
distributed applications.  Similar to IP routing, not all nodes need to be  
distributed applications.  Similar to IP routing, not all nodes need to be  

Revision as of 21:31, 18 April 2014

Distributed Hash Tables (March 18)


Distributed Hash Table Overview

A Distributed Hash Table (DHT) is a fast lookup structure of <key,value> pairs, distributed across many nodes in a network. Keys are hashed to generate the index at which the value can be found. Depending on the nature of the hash function, typically, only exact queries may be returned.

Usually, each node has a partial view of the hash table, as opposed to a full replica. They don't know exactly which other node is responsible for a given key. This has given rise to a number of different routing techniques:

  • A centralized server may maintain a list of all keys and associated nodes at which the value can be found. This method involves a single point of failure.
    • eg. Napster
  • Flooding: Each node may query all connected nodes. This method has performance and scalability shortcomings but had the benefit of being decentralized.
    • eg. Gnutella
  • Consistent Hashing The keyspace can be partitioned such that nodes will maintain the values for keys that hash to similar indices (e.g., within a certain hamming distance). Given a query, nodes do not know specifically on which node a key is located, but they do know a few nodes (a proper subset of the network) located "closer" to the key. The query then continues onto the closest node.
    • eg. Tapestry

Tapestry:

Tapestry is an overlay network which makes use of a DHT to provide routing for distributed applications. Similar to IP routing, not all nodes need to be directly connected to each other: they can query a subset of neighbours for information about which nodes are responsible for certain parts of the keyspace. Routing is performed in such a way that nodes are aware of their distance to the object being queried. Hence objects can be located with low latency without the need to migrate actual object data between nodes. Tapestry has been used in some academic applications such as OceanStore.

    • Tapestry
  • Distributed .
  • Simple key-value store.
  • using DHT ( distributed hash table).
  • look up table contains : key and value
  • DNS as tree but Tapestry as hercically structure.
    • More dtails about Tapestry:
    • how the information flow?
  • each nod has neighbour table which that contains the node neighbour number.
    • Tapestry API:
  • have four operations called
  • each node has ID and each endpoint has GUID (Globally unique identifier).
    • Tapestry look like operating system.
  • it has two models,one is built on UDP protocol and the other on TCP protocol.