Professional Documents
Culture Documents
Session # 9
• Goals
– Consensus between loyal generals
– A small number of traitors cannot cause the loyals to adopt a bad plan
– Do not have to identify the traitors
• Crash fault
– Abrupt halt, never resumes operation
• Omission fault
– Processor “omits” to send required messages to some other
processors
• Malicious fault
– Processor behaves randomly and arbitrarily
– Known as Byzantine faults
Round 1: Commanding
General sends “Retreat” Commanding General 1
Round 2: L3 sends “Retreat” to L2, but
L2 sends nothing Loyal lieutenant obeys
R R commander. (good)
Decide: L3 decides “Retreat”
Lieutenant 3
Lieutenant 2 decides to retreat
R
Acknowledgement: Class lectures of A.D Brown of UofT
Round 1: Commanding
General sends “Retreat” Commanding General 1
Round 2: L3 sends “Retreat” to L2; L2
sends “Attack” to L3 Loyal lieutenant obeys
R R commander. (good)
Decide: L3 decides “Retreat”
Lieutenant 3
Lieutenant 2 decides to retreat
R
A
Round 1: Commanding
General sends “Attack” Commanding General 1
Round 2: L3 sends “Attack” to L2; L2
sends “Retreat” to L3 Loyal lieutenant disobeys
A A commander. (bad)
Decide: L3 decides “Retreat”
Lieutenant 3
Lieutenant 2 decides to retreat
A
R
Acknowledgement: Class lectures of A.D Brown of UofT
Lieutenant 3
Lieutenant 2 decides to retreat
decides to attack R
A
Acknowledgement: Class lectures of A.D Brown of UofT
P1
Stage 0: (n-1) msgs
P2 n-1
Pn
Stage 1: (n-1)(n-2) msgs
n-2
P3 n-2
Pn Pn-1 .
P2
.
. . .
P4 Pn . . Stage m
. . .
.
.
Selfish C
E F
objective P2P overlay
H
A D G layer
B
A C
E F H
AS1
AS4
B AS6
AS2 D
Native IP layer
G AS5
AS3
Improved services
A 50ms
D
20ms
20ms
C
B
• Distributed Control
• Efficient use of resources
• Self-organizing
• Scalability
• Reliability
• Ease of deployment and administration
Search API
Capability &
Configuration NAT/ Firewall Traversal
Operating System
Purely Decentralized
Partially Decentralized
• P2P Network
• All nodes are equal
• Location of arbitrary objects
• No DNS servers
• Resources
• Dynamic participation
• Data indexing
• Classification:
• Centralized
• Local
• Distributed
• Centralized indexing
• One or few central servers to store references (indexes)
• e.g., Napster
Centralized Lookup
– Centralized directory services
– Steps
• Connect to Napster server.
• Upload list of files to server.
• Give server keywords to search the full list with.
• Select “best” of correct answers
– Performance Bottleneck
• Lookup is centralized, but files are copied in P2P manner
• Loose rules
• Flooding or Random walks
• Flood query List of all content
• Ping
• Pong
• Query
• QueryHit
• Push
Scalability:
limited scope flooding
G
Object X
Query F H
Query Hit
E
B
C
D
A
Look up ‘Object X’
• .torrent
• Metadata
• Tracker
Leecher
Seeder
{1,2,3,4,5,6,7,8,9,10}
Seeder: A
{}{1,2,3}
{1,2,3,5}
{}
{1,2,3}
{1,2,3,4}
{1,2,3,4,5}
Downloader C
Downloader B
https://pdos.csail.mit.edu/papers/chord:sigcomm01/chord_sigcomm.pdf
K80 N90
N60
• Random-Graph Model
• Small-world Networks
• Clustering
• Small diameter
• Ordered lattices have clustering coefficients independent
of the network size
• Malicious attacks
Peer A
Peer B
Internet Protected
Network
A TCP Port
Peer X
Malicious Peer C Firewall
QueryHit
1 “star”,” 192.168.100.220:80”
Query: “star”
192.168.100.220:80
(target)
P2
QueryHit
Query: “pop” “pop”, ” 192.168.100.220:80”
Query: “star” “star”,” 192.168.100.220:80”
Query: “pop”
Query: “star”
Malicious
QueryHit: “star”,” 192.168.100.220:80”
QueryHit: “pop”, ” 192.168.100.220:80”
2 P3
Peer
192.168.100.40:4442
original content
polluted content
pollution
company
pollution
server
pollution
company
File sharing
network
pollution pollution
server server
pollution
server
Unsuspecting users
spread pollution! Alice
File sharing
network
Bob
46.100.80.23
index
title location
file1 120.18.89.100
file2 46.100.80.23
file3 234.8.98.20
120.18.89.100
234.8.98.20
46.100.80.23
index
title location
file1 120.18.89.100
file2 46.100.80.23
file3 234.8.98.20
file4 111.22.22.22
120.18.89.100
234.8.98.20
111.22.22.22
Victim Peer
Genuine Blocks
5. Hash Fail
Genuine Blocks
3. Block Request
1. TCP Connection
2. Fake BitMap
4. Fake Block
Genuine Blocks
Attacker
Tracker
Seeder
Free Rider
Sybil
Victim
Centralized
Peers
Peers
Super-peers
Peers
Ordinary
Peers
Fully Hybrid
Decentralized P2P P2P architecture
Dashboard
Type 2
Decision ML model
… Live traffic
Type n
Signal-processing
based features
K-nn
P2P botnets REP trees Extracted
identified ANNs Features
SVMs Network-behavior
based features
Shared Pool of
Computing Resources:
Processors, Memory, Disks
Interconnect
• Price/Performance ratio
• Incremental growth
• The provision of a multipurpose system
• Scientific, commercial, Internet applications
• Scalability
• Packaging
• Control
• Homogeneity
• Security
• Dedicated Vs Enterprise Clusters
Source: Wiki
• Programmability
• Scalable Performance
• Single-System Image (SSI)
• Availability Support
• Cluster Job Management
• Internode Communication
• Fault Tolerance and Recovery
• Cluster Family Classification:
Compute clusters
High Availability clusters
Load-balancing clusters
• Single System
• Single Control
• Symmetry
• Location Transparent
• Hot Standby
• Mutual Takeover
• Fault-Tolerance
• Failure diagnosis
• Failure notification
• Failure recovery
Check-pointing,
rollback recovery etc.
Scalable coherence
interface ring
• Compute Nodes
• Service Nodes
• Cluster Computing
• Cluster development trends
• Design objectives of Computer clusters
• Cluster organization and resource sharing
• Node architecture and MPP packaging
A Rack contains 80
PCs.
GPU Chip design for 20 Tflops performance and 1.6 TB/s memory
bandwidth in the Echelon system
• Components:
• The CPU host nodes
• The GPU nodes
• The cluster interconnect between CPU and GPU nodes
• Dedicated Mode
• Space Sharing
• Time Sharing
• Independent Scheduling
• Gang Scheduling
• Competition with Foreign Jobs
• Functional Distribution
e.g., user server each host node
• Administration is centralized
• Centralized configuration and log files
• Single user interface
• Dynamic Reconfiguration
• Suspend or Kill any job
• Serial jobs
• Parallel jobs
• Interactive jobs
• Batch jobs
• Node Availability
• Migration overhead
• Recruitment threshold
• Heterogeneous
• Support parallel and batch jobs
• Load-balancing
• Dynamic migration
• Dynamic suspension and resumption
• Dynamic addition/deletion of resources
• Command-line interface and a graphics user interface
• Cluster Computing
• Cluster system interconnects
• Hardware, software and Middleware support
• GPU Clusters for massive parallelism
• Cluster job and resource management
Computer Systems
PE
RF
O
R 2100 2100 2100 2100
M
A
N
CE
+
2100 2100 2100 2100
2100
Q Administrative Barriers
o •Individual
S •Group
•Department
•Campus
•State
•National
•Globe
Unreliable Geographically
resources and distributed
Resources are
environments
heterogeneous
• Computational Grids
• Data Grids
• Information or Knowledge Grids
• Business Grids
APPLICATIONS
Applications and Portals
USER LEVEL
Application Development and Deployment Environment
MIDDLEWARE
CORE
Distributed Resources Coupling Services
MIDDLEWARE
SECURITY LAYER
Operating Systems Queuing Systems Libraries & App Kernels … Internet Protocols
• CPU Scavenging
• Virtual Supercomputers
• Grid Resource Aggregation
GRID GARUDA initiative is a collaboration of the Scientific, Engineering and Academic Community to carry-out
research and experimentation on a nationwide grid of computational nodes, mass storage that aims to provide the
distributed data and compute intensive High Performance Computing solutions for the 21st century.
Uniform Access
System Management
Application Construction
Network Management
BITS Pilani, Pilani Campus
Grid Resource Management and Brokering
R2 Application
database
R3 R4
R5 RN
Grid Resource Broker
R6
R1
Resource Broker
• Autonomous resources
• Own resource management policies
• Resource Management System (RMS)
An economy model for assessing grid computing services, developed at the University
of Melbourne.
BITS Pilani, Pilani Campus
A grid resource broker managing
job scheduling and execution
Low-Level Middleware
Applications
Globus
Grid Data
Grid Resource Grid Information
Management
Management Services
(GridFTP, Replica
(GRAM, GASS) (MDS)
Catalog)
• Security (GSI)
• Job submission and management (GRAM)
• Information services (MDS)
• Remote file management (GASS)
• Remote Data Catalogue and Management Tools
• WSRF
• Grid Computing
• Grid Architecture and Service Modeling
• Grid Resource Management and Brokering
• Ubiquitous computing
• Network of sensor or radio connected devices
• Internet of Things (IoT)
• RFID with IPv6
• IP-Identifiable objects
Network The
Mobile Telecom Information
Layer Internet
Network Network
Sensing
Layer
RFID Label Sensor Nodes Road Mapper
• Active RFID
• Passive RFID
• Battery-assisted passive RFID
• RFID tag
• Reader antenna
• Reader
GPRS/UMTS
Cinema
STEP 2:
User having Z-SIM is
recognized by the Encrypted SMS
environment and is invited to
buy a cinema ticket Step 4:
The token is
Do you received and
confirm the
purchase of
“La fabbrica di
sent to the
cioccolato”?
printer via
ZigBee
The ticket is
automatically
printed
• Internet of Things
• IoT for Ubiquitous Computing
• RFID
• Sensors and ZigBee Technologies
• Applications of IoT
• Graph theoretic analysis of social networks
• Facebook, and Twitter case studies
In-Consistent Consistent
• There are two banking transactions ‘T’ and ‘U’ are defined as:
T U
z.deposit(10000);
aBranch.create(Z);
z.deposit(20000);
1. What is the balance of account State the balance of Z after their execution in
this order.
2. Are these consistent with serially equivalent executions of T and U?
• A1 : 20000
• A2 : Serial executions of T and U are:
o T then U: aBranch.create(“Z”); z.deposit(10);z.deposit(20). Z’s final balance is 30000.
o U then T: z.deposit(10); z.deposit(20); aBranch.create(“Z”). Z’s final balance is 0.
o Therefore, the example is not a serially equivalent execution of T and U.
• Give a use case in which multicasts sent by different clients are delivered in
different orders at two group members. Assume that some form of message
retransmissions are in use, but that messages that are not dropped arrive in
sender ordering. Provide a solution how recipients may remedy this situation?
To remedy the situation, each recipient delivers messages to its application in sender order.
When it receives a message that is ahead of the next one expected, it hold it back until it has
received and delivered the earlier re-transmitted messages.
Inconsistent Consistent
The second clock ticks 990,000 times per second, giving an error of 10
msec per second. In a minute this error has grown to 600 msec.
Other way to look is, second clock is 1% slow, so after a minute it is off by
0. 01 x 60 sec, or 600 msec.