You are on page 1of 5

ISS: INTRO TO SCALABLE SYSTEM

Strong Scaling and Weak Scaling Results


NAME: AVINASH KUMAR SR NO.21773 DATE:
NOV 2,2022

For strong scaling,

1)
a) Application: “HTS: A Threaded Multilevel Sparse Hybrid Solver”
b) Number of cores: 32
c) Scaling obtained: linear
d) speedup obtained: 20 times (5%)

HTS have over 20× speedup on 32 cores is incredible for sparse


linear solver packages

2)

a) Application: “A scalable adaptive-matrix SPMV for heterogeneous architectures”


b) Number of cores up to which strong scaling results were shown: 14336
c) Scaling obtained: quadratic
d) Speedup obtained: 2.9x in setup, 1.4x than SpMV
3)
a) Application: “MICCO: An Enhanced Multi-GPU Scheduling Framework for Many-
Body Correlation Functions”
b) Number of cores: 256cores, 8 GPU
c) comment on the kind of scaling obtained (linear or any other),
d) speedup obtained: 1.96x

Scalability. Tensor size is 384. Vector size is 64.

4)
a) Application:” Virtual-Link: A Scalable Multi-Producer Multi-Consumer Message Queue
Architecture for Cross-Core Communication ”
b) Number of cores: 15
c) comment on the kind of scaling obtained: logarithmic
d) speedup obtained: 2.09x

5)
a) Application: “NVMe-CR: A Scalable Ephemeral Storage Runtime for
Checkpoint/Restart with NVMe-over-Fabrics”
b) Number of cores: 16
c) scaling obtained: linear

d) speedup obtained: 2x
For weak scaling,

1)
a) Application:” HTS: A Threaded Multilevel Sparse Hybrid Solver”
b) Weak scaling methodology: input size doubled and also the no of cores used
c) Number of cores up to which the results were shown: 16
d) comment on the weak scaling results: we do observe that as the problem size
increases HTS can continue to scale,

2)
a) Application: “A scalable adaptive-matrix SPMV for heterogeneous architectures”
b) Weak scaling methodology: 488 DoFs per MPI process
c) Number of cores up to which the results were shown:66
d) comment on the weak scaling results: 3x in setup time and 1.5x in SpMV than
PETSc-GPU
3)
a) Application: “PARallel Subgraph Enumeration in CUDA”
b) the weak scaling methodology used: we use synthetic random geometric graphs
(RGGs), A synthetic RGG at scale s has exactly 2s nodes
c) the number of cores: 224
d) comment on the weak scaling results: increasing in number of nodes proportional to
input size. Execution time shows linear result as expected.

4)
a) mention the application: “MICCO: An Enhanced Multi-GPU Scheduling Framework for
Many-Body Correlation Functions”
b) the weak scaling methodology used: Tensor size increases by 128 each time
c) the number of cores up to which the results were shown (8 GPU each having 32
cores=256 cores)
d) comment on the weak scaling results: GFLOPS proportional to increase in tensor size

Tensor size varies from 128 to 768. Vector size is 64.

5)
a) Application:” NVMe-CR: A Scalable Ephemeral Storage Runtime for
Checkpoint/Restart with NVMe-over-Fabrics”
b) Weak scaling methodology used: 32K atoms per process
c) Number of cores:16
d) comment on the weak scaling results: NVMe-CR achieves near perfect efficiency
(0.96 for checkpoint and 0.99 for recovery) at 448 processes

You might also like