A Micro-Benchmark Suite for Evaluating HDFS Operations on

A Micro-Benchmark Suite for
Evaluating HDFS Operations on
Modern Clusters
HDFS Performace Factors
• Performance of HDFS is determined by the
following 3 factors
• Factors related to storage and network
configurations
• Controllable parameters (block size, packet size)
• Data Access Pattern
• A benchmark tool suite to evaluate HDFS
performance metrics in different
configurations is addressed in this paper.
HDFS Benchmarks
• Most popular benchmark to evaluate HDFS
I/O performance is TestDFSIO
• But, it involves the MapReduce framework.
• There is a lack of standardized benchmark
suite to evaluate the performance of just
standalone HDFS.
Designing such a HDFS benchmark
suite
• HDFS has three main operations
• Sequential Write.
• Sequential Read.
• Random Read.
• HDFS performance is measured by the latency
and throughput of these operations
• Performance is influenced by underlying
network, storage, HDFS configuration
parameters and data access patterns
• This benchmark suite
• Focuses on three kinds of data access patterns:
sequential, random and mix (read and write)
• Has options to set HDFS configuration parameters
dynamically
Benchmark Suite
• Five different benchmarks
•
•
•
•
•
Sequential Write Latency (SWL)
Sequential or Random Read Latency (SRL or RRL)
Sequential Write Throughput (SWT)
Sequential Read Throughput (SRT)
Sequential Read - Write Throughput (SRWT)
• All the above benchmarks use HDFS API for
Read and Write
Benchmark Parameter List
The benchmark suite calculates statistics like min,
max and average of latency and throughput.
Experimental Setup
• Intel Westmere Cluster
• Each node has 8 processor cores on 2 Intel Xeon
2.67 GHz Quadcore CPUs, 12 GB main memory,
160 GB hard disk
• Network: 1GigE and IPoIB (32Gbps)
Sequential Write Latency (SWL)
Sequential and Random Read Latency
(SRL / RRL)
Sequential Write Throughput
(SWT with 4 writers)
Sequential Read Throughput (SRT)
Evaluation of Mixed Overload
(SRWT with 4 readers , 4 writers)
Conclusion
• Design, development and implementation of a
micro benchmark suite to evaluate
performance of standalone HDFS
• Flexible infrastructure for the benchmarks to
set HDFS configuration parameters
dynamically
• Performance evaluations with our
benchmarks over different interconnects on
modern clusters
Performance Analysis and Evaluation
of InfiniBand FDR and 40GigE RoCE on
HPC and Cloud Computing Systems
Problem statement
• No study has been done to address the
following questions:
• How much benefit can the user of a HPC / Cloud
installation hope to see by utilizing IB FDR / RoCE
40 GigE over IB QDR and RoCE 10 GigE
interconnects, respectively?
• How does InfiniBand compare with RoCE in terms
of performance?
Experimental Setup
• Hardware
• 16 cores/node – 2 Intel Sandy Bridge-EP 2.6 Ghz
CPUs
• 32 GB main memory, 20 MB L3 shared cache
• 1 PCIe Gen3 (128 Gbps)
Performance Results
• Network Level Performance
• Latency
• MPI Level Performace
• Point to point MPI
• MPI Collectives
• NAS Parallel Benchmarks
• Cloud Computing Middlewares
• HDFS write using TestDFSIO
Latency
Point to point MPI : Latency
Point to point MPI: Bandwidth
MPI Collective : Scatter
NAS Parallel Benchmarks Class C
HDFS Write Operation
Conclusion
• Carried out a comprehensive performance
evaluation of four possible modes of
communication
• Latest InfiniBand FDR interconnect gives the best
performance
• Network level evaluations and for HPC
applications: RoCE 40 GigE performance better
than IB QDR
• Cloud computing middleware: IPoIB QDR
performance better than RoCE 40 GigE