Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

BigDataBench: a Big Data Benchmark Suite from Internet Services (1401.1406v2)

Published 6 Jan 2014 in cs.DB

Abstract: As architecture, systems, and data management communities pay greater attention to innovative big data systems and architectures, the pressure of benchmarking and evaluating these systems rises. Considering the broad use of big data systems, big data benchmarks must include diversity of data and workloads. Most of the state-of-the-art big data benchmarking efforts target evaluating specific types of applications or system software stacks, and hence they are not qualified for serving the purposes mentioned above. This paper presents our joint research efforts on this issue with several industrial partners. Our big data benchmark suite BigDataBench not only covers broad application scenarios, but also includes diverse and representative data sets. BigDataBench is publicly available from http://prof.ict.ac.cn/BigDataBench . Also, we comprehensively characterize 19 big data workloads included in BigDataBench with varying data inputs. On a typical state-of-practice processor, Intel Xeon E5645, we have the following observations: First, in comparison with the traditional benchmarks: including PARSEC, HPCC, and SPECCPU, big data applications have very low operation intensity; Second, the volume of data input has non-negligible impact on micro-architecture characteristics, which may impose challenges for simulation-based big data architecture research; Last but not least, corroborating the observations in CloudSuite and DCBench (which use smaller data inputs), we find that the numbers of L1 instruction cache misses per 1000 instructions of the big data applications are higher than in the traditional benchmarks; also, we find that L3 caches are effective for the big data applications, corroborating the observation in DCBench.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (15)
  1. Lei Wang (975 papers)
  2. Jianfeng Zhan (92 papers)
  3. Chunjie Luo (39 papers)
  4. Yuqing Zhu (34 papers)
  5. Qiang Yang (202 papers)
  6. Yongqiang He (8 papers)
  7. Wanling Gao (47 papers)
  8. Zhen Jia (34 papers)
  9. Yingjie Shi (8 papers)
  10. Shujie Zhang (6 papers)
  11. Chen Zheng (52 papers)
  12. Gang Lu (21 papers)
  13. Kent Zhan (1 paper)
  14. Xiaona Li (3 papers)
  15. Bizhu Qiu (2 papers)
Citations (583)

Summary

BigDataBench: A Comprehensive Big Data Benchmark Suite

The paper "BigDataBench: a Big Data Benchmark Suite from Internet Services" addresses the growing demand in the architecture, systems, and data management communities for evaluating big data systems. As big data applications evolve rapidly and become more diverse, the complexities in benchmarking these systems also increase. This paper presents BigDataBench, a comprehensive benchmarking suite designed to provide fair and inclusive evaluations.

BigDataBench differentiates itself from prior efforts by covering a broad spectrum of application scenarios, data types, and workloads. The suite includes 19 benchmarks extracted from dimensions such as application scenarios, operations, algorithms, data types, data sources, software stacks, and application types, ensuring a holistic approach to evaluating big data systems. Unlike other benchmarks like HiBench, CloudSuite, and DCBench, BigDataBench offers a broader coverage by integrating diverse real-world data sets and workloads. These efforts provide a crucial foundation for consistently measuring and comparing the performance of big data systems.

Key Observations

On testing with typical state-of-practice processors like the Intel Xeon E5645, three significant observations are made:

  1. Operation Intensity: Big data applications exhibit low operation intensity compared to traditional benchmarks such as PARSEC and HPCC. This characteristic is attributed to their high ratio of memory access relative to instruction execution, highlighting different computational demand patterns.
  2. Impact of Data Volume: The volume of input data markedly affects micro-architectural characteristics. As data size increases, the performance characteristics such as instruction throughput and cache behaviors vary significantly. For example, variations in input size resulted in a 2.9 times difference in Million Instructions Per Second (MIPS) for the Grep workload.
  3. Cache Efficiency: The effectiveness of L3 caches is notable in big data workloads, unlike traditional benchmarks that might not leverage them as effectively. Interestingly, big data workloads show higher L1 instruction cache (L1I) misses compared to traditional workloads, suggesting a demand for improved front-end performance.

Implications and Future Directions

The findings suggest that big data systems require different architectural considerations due to their unique operational characteristics, including high data movement demands and varying cache efficiencies. These factors impact both theoretical explorations and practical implementations in the field.

BigDataBench’s methodology embracing real-world data and diverse workloads sets a precedent for future developments in AI and data-centric architecture research. The researchers acknowledge the need for extensibility and anticipate incorporating emerging techniques and data types into the suite, thereby ensuring ongoing relevance in an evolving landscape.

The BigDataBench initiative provides both breadth and depth in benchmarking, which is critical for the advancements in optimizing and building next-generation big data systems. The comprehensive suite aims to guide hardware and software design by accurately reflecting the realities of contemporary data workloads. Continuing to develop and adapt benchmarks like BigDataBench will remain vital as the big data paradigm continues to expand and evolve.