Research Artifacts in Software Engineering Publications: Status and Trends (2404.06852v1)
Abstract: The Software Engineering (SE) community has been embracing the open science policy and encouraging researchers to disclose artifacts in their publications. However, the status and trends of artifact practice and quality remain unclear, lacking insights on further improvement. In this paper, we present an empirical study to characterize the research artifacts in SE publications. Specifically, we manually collect 1,487 artifacts from all 2,196 papers published in top-tier SE conferences (ASE, FSE, ICSE, and ISSTA) from 2017 to 2022. We investigate the common practices (e.g., URL location and format, storage websites), maintenance activities (e.g., last update time and URL validity), popularity (e.g., the number of stars on GitHub and characteristics), and quality (e.g., documentation and code smell) of these artifacts. Based on our analysis, we reveal a rise in publications providing artifacts. The usage of Zenodo for sharing artifacts has significantly increased. However, artifacts stored in GitHub tend to receive few stars, indicating a limited influence on real-world SE applications. We summarize the results and provide suggestions to different stakeholders in conjunction with current guidelines.
- Software artifact mining in software engineering conferences: A meta-analysis, in: Proceedings of the 16th ACM/IEEE International Symposium on Empirical Software Engineering and Measurement, pp. 227–237.
- ACM, 2020. Artifact review and badging - version 1.1. https://www.acm.org/publications/policies/artifact-review-and-badging-current.
- Software documentation issues unveiled, in: Proceedings of 2019 IEEE/ACM 41st International Conference on Software Engineering (ICSE19), pp. 1199–1210.
- Resources for reproducibility of experiments in empirical software engineering: Topics derived from a secondary study. IEEE Access 8, 8992–9004.
- Wayback machine. https://archive.org/web/.
- 1,500 scientists lift the lid on reproducibility. Nature 533.
- CSRankings, 2023. Computer science rankings. https://csrankings.org.
- Urls link rot: implications for electronic publishing. World Digital Libraries-An International Journal 8, 59–66.
- DBLP, 2023. computer science bibliography. https://dblp.org/.
- Software heritage: Why and how to preserve software source code, in: iPRES 2017-14th International Conference on Digital Preservation, pp. 1–10.
- What Makes a Popular Academic AI Repository? Empirical Software Engineering 26.
- A large-scale study of the evolution of web pages, in: Proceedings of the Twelfth International World Wide Web Conference (WWW03), pp. 669–678.
- A survey on artifacts from conext, icn, imc, and sigcomm conferences in 2017. ACM SIGCOMM Computer Communication Review 48, 75–80.
- FSE21, 2021a. Esec/fse 2021 open science guidelines. https://2021.esec-fse.org/attending/Open_Science_Guidelines.
- FSE21, 2021b. Fse 2021 artifact evaluation track:submission and reviewing guidelines. https://2021.esec-fse.org/getImage/orig/fse_artifacts_submission_reviewing_guidelines.pdf.
- FSE22, 2022. Fse 2022 open science policy. https://2022.esec-fse.org/track/fse-2022-research-papers#FAQ.
- Community expectations for research artifacts and evaluation processes, in: Proceedings of the 28th ACM Joint Meeting on European Software Engineering Conference and Symposium on the Foundations of Software Engineering (FSE20), pp. 469–480.
- ICSE21, 2020. Icse 2021 open science policies. https://conf.researchr.org/track/icse-2021/icse-2021-open-science-policies.
- ICSE22, 2021a. Icse 2022 open science policies. https://conf.researchr.org/track/icse-2022/icse-2022-open-science-policies.
- ICSE22, 2021b. Icse 2022 review process and guidelines. https://conf.researchr.org/getImage/icse-2022/orig/ICSE+2022+Review+Process+and+Guidelines-2.pdf.
- ICSE23, 2022. Icse 2023 open science policy. https://conf.researchr.org/track/icse-2023/icse-2023-technical-track#ICSE-2023-open-science-policy.
- Robustifying links to combat reference rot. Code4Lib Journal 50.
- The general index of software engineering papers, in: Proceedings of the 19th International Conference on Mining Software Repositories, pp. 98–102.
- Precise concolic unit testing of c programs using extended units and symbolic alarm filtering, in: Proceedings of the 40th International Conference on Software Engineering (ICSE18), pp. 315–326.
- Scholarly context not found: one in five articles suffers from reference rot. PloS one 9.
- Practitioners’ expectations on automated fault localization, in: Proceedings of the 25th International Symposium on Software Testing and Analysis (ISSTA16), pp. 165–176.
- Artifact evaluation for software conferences. ACM SIGSOFT Software Engineering Notes 38, 7–10.
- The real software crisis: Repeatability as a core value. Communications of the ACM 58, 34–36.
- On the Reproducibility and Replicability of Deep Learning in Software Engineering. ACM Transactions on Software Engineering and Methodology 31, 1–46.
- PEP8, 2013. Pep8 – style guide for python code. https://peps.python.org/pep-0008/.
- PMD, 2022. Pmd. https://pmd.github.io/.
- Pylint, 2022. Pylint. https://github.com/PyCQA/pylint.
- Pylint, 2023. Pylint messages overview. https://pylint.pycqa.org/en/latest/user_guide/messages/messages_overview.html.
- Analyzing the persistence of referenced web resources with memento. arXiv preprint arXiv:1105.3459 .
- A systematic review on the code smell effect. Journal of Systems and Software 144, 450–477.
- Evaluating the artifacts of sigcomm papers. ACM SIGCOMM Computer Communication Review 49, 44–47.
- Understanding and improving artifact sharing in software engineering research. Empirical Softw. Eng. 26.
- Software engineering and r programming: A call for research. R J. 13, 600.
- Mugeng Liu (4 papers)
- Xiaolong Huang (29 papers)
- Wei He (188 papers)
- Yibing Xie (1 paper)
- Jie M. Zhang (39 papers)
- Xiang Jing (1 paper)
- Zhenpeng Chen (39 papers)
- Yun Ma (38 papers)