Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A deep tree-based model for software defect prediction (1802.00921v1)

Published 3 Feb 2018 in cs.SE

Abstract: Defects are common in software systems and can potentially cause various problems to software users. Different methods have been developed to quickly predict the most likely locations of defects in large code bases. Most of them focus on designing features (e.g. complexity metrics) that correlate with potentially defective code. Those approaches however do not sufficiently capture the syntax and different levels of semantics of source code, an important capability for building accurate prediction models. In this paper, we develop a novel prediction model which is capable of automatically learning features for representing source code and using them for defect prediction. Our prediction system is built upon the powerful deep learning, tree-structured Long Short Term Memory network which directly matches with the Abstract Syntax Tree representation of source code. An evaluation on two datasets, one from open source projects contributed by Samsung and the other from the public PROMISE repository, demonstrates the effectiveness of our approach for both within-project and cross-project predictions.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Hoa Khanh Dam (17 papers)
  2. Trang Pham (17 papers)
  3. Shien Wee Ng (2 papers)
  4. Truyen Tran (112 papers)
  5. John Grundy (127 papers)
  6. Aditya Ghose (22 papers)
  7. Taeksu Kim (1 paper)
  8. Chul-Joo Kim (2 papers)
Citations (107)

Summary

We haven't generated a summary for this paper yet.