Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Context-Aware Parse Trees (2003.11118v1)

Published 24 Mar 2020 in cs.PL and cs.AI

Abstract: The simplified parse tree (SPT) presented in Aroma, a state-of-the-art code recommendation system, is a tree-structured representation used to infer code semantics by capturing program \emph{structure} rather than program \emph{syntax}. This is a departure from the classical abstract syntax tree, which is principally driven by programming language syntax. While we believe a semantics-driven representation is desirable, the specifics of an SPT's construction can impact its performance. We analyze these nuances and present a new tree structure, heavily influenced by Aroma's SPT, called a \emph{context-aware parse tree} (CAPT). CAPT enhances SPT by providing a richer level of semantic representation. Specifically, CAPT provides additional binding support for language-specific techniques for adding semantically-salient features, and language-agnostic techniques for removing syntactically-present but semantically-irrelevant features. Our research quantitatively demonstrates the value of our proposed semantically-salient features, enabling a specific CAPT configuration to be 39\% more accurate than SPT across the 48,610 programs we analyzed.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (11)
  1. Fangke Ye (3 papers)
  2. Shengtian Zhou (8 papers)
  3. Anand Venkat (5 papers)
  4. Ryan Marcus (33 papers)
  5. Paul Petersen (2 papers)
  6. Jesmin Jahan Tithi (19 papers)
  7. Tim Mattson (10 papers)
  8. Tim Kraska (78 papers)
  9. Pradeep Dubey (31 papers)
  10. Vivek Sarkar (16 papers)
  11. Justin Gottschlich (21 papers)
Citations (2)

Summary

We haven't generated a summary for this paper yet.