Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Predicting and Explaining Mobile UI Tappability with Vision Modeling and Saliency Analysis (2204.02448v1)

Published 5 Apr 2022 in cs.HC, cs.AI, cs.CV, and cs.LG

Abstract: We use a deep learning based approach to predict whether a selected element in a mobile UI screenshot will be perceived by users as tappable, based on pixels only instead of view hierarchies required by previous work. To help designers better understand model predictions and to provide more actionable design feedback than predictions alone, we additionally use ML interpretability techniques to help explain the output of our model. We use XRAI to highlight areas in the input screenshot that most strongly influence the tappability prediction for the selected region, and use k-Nearest Neighbors to present the most similar mobile UIs from the dataset with opposing influences on tappability perception.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Eldon Schoop (10 papers)
  2. Xin Zhou (319 papers)
  3. Gang Li (579 papers)
  4. Zhourong Chen (12 papers)
  5. Björn Hartmann (9 papers)
  6. Yang Li (1140 papers)
Citations (26)