AURORA: Navigating UI Tarpits via Automated Neural Screen Understanding
arxiv(2024)
摘要
Nearly a decade of research in software engineering has focused on automating
mobile app testing to help engineers in overcoming the unique challenges
associated with the software platform. Much of this work has come in the form
of Automated Input Generation tools (AIG tools) that dynamically explore app
screens. However, such tools have repeatedly been demonstrated to achieve
lower-than-expected code coverage - particularly on sophisticated proprietary
apps. Prior work has illustrated that a primary cause of these coverage
deficiencies is related to so-called tarpits, or complex screens that are
difficult to navigate.
In this paper, we take a critical step toward enabling AIG tools to
effectively navigate tarpits during app exploration through a new form of
automated semantic screen understanding. We introduce AURORA, a technique that
learns from the visual and textual patterns that exist in mobile app UIs to
automatically detect common screen designs and navigate them accordingly. The
key idea of AURORA is that there are a finite number of mobile app screen
designs, albeit with subtle variations, such that the general patterns of
different categories of UI designs can be learned. As such, AURORA employs a
multi-modal, neural screen classifier that is able to recognize the most common
types of UI screen designs. After recognizing a given screen, it then applies a
set of flexible and generalizable heuristics to properly navigate the screen.
We evaluated AURORA both on a set of 12 apps with known tarpits from prior
work, and on a new set of five of the most popular apps from the Google Play
store. Our results indicate that AURORA is able to effectively navigate tarpit
screens, outperforming prior approaches that avoid tarpits by 19.6
method coverage. The improvements can be attributed to AURORA's UI design
classification and heuristic navigation techniques.
更多查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要