Beihang University, China

Slides:



Advertisements
Similar presentations
Automated Test Data Generation Maili Markvardt. Outline Introduction Test data generation problem Black-box approach White-box approach.
Advertisements

Learning Rules from System Call Arguments and Sequences for Anomaly Detection Gaurav Tandon and Philip Chan Department of Computer Sciences Florida Institute.
Automatic and Scalable Fault Detection for Mobile Applications MobiSys’ 14 Presented by Haocheng Huang
Hybrid Concolic Testing Rupak Majumdar Koushik Sen UC Los Angeles UC Berkeley.
An Experimental Evaluation on Reliability Features of N-Version Programming Xia Cai, Michael R. Lyu and Mladen A. Vouk ISSRE’2005.
Delta Debugging - Demo Presented by: Xia Cheng. Motivation Automation is difficult Automation is difficult fail analysis needs complete understanding.
1 Indirect Adaptive Routing on Large Scale Interconnection Networks Nan Jiang, William J. Dally Computer System Laboratory Stanford University John Kim.
1 Dot Plots For Time Series Analysis Dragomir Yankov, Eamonn Keogh, Stefano Lonardi Dept. of Computer Science & Eng. University of California Riverside.
AUTOMATIC CONCOLIC TEST GENERATION WITH VIRTUAL PROTOTYPES FOR POST-SILICON VALIDATION Reviewer: Shin-Yann Ho Instructor: Jie-Hong Jiang.
XFindBugs: eXtended FindBugs for AspectJ Haihao Shen, Sai Zhang, Jianjun Zhao, Jianhong Fang, Shiyuan Yao Software Theory and Practice Group (STAP) Shanghai.
Approximating the Algebraic Solution of Systems of Interval Linear Equations with Use of Neural Networks Nguyen Hoang Viet Michal Kleiber Institute of.
Dynodroid: An Input Generation System for Android Apps
AMOST Experimental Comparison of Code-Based and Model-Based Test Prioritization Bogdan Korel Computer Science Department Illinois Institute of Technology.
The Software Engineering Group Prof. T.H. Tse (PhD LSE) Research Program testing, debugging, and analysis Application areas include object-oriented software,
Improving Effectiveness of Regression Testing of Telecommunications Systems Software Sami Torniainen Supervisor: Professor Raimo Kantola.
“Simulation Approach for Impact Analysis of Requirement Volatility Considering Dependency Change” Tiemen Roest.
Evaluating Impact of Storage on Smartphone Energy Efficiency David T. Nguyen.
Change Impact Analysis for AspectJ Programs Sai Zhang, Zhongxian Gu, Yu Lin and Jianjun Zhao Shanghai Jiao Tong University.
Flow-Augmented Call Graph: A New Foundation for Taming API Complexity Qirun Zhang
BFTCloud: A Byzantine Fault Tolerance Framework for Voluntary-Resource Cloud Computing Yilei Zhang, Zibin Zheng, and Michael R. Lyu
University of Central Florida TaintDroid: An Information-Flow Tracking System for Realtime Privacy Monitoring on Smartphones Written by Enck, Gilbert,
PUMA: Programmable UI- Automation for Large Scale Dynamic Analysis of Mobile Apps MobiSys’ 14 Presented by Haocheng Huang
On Distinguishing the Multiple Radio Paths in RSS-based Ranging Dian Zhang, Yunhuai Liu, Xiaonan Guo, Min Gao and Lionel M. Ni College of Software, Shenzhen.
From Design to Production Practicing what we preach at HP Shane Evans – Product Manager Oded Keret – Functional Architect.
Automatically Repairing Broken Workflows for Evolving GUI Applications Sai Zhang University of Washington Joint work with: Hao Lü, Michael D. Ernst.
Examining the Landscape and Impact of Android App Plagiarism Clint Gibler Ryan Stevens Jonathan Crussell Hao Chen Hui Zang Heesook Choi 1.
Extended Finite-State Machine Inference with Parallel Ant Colony Based Algorithms PPSN’14 September 13, 2014 Daniil Chivilikhin PhD student ITMO.
Computer Science and Engineering TreeSpan Efficiently Computing Similarity All-Matching Gaoping Zhu #, Xuemin Lin #, Ke Zhu #, Wenjie Zhang #, Jeffrey.
Comparing model-based and dynamic event-extraction based GUI testing techniques : An empirical study Gigon Bae, Gregg Rothermel, Doo-Hwan Bae The Journal.
11 Online Computing and Predicting Architectural Vulnerability Factor of Microprocessor Structures Songjun Pan Yu Hu Xiaowei Li {pansongjun, huyu,
Recommender Systems with Social Regularization Hao Ma, Dengyong Zhou, Chao Liu Microsoft Research Michael R. Lyu The Chinese University of Hong Kong Irwin.
Symbolic Model Checking of Software Nishant Sinha with Edmund Clarke, Flavio Lerda, Michael Theobald Carnegie Mellon University.
SOFTWARE TESTING AND QUALITY ASSURANCE. Software Testing.
Hao Ma, Dengyong Zhou, Chao Liu Microsoft Research Michael R. Lyu
AGENT SERVICES FRAMEWORK FOR HANDHELD DEVICES SUPERVISOR: DR MUHAMMAD ASLAM Domain: Artificial Intelligence(AI) Muhammad Ali (09-MS-CE-11),
Experience Report: System Log Analysis for Anomaly Detection
Appium Studio Appium testing made easy at any scale.
Internet Quarantine: Requirements for Containing Self-Propagating Code
More Security and Programming Language Work on SmartPhones
Regression Testing with its types
Jacob R. Lorch Microsoft Research
Distributed Network Traffic Feature Extraction for a Real-time IDS
Spatial Discovery in 60 GHz
Daniil Chivilikhin and Vladimir Ulyantsev
Harvesting Runtime Values in Android Applications That Feature Anti-Analysis Techniques Presented by Vikraman Mohan.
Towards Trustworthy Program Repair
Discriminative Training of Chow-Liu tree Multinet Classifiers
Automated Pattern Based Mobile Testing
Mid-term Exam Account for 20% of the grade 100 points in total
Chapter 10: Analysis of Variance: Comparing More Than Two Means
SENIOR MANAGER - SOFTWARE TESTING PRACTICE
TriggerScope Towards detecting logic bombs in android applications
A Framework for Automatic Resource and Accuracy Management in A Cloud Environment Smita Vijayakumar.
Collaborative Offloading for Distributed Mobile-Cloud Apps
Sentio: Distributed Sensor Virtualization for Mobile Apps
Pankaj Kumar, Tech Lead Bhuvaneswari Radhakrishnan, Senior Engineer
Speaker: Bo Jiang* Co-authors: Yuxuan Wu*, Teng Li*, W.K. Chan†
Vlad Nae, Radu Prodan, Thomas Fahringer Institute of Computer Science
Bo Wang1, Yingfei Xiong2, Zhenjiang Hu3, Haiyan Zhao1,
Network Profiler: Towards Automatic Fingerprinting of Android Apps
Agile Testing Using Virtualization
Comparison between conventional wide-field detection and multifocus detection. Comparison between conventional wide-field detection and multifocus detection.
Smita Vijayakumar Qian Zhu Gagan Agrawal
Tracking users for a targeted dissemination
Personalized Celebrity Video Search Based on Cross-space Mining
کتابهای تازه خریداری شده دروس عمومی 1397
Institute of Computing Tech.
Off-path Leakage Power Aware Routing for SRAM-based FPGAs
Spatial Discovery in 60 GHz
Mitigating the Effects of Flaky Tests on Mutation Testing
Presentation transcript:

Beihang University, China QRS 2017 Which Factor Impacts GUI Traversal-Based Test Case Generation Technique Most? — A Controlled Experiment on Android Applications Bo Jiang & Yaoyue Zhang, Beihang University W.K. Chan, City Univeristy of Hong Kong Zhenyu Zhang, Chinese Academy of Sciences, Institute of Software

1 2 3 4 5 Outline DESIGN FACTORS BACKGROUND & RELATED WORK TEST CASE GENERATION FRAMEWORK 3 DESIGN FACTORS 4 CONTROLLED EXPERIMENT 5 CONCLUSION

1 2 3 4 5 Outline DESIGN FACTORS BACKGROUND & RELATED WORK TEST CASE GENERATION FRAMEWORK 3 DESIGN FACTORS 4 CONTROLLED EXPERIMENT 5 CONCLUSION

Background 4 1 Android Applications requires testing for ensure quality Automated test case generation techniques are major research focus. 2 3 StateTraversal: one of the most popular type of techniques.

Related Work 5 1 Automated Test Input Generation for Android: Are We There Yet? R. C. Shauvik et al. (ASE2015) SwiftHand: Guided GUI Testing of Android Apps with Minimal Restart and Approximate Learning. W. Choi et al. (OOPSLA2013) 2 PUMA: Programmable UI-automation for large-scale dynamic analysis of mobile apps. S. Hao, B. Liu et al. (MobiSys2014) 3 Acteve: Automated Concolic Testing of Smartphone Apps. S. Anand et al. (FSE2012) 4

1 2 3 4 5 Outline DESIGN FACTORS BACKGROUND & RELATED WORK TEST CASE GENERATION FRAMEWORK 3 DESIGN FACTORS 4 CONTROLLED EXPERIMENT 5 CONCLUSION

Test Case Generation Framework 7 PUMA Framework

Test Case Generation Framework 8 Generic Framework Generic GUI Traversal-based Test Case generation framework of PUMA 01 State equivalence 02 Search strategy 03 Waiting time

1 2 3 4 5 Outline BACKGROUND & RELATED WORK TEST CASE GENERATION FRAMEWORK 3 DESIGN FACTORS 4 CONTROLLED EXPERIMENT 5 CONCLUSION

02 01 03 01 Design Factors Cosine ActivityID UI Hierarchy 10 01 State Equivalence 02 Cosine Eigenvectors of the UI widgets. Threshold: 0.95 DECAF&PUMA ActivityID UIAutomator's API getCurrentActivityName(). String comparison. A3E UI Hierarchy Use Widgets tree structure to represent . The widgets trees are the same. SwiftHand 01 03

Design Factors 11 02 Search Strategy BFS PUMA DFS A3E Rand monkey

03 Design Factors Waiting Time (between Two Events) Strategy Used By 11 03 Waiting Time (between Two Events) Strategy Used By watiForIdle PUMA wait200ms Monkey used by Shauvik et al in ASE 2015 wait3000ms Acteve wait5000ms Swifthand

Three Factors and Their Levles Design Factors 12 Three Factors and Their Levles Factor Level Factor 1: State Equivalence Factor 2: Search Strategy Factor 3: Waiting Time Cosine BFS waitForIdle 1 UI Hierarchy DFS wait200ms 2 ActivityID Random wait3000ms 3 — wait5000ms

1 2 3 4 5 Outline DESIGN FACTORS BACKGROUND & RELATED WORK TEST CASE GENERATION FRAMEWORK 3 DESIGN FACTORS 4 CONTROLLED EXPERIMENT 5 CONCLUSION

Controlled Experiment 14 01 Benchmarks and Experimental Setup 33 real-world open-source mobile apps from Dynodroid, A3E, ACTEve, SwiftHand. We implemented all factor levels in the PUMA framework. Two virtual machines installed with ubuntu 14.04 operating systems.

Controlled Experiment 15 02 Experimental Procedure 36 (i.e., 3*3*4) combinations of factor levels for the three factors. Took 1188 testing hours in total on 2 virtual machines. ANOVAs(one-way ANalyses Of VAriances) Multiple comparison

Controlled Experiment 16 03 Results and Analysis-state equivalence

Controlled Experiment 17 03 Results and Analysis-state equivalence

Controlled Experiment 18 03 Results and Analysis-state equivalence Cosine Similarity > UI Hierarchy > ActivityID in Failure detection ability Code coverage rate

Controlled Experiment 19 03 Results and Analysis-search strategy

Controlled Experiment 20 03 Results and Analysis-search strategy

Controlled Experiment 21 03 Results and Analysis-search strategy Randomized search strategy was statistically comparable to BFS and DFS in Failure detection ability Code coverage rate

Controlled Experiment 22 03 Results and Analysis-waiting time

Controlled Experiment 23 03 Results and Analysis-waiting time

Controlled Experiment 25 03 Results and Analysis-waiting time The strategy to wait until GUI state is stable before sending the next input event is not statistically more effective than the strategy of waiting for a fixed time interval in Failure detection ability Code coverage rate

Controlled Experiment 25 03 Results and Analysis-Best Treatment-Failure Detection Rate

Controlled Experiment 26 03 Results and Analysis-Best Treatment-Code Coverage

Controlled Experiment 27 03 Results and Analysis-Best Treatment There were many combinations of factor levels can attain the same high level of failure detection rate and high level of statement code coverage There could be many good configurations in configuring StateTraversal.

1 2 3 4 5 Outline DESIGN FACTORS BACKGROUND & RELATED WORK TEST CASE GENERATION FRAMEWORK 3 DESIGN FACTORS 4 CONTROLLED EXPERIMENT 5 CONCLUSION

Conclusion 29 State equivalence :Different state equivalence definitionswill significantly affect the failure detection rates and the code coverage. Search strategy: BFS and DFS are comparable to Random. Waiting time: Waiting for idle and waiting for a fixed time period have no significant difference. Failure detection rate: <Cosine Similarity, BFS, wait5000ms>(best). Code coverage:<Cosine Similarity, DFS, wait5000ms>(best).

THANKS Q&A