GitHub - ai-agents-2030/SPA-Bench: SPA-Bench: A Comprehensive Benchmark for SmartPhone Agent Evaluation (original) (raw)
πΏπͺ SPA-Bench: A Comprehensive Benchmark for Smartphone Agent Evaluation
π Welcome to the SPA-Bench repository, a benchmark designed to evaluate the performance of smartphone agents. This project offers a structured approach to assessing the efficiency, robustness, and accuracy of various smartphone agents across a variety of scenarios and conditions.
π’ News
- [11 Feb '25] SPA-Bench has been accepted to ICLR 2025 and selected as a Spotlight! Congrats to all co-authorsπ See you in SingaporeπΈπ¬
- [02 Dec '24] We have partially released the core code, including AppAgent integration. The full version will be made available later this month. Stay tuned for updates!
- [11 Oct '24] SPA-Bench has been accepted by the NeurIPS 2024 Workshop Open-World Agents!
β© Quick Start
π οΈ Installation
git clone --recurse-submodules https://github.com/ai-agents-2030/SPA-Bench.git
π Documentation
π‘ About SPA-Bench
SPA-Bench provides a thorough evaluation framework for smartphone agents, covering key metrics and test scenarios that reflect real-world usage patterns and challenges. This benchmark supplies essential tools and datasets to support consistent evaluation of agent performance across a wide range of tasks and applications.
π¬ Core Features
π Diverse and Realistic Task Design
- π¦ 340 Tasks - 300 Single-app Tasks and 40 Cross-app Tasks
- π 66 Apps β 52 Third-party Apps, 7 Google Apps and 7 System Apps
- π 2 Languages β Chinese and English apps
- π Increased Difficulty Levels
- π¨ Human-Annotated Trajectories & Key Components
π€ Plug-and-Play Agent Framework
- π§ 11 Smartphone Agents Ready for Evaluation
- π§© Easy Integration of Your Own Agents with Minimal Code Changes
- π± Scalable Design β Multi-device support & Emulator Compatibility
- πΈ Android Snapshot β Local Environment Setup and Data Reset for Consistent Testing
β Automatic and Scalable Evaluation Pipeline
- π 7 Evaluation Metrics for a Comprehensive Analysis
- π Coarse-and-Fine Success Detection β Requires No Further Human Effort
- π Trajectory Splitting & Subtask Evaluation β Tailored for Long-Sequence Tasks
- π Performance Metrics:
- Single-app Tasks β F1-scores: 0.926 (English), 0.884 (Chinese)
- Cross-app Tasks β F1-scores: 0.833 (English), 0.857 (Chinese)
π Coming Soon
- Full Agent Integrations
- Snapshot for Android Emulator
- Task Collection
- Agent Framework
- Evaluation Pipeline
π Citation
@inproceedings{chen2025spabench, title={SPA-Bench: A Comprehensive Benchmark for SmartPhone Agent Evaluation}, author={Jingxuan Chen and Derek Yuen and Bin Xie and Yuhao Yang and Gongwei Chen and Zhihao Wu and Li Yixing and Xurui Zhou and Weiwen Liu and Shuai Wang and Kaiwen Zhou and Rui Shao and Liqiang Nie and Yasheng Wang and Jianye HAO and Jun Wang and Kun Shao}, booktitle={The Thirteenth International Conference on Learning Representations}, year={2025}, }