Datasets:
Tasks:
Object Detection
Modalities:
Image
Formats:
imagefolder
Languages:
English
Size:
1K - 10K
ArXiv:
File size: 9,355 Bytes
bf02f4d 2b83432 bf02f4d 2b83432 bf02f4d a080a11 bf02f4d 2b83432 bf02f4d 2b83432 bf02f4d a080a11 2b83432 bf02f4d 2b83432 bf02f4d 2b83432 bf02f4d 2b83432 bf02f4d 2b83432 bf02f4d 2b83432 bf02f4d 2b83432 bf02f4d 2b83432 bf02f4d 2b83432 bf02f4d 2b83432 bf02f4d 2b83432 bf02f4d 2b83432 bf02f4d 2b83432 bf02f4d 2b83432 bf02f4d 2b83432 bf02f4d 2b83432 bf02f4d 2b83432 bf02f4d 2b83432 bf02f4d 2b83432 bf02f4d 2b83432 bf02f4d 2b83432 bf02f4d 2b83432 bf02f4d 2b83432 bf02f4d 2b83432 bf02f4d 2b83432 bf02f4d 2b83432 bf02f4d 2b83432 bf02f4d 2b83432 bf02f4d 2b83432 bf02f4d 2b83432 bf02f4d 2b83432 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 |
---
annotations_creators: []
language: en
size_categories:
- 1K<n<10K
task_categories:
- object-detection
task_ids: []
pretty_name: guiact_smartphone_test
tags:
- fiftyone
- visual-agents
- os-agents
- gui-grounding
- image
- object-detection
dataset_summary: '
This is a [FiftyOne](https://github.com/voxel51/fiftyone) dataset with 2079 samples.
## Installation
If you haven''t already, install FiftyOne:
```bash
pip install -U fiftyone
```
## Usage
```python
import fiftyone as fo
from fiftyone.utils.huggingface import load_from_hub
# Load the dataset
# Note: other available arguments include ''max_samples'', etc
dataset = load_from_hub("Voxel51/guiact_smartphone_test")
# Launch the App
session = fo.launch_app(dataset)
```
'
---
# GUIAct Smartphone Dataset - Test Split
<!-- Provide a quick summary of the dataset. -->

This is a [FiftyOne](https://github.com/voxel51/fiftyone) dataset with 2079 samples.
## Installation
If you haven't already, install FiftyOne:
```bash
pip install -U fiftyone
```
## Usage
```python
import fiftyone as fo
from fiftyone.utils.huggingface import load_from_hub
# Load the dataset
# Note: other available arguments include 'max_samples', etc
dataset = load_from_hub("Voxel51/guiact_smartphone_test")
# Launch the App
session = fo.launch_app(dataset)
```
## Dataset Details
### Dataset Description
The GUIAct Smartphone dataset is a collection of multi-step action instructions for smartphone GUI navigation tasks. It was created by converting and adapting a subset of the AITW ([Androids in the Wild](https://arxiv.org/abs/2307.10088)) dataset, specifically focusing on data with the "General" tag. The dataset contains sequences of actions (such as tap, swipe, input) paired with smartphone screenshots, designed to train vision language models to navigate smartphone interfaces. Each instruction requires multiple actions to complete the task, with actions mapped to a standardized format that includes position information. The dataset serves as a crucial component of the larger GUICourse collection for training versatile GUI agents.
- **Curated by:** Wentong Chen, Junbo Cui, Jinyi Hu, and other researchers from Tsinghua University, Renmin University of China, and other institutions as listed in the GUICourse paper
- **Shared by:** The authors of the GUICourse paper
- **Language(s) (NLP):** en
- **License:** CC BY 4.0
### Dataset Sources
- **Repository:** https://github.com/yiye3/GUICourse and https://huggingface.co/datasets/yiye2023/GUIAct
- **Paper:** "GUICourse: From General Vision Language Model to Versatile GUI Agent" (arXiv:2406.11317v1)
## Uses
### Direct Use
The GUIAct Smartphone dataset is intended to be used for training and evaluating GUI agents that can navigate smartphone interfaces using visual inputs. Specific use cases include:
1. Training vision language models to understand smartphone interface elements and their functions
2. Teaching models to execute multi-step operations on smartphones based on natural language instructions
3. Developing assistive technologies that can help users navigate smartphone interfaces
4. Benchmarking the performance of GUI agents on smartphone navigation tasks
### Out-of-Scope Use
The dataset is not intended to:
1. Train models to access private user data on smartphones
2. Enable unauthorized access to smartphone systems
3. Replace human decision-making for critical smartphone operations
4. Train agents to perform tasks beyond the scope of normal smartphone interface interactions
5. Work with smartphone operating systems or applications not represented in the dataset
## Dataset Structure
The GUIAct Smartphone dataset contains 9,157 multi-step action instructions which correspond to approximately 67,000 training samples. Each sample consists of:
1. A smartphone screenshot (image)
2. One or more actions to be performed on that screenshot
3. Natural language instruction describing the task to be accomplished
The action space includes standardized actions such as:
- **tap**: Single-point touch on the screen with position coordinates
- **swipe**: Multi-point gesture with start and end coordinates
- **input**: Text entry with content specified
- **enter**: Submission action
- **answer**: Task completion indication with text output
Actions include position information represented in either absolute pixel coordinates or relative position format (normalized to a range of 0-1 or 0-1000).
## FiftyOne Dataset Structure
# GUIAct Smartphone Test Dataset Structure
**Basic Info:** 2,079 mobile UI screenshots with interaction annotations
**Core Fields:**
- `id`: ObjectIdField - Unique identifier
- `filepath`: StringField - Image path
- `tags`: ListField(StringField) - Sample categories
- `metadata`: EmbeddedDocumentField - Image properties (size, dimensions)
- `episode`: StringField - Unique identifier for interaction sequence
- `step`: IntField - Sequential position within episode
- `question`: StringField - Natural language task description
- `current_action`: StringField - Current action to perform
- `structured_history`: ListField - Previous actions in sequence
- `ui_elements`: EmbeddedDocumentField(Detections) containing multiple Detection objects:
- `label`: Element type (ICON_*, TEXT)
- `bounding_box`: a list of relative bounding box coordinates in [0, 1] in the following format: `[<top-left-x>, <top-left-y>, <width>, <height>]`
- `text`: Text content of element if present
- `action`: EmbeddedDocumentField(Keypoints) containing interaction points:
- `label`: Action type (tap, swipe)
- `points`: A list of `(x, y)` keypoints in `[0, 1] x [0, 1]`, `[[x, y]]` for tap or `[[x1, y1], [x2, y2]]` for swipe
The dataset captures smartphone screen interactions with detailed UI element annotations and action sequences for mobile interaction research.
## Dataset Creation
### Curation Rationale
The GUIAct Smartphone dataset was created to provide training data for visual-based GUI agents that can navigate smartphone interfaces. The authors identified limitations in existing datasets for training such agents, including scenarios that were too simplistic, too narrow in domain focus, or too small in size. By adapting data from the larger AITW dataset, they aimed to provide a substantial collection of realistic smartphone navigation tasks for training robust GUI agents.
### Source Data
#### Data Collection and Processing
The GUIAct Smartphone dataset was derived from the AITW dataset by:
1. Selecting data with the "General" tag from AITW
2. Filtering out screenshots that did not include the bottom navigation bar
3. Converting the original AITW action format to the unified action space defined for GUICourse:
- "dual-point gesture" actions were split into "tap" and "swipe" based on distance
- "type" actions were renamed to "input" actions
- "enter" actions were preserved
- "go back/home" actions were converted to "tap" actions on the bottom navigation bar
- "task complete/impossible" actions were converted to "answer" actions
#### Personal and Sensitive Information
The paper does not explicitly address whether the smartphone screenshots contain personal or sensitive information. However, since the dataset is derived from AITW and intended for research purposes, it likely avoids including personally identifiable information in the screenshots.
## Bias, Risks, and Limitations
- The dataset may reflect biases present in smartphone interface design
- The tasks represented may not cover the full diversity of smartphone usage patterns across different user demographics
- The dataset is likely limited to specific smartphone operating systems and applications
- Performance of models trained on this data may not generalize to significantly different smartphone interfaces or to future interface designs
- The action space simplification (e.g., converting dual-point gestures to tap/swipe) may not capture the full complexity of human smartphone interaction
### Recommendations
Users should be aware that:
- Models trained on this dataset may perform differently across different smartphone operating systems or UI designs
- The simplified action space may limit the range of interactions possible with trained models
- Evaluation should consider both action accuracy and overall task completion
- For deployment in assistive technologies, additional safety measures and user feedback mechanisms are advisable
- The dataset should be used as part of a broader training approach that includes ethical considerations for AI assistance
## Citation
**BibTeX:**
```bibtex
@article{chen2024guicourse,
title={GUICourse: From General Vision Language Model to Versatile GUI Agent},
author={Chen, Wentong and Cui, Junbo and Hu, Jinyi and Qin, Yujia and Fang, Junjie and Zhao, Yue and Wang, Chongyi and Liu, Jun and Chen, Guirong and Huo, Yupeng and Yao, Yuan and Lin, Yankai and Liu, Zhiyuan and Sun, Maosong},
journal={arXiv preprint arXiv:2406.11317},
year={2024}
}
```
**APA:**
Chen, W., Cui, J., Hu, J., Qin, Y., Fang, J., Zhao, Y., Wang, C., Liu, J., Chen, G., Huo, Y., Yao, Y., Lin, Y., Liu, Z., & Sun, M. (2024). GUICourse: From General Vision Language Model to Versatile GUI Agent. arXiv preprint arXiv:2406.11317.
|