harpreetsahota's picture
Update README.md
652067e verified
---
annotations_creators: []
language: en
size_categories:
- 1K<n<10K
task_categories:
- image-classification
- image-segmentation
task_ids: []
pretty_name: AVM_Segmentation
tags:
- fiftyone
- image
- image-classification
- image-segmentation
dataset_summary: '
This is a [FiftyOne](https://github.com/voxel51/fiftyone) dataset with 6763 samples.
## Installation
If you haven''t already, install FiftyOne:
```bash
pip install -U fiftyone
```
## Usage
```python
import fiftyone as fo
from fiftyone.utils.huggingface import load_from_hub
# Load the dataset
# Note: other available arguments include ''max_samples'', etc
dataset = load_from_hub("harpreetsahota/AVM_Segmentation_train")
# Launch the App
session = fo.launch_app(dataset)
```
'
---
# Dataset Card for AVM_Segmentation# Dataset Card for AVM (Around View Monitoring) Semantic Segmentation Dataset
![image/png](avm_segmentation-mq.gif)
This repository provides a FiftyOne-compatible version of the AVM semantic segmentation dataset for autonomous parking systems, with enhanced metadata and visualization capabilities.
This is a [FiftyOne](https://github.com/voxel51/fiftyone) dataset with 6763 samples.
## Installation
If you haven't already, install FiftyOne:
```bash
pip install -U fiftyone
```
## Usage
```python
import fiftyone as fo
from fiftyone.utils.huggingface import load_from_hub
# Load the dataset
# Note: other available arguments include 'max_samples', etc
dataset = load_from_hub("harpreetsahota/AVM_Segmentation_train")
# Launch the App
session = fo.launch_app(dataset)
```
## Dataset Details
### Dataset Description
The AVM dataset is a specialized computer vision dataset designed for training semantic segmentation models for autonomous parking systems. It contains bird's-eye view images from around-view monitoring cameras with pixel-level annotations for parking space detection and obstacle avoidance.
* **Curated by:** Chulhoon Jang and team at [original repository](https://github.com/ChulhoonJang/avm_dataset)
* **FiftyOne Integration by:** Harpreet Sahota (Voxel51)
* **License:** Please refer to the [original dataset repository](https://github.com/ChulhoonJang/avm_dataset) for license information (which currently has no License)
### Dataset Sources
* **Original Repository:** [https://github.com/ChulhoonJang/avm_dataset](https://github.com/ChulhoonJang/avm_dataset)
## Uses
### Direct Use
This dataset is designed for:
- **Autonomous Parking Systems**: Training models to detect and navigate into parking spaces
- **Semantic Segmentation Research**: Benchmarking segmentation algorithms on fisheye/bird's-eye view images
- **Parking Space Detection**: Identifying available vs occupied parking spots
- **Obstacle Detection**: Recognizing curbs, pillars, walls, and other vehicles
- **360° Surround View Systems**: Enhancing camera-based parking assistance features
### Out-of-Scope Use
This dataset should NOT be used for:
- Forward-facing autonomous driving (it's specifically bird's-eye view)
- General object detection (annotations are polygon-based for segmentation)
- High-speed navigation (designed for low-speed parking scenarios)
- Pedestrian detection (pedestrians are not annotated)
## Dataset Structure
### Overview
- **Total Images**: 6,763 (320 x 160 pixels)
- **Training Set**: 4,057 images
- **Test Set**: 2,706 images
- **Outdoor Images**: 3,614
- **Indoor Images**: 3,149
### Semantic Classes
The dataset contains 5 semantic classes with specific RGB color mappings:
| Class | Description | RGB Color | Hex Color |
|-------|------------|-----------|-----------|
| 0 | Free Space (drivable area) | [0, 0, 255] | #0000FF (Blue) |
| 1 | Marker (parking lines) | [255, 255, 255] | #FFFFFF (White) |
| 2 | Vehicle (other cars) | [255, 0, 0] | #FF0000 (Red) |
| 3 | Other (curbs, pillars, walls) | [0, 255, 0] | #00FF00 (Green) |
| 4 | Ego Vehicle (camera car) | [0, 0, 0] | #000000 (Black) |
### FiftyOne Fields
When parsed into FiftyOne, each sample includes:
| Field | Type | Description |
|-------|------|-------------|
| `filepath` | string | Path to the image file |
| `split` | string | "train" or "test" |
| `sample_id` | int | Unique identifier from filename |
| `environment` | Classification | "indoor" or "outdoor" (heuristic based on curb presence) |
| `parking_type` | Classification | "perpendicular" or "parallel" |
| `slot_type` | Classification | "closed", "opened", or "no_marker" |
| `polygon_annotations` | Polylines | Normalized polygon coordinates for each object |
| `ground_truth` | Segmentation | Pixel-level segmentation mask |
| `classes_present` | list | Classes present in the image |
| `num_markers` | int | Count of parking marker polygons |
| `num_vehicles` | int | Count of vehicle polygons |
| `has_curb` | bool | Whether curb is present |
| `has_ego_vehicle` | bool | Whether ego vehicle is annotated |
## Dataset Creation
### Curation Rationale
The dataset was created to address the lack of bird's-eye view datasets for autonomous parking systems. Most existing datasets focus on forward-facing cameras, but parking assistance requires a top-down perspective to accurately detect parking spaces and navigate safely.
### Source Data
#### Data Collection and Processing
- **Camera Setup**: Around View Monitoring (AVM) system with fisheye cameras
- **View Angle**: Bird's-eye view (top-down perspective)
- **Resolution**: 320 x 160 pixels (optimized for embedded systems)
- **Environments**: Real parking lots (both indoor parking garages and outdoor lots)
- **Conditions**: Various lighting conditions, weather (sunny, cloudy, rainy)
#### Who are the source data producers?
The original dataset was produced by researchers developing autonomous parking systems, likely in an academic or industrial research setting.
### Annotations
#### Annotation Process
1. **Polygon Annotation**: Each object is annotated with precise polygon boundaries in YAML format
2. **Semantic Masks**: Ground truth masks are generated from polygon annotations
3. **Multi-polygon Support**: Multiple instances of the same class are supported (e.g., multiple vehicles)
4. **Coordinate System**: Polygons use image coordinates (0-319 x 0-159)
#### Who are the annotators?
Information about specific annotators is not provided in the original dataset documentation.
## Personal and Sensitive Information
The dataset contains images from parking lots but does not include:
- License plate information (resolution too low)
- Personally identifiable information
- Pedestrian annotations
- Location-specific information
## Bias, Risks, and Limitations
### Known Limitations
1. **Limited Resolution**: 320x160 pixels may not capture fine details
2. **Geographic Bias**: Dataset may be from specific geographic regions
3. **Weather Conditions**: Limited representation of extreme weather
4. **Vehicle Types**: May not include all vehicle types (trucks, motorcycles, etc.)
5. **Parking Styles**: Primarily perpendicular and parallel parking
### Technical Challenges
- **Indoor Reflections**: Reflected lights can be mistaken for parking markers
- **Fisheye Distortion**: Bird's-eye view introduces geometric distortions
- **Class Imbalance**: Some classes (like curbs) appear less frequently
## Recommendations
1. **Augmentation**: Apply data augmentation to improve model robustness
2. **Validation**: Test models on diverse parking environments not in the dataset
3. **Resolution**: Consider upscaling techniques if higher resolution is needed
4. **Edge Cases**: Be aware that the dataset may not cover all parking scenarios
### Exploring the Dataset
```python
# View class distribution
print(dataset.count_values("classes_present"))
# Filter indoor vs outdoor
indoor = dataset.match(F("environment.label") == "indoor")
outdoor = dataset.match(F("environment.label") == "outdoor")
# Samples with multiple vehicles
multi_vehicle = dataset.match(F("num_vehicles") > 2)
```
## Citation
### BibTeX
```bibtex
@dataset{avm_dataset,
title={AVM (Around View Monitoring) System Datasets for Auto Parking},
author={Chulhoon Jang and others},
year={2020},
url={https://github.com/ChulhoonJang/avm_dataset}
}
```
### APA
Jang, C., et al. (2020). AVM (Around View Monitoring) System Datasets for Auto Parking. GitHub. https://github.com/ChulhoonJang/avm_dataset
## More Information
### Related Resources
- [Original Dataset Repository](https://github.com/ChulhoonJang/avm_dataset)
- [FiftyOne Documentation](https://docs.voxel51.com)
- Implementation code for semantic segmentation models (link in original repo)
### Dataset Statistics
- Average polygons per class:
- Ego vehicle: 1.0 polygons (fixed position)
- Markers: 2.6 polygons per image
- Vehicles: 2.1 polygons per image
- Curbs: 1.4 polygons per image (when present)
## Dataset Card Authors
- **FiftyOne Integration**: Harpreet Sahota (Voxel51)
- **Original Dataset**: Chulhoon Jang and team
## Dataset Card Contact
- **Original dataset**: See [original repository](https://github.com/ChulhoonJang/avm_dataset)
---
## Acknowledgments
Thanks to the original dataset creators for making this valuable resource available to the research community. The FiftyOne integration enhances the dataset's usability for modern computer vision workflows.