Update README.md
Browse files
README.md
CHANGED
@@ -9,13 +9,13 @@
|
|
9 |
|
10 |
</div>
|
11 |
|
12 |
-
## Model Description
|
13 |
|
14 |
We introduce **WALL-OSS**, an end-to-end embodied foundation model that leverages large-scale multimodal pretraining to achieve (1) embodiment-aware vision--language understanding, (2) strong language--action association, and (3) robust manipulation capability.
|
15 |
-
Our approach employs a tightly coupled architecture and multi-strategies training curriculum that enables
|
16 |
Our results show that WALL-OSS attains high success on complex long-horizon manipulations, demonstrates strong instruction-following capabilities, complex understanding and reasoning, and outperforms strong baselines, thereby providing a reliable and scalable path from VLMs to embodied foundation models.
|
17 |
|
18 |
-
## Quick Start
|
19 |
|
20 |
### Installation
|
21 |
|
@@ -52,7 +52,7 @@ model = model.to(device).bfloat16()
|
|
52 |
# Your inference code here...
|
53 |
```
|
54 |
|
55 |
-
## Supervised Fine-Tuning (SFT)
|
56 |
|
57 |
For training Wall-X on your robotics datasets, please refer to our comprehensive training guide:
|
58 |
|
@@ -70,7 +70,7 @@ The training process includes:
|
|
70 |
bash ./workspace/lerobot_example/run.sh
|
71 |
```
|
72 |
|
73 |
-
## Inference
|
74 |
|
75 |
For detailed inference examples and model evaluation:
|
76 |
|
@@ -139,7 +139,7 @@ python ./scripts/draw_openloop_plot.py
|
|
139 |
|
140 |
**📁 [View all inference scripts](https://github.com/X-Square-Robot/wall-x/tree/main/scripts)**
|
141 |
|
142 |
-
## Complete Documentation
|
143 |
|
144 |
For comprehensive setup, training, and inference instructions:
|
145 |
|
@@ -152,7 +152,7 @@ The repository contains:
|
|
152 |
- **Configuration Templates**: Ready-to-use configs for different robot setups
|
153 |
- **Troubleshooting Guide**: Common issues and solutions
|
154 |
|
155 |
-
##
|
156 |
|
157 |
If you find WALL-OSS models useful, please cite:
|
158 |
|
|
|
9 |
|
10 |
</div>
|
11 |
|
12 |
+
## 🤖 Model Description
|
13 |
|
14 |
We introduce **WALL-OSS**, an end-to-end embodied foundation model that leverages large-scale multimodal pretraining to achieve (1) embodiment-aware vision--language understanding, (2) strong language--action association, and (3) robust manipulation capability.
|
15 |
+
Our approach employs a tightly coupled architecture and multi-strategies training curriculum that enables Unified Cross-Level CoT—seamlessly unifying instruction reasoning, subgoal decomposition, and fine-grained action synthesis within a single differentiable framework.
|
16 |
Our results show that WALL-OSS attains high success on complex long-horizon manipulations, demonstrates strong instruction-following capabilities, complex understanding and reasoning, and outperforms strong baselines, thereby providing a reliable and scalable path from VLMs to embodied foundation models.
|
17 |
|
18 |
+
## 🚀 Quick Start
|
19 |
|
20 |
### Installation
|
21 |
|
|
|
52 |
# Your inference code here...
|
53 |
```
|
54 |
|
55 |
+
## 🎯 Supervised Fine-Tuning (SFT)
|
56 |
|
57 |
For training Wall-X on your robotics datasets, please refer to our comprehensive training guide:
|
58 |
|
|
|
70 |
bash ./workspace/lerobot_example/run.sh
|
71 |
```
|
72 |
|
73 |
+
## 🔮 Inference
|
74 |
|
75 |
For detailed inference examples and model evaluation:
|
76 |
|
|
|
139 |
|
140 |
**📁 [View all inference scripts](https://github.com/X-Square-Robot/wall-x/tree/main/scripts)**
|
141 |
|
142 |
+
## 📚 Complete Documentation
|
143 |
|
144 |
For comprehensive setup, training, and inference instructions:
|
145 |
|
|
|
152 |
- **Configuration Templates**: Ready-to-use configs for different robot setups
|
153 |
- **Troubleshooting Guide**: Common issues and solutions
|
154 |
|
155 |
+
## 📄 Cite Us
|
156 |
|
157 |
If you find WALL-OSS models useful, please cite:
|
158 |
|