Shalfunnn commited on
Commit
aa10d79
·
verified ·
1 Parent(s): d304709

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -7
README.md CHANGED
@@ -9,13 +9,13 @@
9
 
10
  </div>
11
 
12
- ## Model Description
13
 
14
  We introduce **WALL-OSS**, an end-to-end embodied foundation model that leverages large-scale multimodal pretraining to achieve (1) embodiment-aware vision--language understanding, (2) strong language--action association, and (3) robust manipulation capability.
15
- Our approach employs a tightly coupled architecture and multi-strategies training curriculum that enables {Unified Cross-Level CoT}—seamlessly unifying instruction reasoning, subgoal decomposition, and fine-grained action synthesis within a single differentiable framework.
16
  Our results show that WALL-OSS attains high success on complex long-horizon manipulations, demonstrates strong instruction-following capabilities, complex understanding and reasoning, and outperforms strong baselines, thereby providing a reliable and scalable path from VLMs to embodied foundation models.
17
 
18
- ## Quick Start
19
 
20
  ### Installation
21
 
@@ -52,7 +52,7 @@ model = model.to(device).bfloat16()
52
  # Your inference code here...
53
  ```
54
 
55
- ## Supervised Fine-Tuning (SFT)
56
 
57
  For training Wall-X on your robotics datasets, please refer to our comprehensive training guide:
58
 
@@ -70,7 +70,7 @@ The training process includes:
70
  bash ./workspace/lerobot_example/run.sh
71
  ```
72
 
73
- ## Inference
74
 
75
  For detailed inference examples and model evaluation:
76
 
@@ -139,7 +139,7 @@ python ./scripts/draw_openloop_plot.py
139
 
140
  **📁 [View all inference scripts](https://github.com/X-Square-Robot/wall-x/tree/main/scripts)**
141
 
142
- ## Complete Documentation
143
 
144
  For comprehensive setup, training, and inference instructions:
145
 
@@ -152,7 +152,7 @@ The repository contains:
152
  - **Configuration Templates**: Ready-to-use configs for different robot setups
153
  - **Troubleshooting Guide**: Common issues and solutions
154
 
155
- ## 📚 Cite Us
156
 
157
  If you find WALL-OSS models useful, please cite:
158
 
 
9
 
10
  </div>
11
 
12
+ ## 🤖 Model Description
13
 
14
  We introduce **WALL-OSS**, an end-to-end embodied foundation model that leverages large-scale multimodal pretraining to achieve (1) embodiment-aware vision--language understanding, (2) strong language--action association, and (3) robust manipulation capability.
15
+ Our approach employs a tightly coupled architecture and multi-strategies training curriculum that enables Unified Cross-Level CoT—seamlessly unifying instruction reasoning, subgoal decomposition, and fine-grained action synthesis within a single differentiable framework.
16
  Our results show that WALL-OSS attains high success on complex long-horizon manipulations, demonstrates strong instruction-following capabilities, complex understanding and reasoning, and outperforms strong baselines, thereby providing a reliable and scalable path from VLMs to embodied foundation models.
17
 
18
+ ## 🚀 Quick Start
19
 
20
  ### Installation
21
 
 
52
  # Your inference code here...
53
  ```
54
 
55
+ ## 🎯 Supervised Fine-Tuning (SFT)
56
 
57
  For training Wall-X on your robotics datasets, please refer to our comprehensive training guide:
58
 
 
70
  bash ./workspace/lerobot_example/run.sh
71
  ```
72
 
73
+ ## 🔮 Inference
74
 
75
  For detailed inference examples and model evaluation:
76
 
 
139
 
140
  **📁 [View all inference scripts](https://github.com/X-Square-Robot/wall-x/tree/main/scripts)**
141
 
142
+ ## 📚 Complete Documentation
143
 
144
  For comprehensive setup, training, and inference instructions:
145
 
 
152
  - **Configuration Templates**: Ready-to-use configs for different robot setups
153
  - **Troubleshooting Guide**: Common issues and solutions
154
 
155
+ ## 📄 Cite Us
156
 
157
  If you find WALL-OSS models useful, please cite:
158