|
--- |
|
base_model: osmosis-ai/Osmosis-Structure-0.6B |
|
license: apache-2.0 |
|
model_creator: osmosis-ai |
|
model_name: Osmosis-Structure-0.6B |
|
quantized_by: Second State Inc. |
|
--- |
|
# 🧠 Osmosis-Structure 0.6B (GGUF) |
|
|
|
**Osmosis-Structure 0.6B** is a lightweight language model optimized for inference using the **GGUF** format. It's suitable for edge deployment, research, and low-resource environments. |
|
|
|
--- |
|
|
|
## 📦 Model Overview |
|
|
|
- **Model Size:** 0.6 Billion parameters |
|
- **Quantization:** Q4_K_M |
|
- **Format:** GGUF |
|
- **Tokenizer:** SentencePiece (`tokenizer.model`) |
|
- **Usage:** Optimized for fast inference with low memory requirements |
|
|
|
--- |
|
|
|
## 🧰 How to Use |
|
|
|
This model is in **GGUF format**, which is supported by: |
|
|
|
- [`llama.cpp`](https://github.com/ggerganov/llama.cpp) |
|
- [`llama-cpp-python`](https://github.com/abetlen/llama-cpp-python) |
|
- [`llm`](https://github.com/simonw/llm) |
|
|
|
### Example command using `llama.cpp`: |
|
|
|
```bash |
|
./main -m Osmosis-Structure-0.6B-Q4_K_M.gguf -p "Explain the structure of a water molecule." |
|
``` |
|
|
|
--- |
|
|
|
## 📁 Files Included |
|
|
|
- `Osmosis-Structure-0.6B-Q4_K_M.gguf` |
|
- `tokenizer.model` |
|
- `README.md` |
|
|
|
--- |
|
|
|
## ⚠️ License & Usage |
|
|
|
Please review the included license file (if any) for usage terms. This model is provided for educational and research purposes. |
|
|
|
--- |
|
|
|
## ✨ Maintained by |
|
|
|
Model hosted by [XythicK](https://xythick.vercel.app) |
|
Powered by open-source magic ⚡ |
|
|