Update model card for v0.9
Browse files
README.md
CHANGED
@@ -1,3 +1,4 @@
|
|
|
|
1 |
---
|
2 |
base_model: unsloth/Qwen2.5-Coder-32B-Instruct-bnb-4bit
|
3 |
tags:
|
@@ -19,25 +20,25 @@ Pre-fine-tuned LoRA adapters for unsloth/Qwen2.5-Coder-32B-Instruct.
|
|
19 |
- https://huggingface.co/datasets/Jake5/wmx-doc-user
|
20 |
- https://huggingface.co/datasets/Jake5/wmx-doc-robot
|
21 |
|
22 |
-
## Version v0.
|
23 |
- Source: lora_model
|
24 |
- Base model: unsloth/Qwen2.5-Coder-32B-Instruct
|
25 |
-
- Uploaded on: 2025-09-
|
26 |
|
27 |
## Usage
|
28 |
```python
|
29 |
from peft import PeftModel
|
30 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
31 |
base_model = AutoModelForCausalLM.from_pretrained("unsloth/Qwen2.5-Coder-32B-Instruct")
|
32 |
-
model = PeftModel.from_pretrained(base_model, "Jake5/Qwen2.5-Coder-32B-Instruct-WMX", subfolder="adapters_v0.
|
33 |
-
tokenizer = AutoTokenizer.from_pretrained("Jake5/Qwen2.5-Coder-32B-Instruct-WMX", subfolder="adapters_v0.
|
34 |
```
|
35 |
|
36 |
## vLLM Serving
|
37 |
```bash
|
38 |
python -m vllm.entrypoints.openai.api_server \
|
39 |
--model unsloth/Qwen2.5-Coder-32B-Instruct \
|
40 |
-
--lora-modules my-lora=Jake5/Qwen2.5-Coder-32B-Instruct-WMX/adapters_v0.
|
41 |
--dtype bfloat16 \
|
42 |
--port 8000
|
43 |
```
|
|
|
1 |
+
|
2 |
---
|
3 |
base_model: unsloth/Qwen2.5-Coder-32B-Instruct-bnb-4bit
|
4 |
tags:
|
|
|
20 |
- https://huggingface.co/datasets/Jake5/wmx-doc-user
|
21 |
- https://huggingface.co/datasets/Jake5/wmx-doc-robot
|
22 |
|
23 |
+
## Version v0.9
|
24 |
- Source: lora_model
|
25 |
- Base model: unsloth/Qwen2.5-Coder-32B-Instruct
|
26 |
+
- Uploaded on: 2025-09-12
|
27 |
|
28 |
## Usage
|
29 |
```python
|
30 |
from peft import PeftModel
|
31 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
32 |
base_model = AutoModelForCausalLM.from_pretrained("unsloth/Qwen2.5-Coder-32B-Instruct")
|
33 |
+
model = PeftModel.from_pretrained(base_model, "Jake5/Qwen2.5-Coder-32B-Instruct-WMX", subfolder="adapters_v0.9")
|
34 |
+
tokenizer = AutoTokenizer.from_pretrained("Jake5/Qwen2.5-Coder-32B-Instruct-WMX", subfolder="adapters_v0.9")
|
35 |
```
|
36 |
|
37 |
## vLLM Serving
|
38 |
```bash
|
39 |
python -m vllm.entrypoints.openai.api_server \
|
40 |
--model unsloth/Qwen2.5-Coder-32B-Instruct \
|
41 |
+
--lora-modules my-lora=Jake5/Qwen2.5-Coder-32B-Instruct-WMX/adapters_v0.9 \
|
42 |
--dtype bfloat16 \
|
43 |
--port 8000
|
44 |
```
|