Update model card for v0.5
Browse files
README.md
CHANGED
@@ -1,15 +1,3 @@
|
|
1 |
-
---
|
2 |
-
base_model: unsloth/Qwen2.5-Coder-32B-Instruct-bnb-4bit
|
3 |
-
tags:
|
4 |
-
- text-generation-inference
|
5 |
-
- transformers
|
6 |
-
- unsloth
|
7 |
-
- qwen2
|
8 |
-
- trl
|
9 |
-
license: apache-2.0
|
10 |
-
language:
|
11 |
-
- en
|
12 |
-
---
|
13 |
|
14 |
# Qwen2.5-Coder-32B-Instruct-WMX
|
15 |
Pre-fine-tuned LoRA adapters for unsloth/Qwen2.5-Coder-32B-Instruct.
|
@@ -19,25 +7,25 @@ Pre-fine-tuned LoRA adapters for unsloth/Qwen2.5-Coder-32B-Instruct.
|
|
19 |
- https://huggingface.co/datasets/Jake5/wmx-doc-user
|
20 |
- https://huggingface.co/datasets/Jake5/wmx-doc-robot
|
21 |
|
22 |
-
## Version v0.
|
23 |
- Source: lora_model
|
24 |
- Base model: unsloth/Qwen2.5-Coder-32B-Instruct
|
25 |
-
- Uploaded on: 2025-09-
|
26 |
|
27 |
## Usage
|
28 |
```python
|
29 |
from peft import PeftModel
|
30 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
31 |
base_model = AutoModelForCausalLM.from_pretrained("unsloth/Qwen2.5-Coder-32B-Instruct")
|
32 |
-
model = PeftModel.from_pretrained(base_model, "Jake5/Qwen2.5-Coder-32B-Instruct-WMX", subfolder="adapters_v0.
|
33 |
-
tokenizer = AutoTokenizer.from_pretrained("Jake5/Qwen2.5-Coder-32B-Instruct-WMX", subfolder="adapters_v0.
|
34 |
```
|
35 |
|
36 |
## vLLM Serving
|
37 |
```bash
|
38 |
python -m vllm.entrypoints.openai.api_server \
|
39 |
--model unsloth/Qwen2.5-Coder-32B-Instruct \
|
40 |
-
--lora-modules my-lora=Jake5/Qwen2.5-Coder-32B-Instruct-WMX/adapters_v0.
|
41 |
--dtype bfloat16 \
|
42 |
--port 8000
|
43 |
```
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
|
2 |
# Qwen2.5-Coder-32B-Instruct-WMX
|
3 |
Pre-fine-tuned LoRA adapters for unsloth/Qwen2.5-Coder-32B-Instruct.
|
|
|
7 |
- https://huggingface.co/datasets/Jake5/wmx-doc-user
|
8 |
- https://huggingface.co/datasets/Jake5/wmx-doc-robot
|
9 |
|
10 |
+
## Version v0.5
|
11 |
- Source: lora_model
|
12 |
- Base model: unsloth/Qwen2.5-Coder-32B-Instruct
|
13 |
+
- Uploaded on: 2025-09-08
|
14 |
|
15 |
## Usage
|
16 |
```python
|
17 |
from peft import PeftModel
|
18 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
19 |
base_model = AutoModelForCausalLM.from_pretrained("unsloth/Qwen2.5-Coder-32B-Instruct")
|
20 |
+
model = PeftModel.from_pretrained(base_model, "Jake5/Qwen2.5-Coder-32B-Instruct-WMX", subfolder="adapters_v0.5")
|
21 |
+
tokenizer = AutoTokenizer.from_pretrained("Jake5/Qwen2.5-Coder-32B-Instruct-WMX", subfolder="adapters_v0.5")
|
22 |
```
|
23 |
|
24 |
## vLLM Serving
|
25 |
```bash
|
26 |
python -m vllm.entrypoints.openai.api_server \
|
27 |
--model unsloth/Qwen2.5-Coder-32B-Instruct \
|
28 |
+
--lora-modules my-lora=Jake5/Qwen2.5-Coder-32B-Instruct-WMX/adapters_v0.5 \
|
29 |
--dtype bfloat16 \
|
30 |
--port 8000
|
31 |
```
|