Jake5 commited on
Commit
9e24005
·
verified ·
1 Parent(s): 70a1567

Update model card for v0.5

Browse files
Files changed (1) hide show
  1. README.md +5 -17
README.md CHANGED
@@ -1,15 +1,3 @@
1
- ---
2
- base_model: unsloth/Qwen2.5-Coder-32B-Instruct-bnb-4bit
3
- tags:
4
- - text-generation-inference
5
- - transformers
6
- - unsloth
7
- - qwen2
8
- - trl
9
- license: apache-2.0
10
- language:
11
- - en
12
- ---
13
 
14
  # Qwen2.5-Coder-32B-Instruct-WMX
15
  Pre-fine-tuned LoRA adapters for unsloth/Qwen2.5-Coder-32B-Instruct.
@@ -19,25 +7,25 @@ Pre-fine-tuned LoRA adapters for unsloth/Qwen2.5-Coder-32B-Instruct.
19
  - https://huggingface.co/datasets/Jake5/wmx-doc-user
20
  - https://huggingface.co/datasets/Jake5/wmx-doc-robot
21
 
22
- ## Version v0.4
23
  - Source: lora_model
24
  - Base model: unsloth/Qwen2.5-Coder-32B-Instruct
25
- - Uploaded on: 2025-09-05
26
 
27
  ## Usage
28
  ```python
29
  from peft import PeftModel
30
  from transformers import AutoModelForCausalLM, AutoTokenizer
31
  base_model = AutoModelForCausalLM.from_pretrained("unsloth/Qwen2.5-Coder-32B-Instruct")
32
- model = PeftModel.from_pretrained(base_model, "Jake5/Qwen2.5-Coder-32B-Instruct-WMX", subfolder="adapters_v0.4")
33
- tokenizer = AutoTokenizer.from_pretrained("Jake5/Qwen2.5-Coder-32B-Instruct-WMX", subfolder="adapters_v0.4")
34
  ```
35
 
36
  ## vLLM Serving
37
  ```bash
38
  python -m vllm.entrypoints.openai.api_server \
39
  --model unsloth/Qwen2.5-Coder-32B-Instruct \
40
- --lora-modules my-lora=Jake5/Qwen2.5-Coder-32B-Instruct-WMX/adapters_v0.4 \
41
  --dtype bfloat16 \
42
  --port 8000
43
  ```
 
 
 
 
 
 
 
 
 
 
 
 
 
1
 
2
  # Qwen2.5-Coder-32B-Instruct-WMX
3
  Pre-fine-tuned LoRA adapters for unsloth/Qwen2.5-Coder-32B-Instruct.
 
7
  - https://huggingface.co/datasets/Jake5/wmx-doc-user
8
  - https://huggingface.co/datasets/Jake5/wmx-doc-robot
9
 
10
+ ## Version v0.5
11
  - Source: lora_model
12
  - Base model: unsloth/Qwen2.5-Coder-32B-Instruct
13
+ - Uploaded on: 2025-09-08
14
 
15
  ## Usage
16
  ```python
17
  from peft import PeftModel
18
  from transformers import AutoModelForCausalLM, AutoTokenizer
19
  base_model = AutoModelForCausalLM.from_pretrained("unsloth/Qwen2.5-Coder-32B-Instruct")
20
+ model = PeftModel.from_pretrained(base_model, "Jake5/Qwen2.5-Coder-32B-Instruct-WMX", subfolder="adapters_v0.5")
21
+ tokenizer = AutoTokenizer.from_pretrained("Jake5/Qwen2.5-Coder-32B-Instruct-WMX", subfolder="adapters_v0.5")
22
  ```
23
 
24
  ## vLLM Serving
25
  ```bash
26
  python -m vllm.entrypoints.openai.api_server \
27
  --model unsloth/Qwen2.5-Coder-32B-Instruct \
28
+ --lora-modules my-lora=Jake5/Qwen2.5-Coder-32B-Instruct-WMX/adapters_v0.5 \
29
  --dtype bfloat16 \
30
  --port 8000
31
  ```