Jake5 commited on
Commit
42a3a14
·
verified ·
1 Parent(s): b2a0fc1

Update model card for v0.9

Browse files
Files changed (1) hide show
  1. README.md +6 -5
README.md CHANGED
@@ -1,3 +1,4 @@
 
1
  ---
2
  base_model: unsloth/Qwen2.5-Coder-32B-Instruct-bnb-4bit
3
  tags:
@@ -19,25 +20,25 @@ Pre-fine-tuned LoRA adapters for unsloth/Qwen2.5-Coder-32B-Instruct.
19
  - https://huggingface.co/datasets/Jake5/wmx-doc-user
20
  - https://huggingface.co/datasets/Jake5/wmx-doc-robot
21
 
22
- ## Version v0.8
23
  - Source: lora_model
24
  - Base model: unsloth/Qwen2.5-Coder-32B-Instruct
25
- - Uploaded on: 2025-09-11
26
 
27
  ## Usage
28
  ```python
29
  from peft import PeftModel
30
  from transformers import AutoModelForCausalLM, AutoTokenizer
31
  base_model = AutoModelForCausalLM.from_pretrained("unsloth/Qwen2.5-Coder-32B-Instruct")
32
- model = PeftModel.from_pretrained(base_model, "Jake5/Qwen2.5-Coder-32B-Instruct-WMX", subfolder="adapters_v0.8")
33
- tokenizer = AutoTokenizer.from_pretrained("Jake5/Qwen2.5-Coder-32B-Instruct-WMX", subfolder="adapters_v0.8")
34
  ```
35
 
36
  ## vLLM Serving
37
  ```bash
38
  python -m vllm.entrypoints.openai.api_server \
39
  --model unsloth/Qwen2.5-Coder-32B-Instruct \
40
- --lora-modules my-lora=Jake5/Qwen2.5-Coder-32B-Instruct-WMX/adapters_v0.8 \
41
  --dtype bfloat16 \
42
  --port 8000
43
  ```
 
1
+
2
  ---
3
  base_model: unsloth/Qwen2.5-Coder-32B-Instruct-bnb-4bit
4
  tags:
 
20
  - https://huggingface.co/datasets/Jake5/wmx-doc-user
21
  - https://huggingface.co/datasets/Jake5/wmx-doc-robot
22
 
23
+ ## Version v0.9
24
  - Source: lora_model
25
  - Base model: unsloth/Qwen2.5-Coder-32B-Instruct
26
+ - Uploaded on: 2025-09-12
27
 
28
  ## Usage
29
  ```python
30
  from peft import PeftModel
31
  from transformers import AutoModelForCausalLM, AutoTokenizer
32
  base_model = AutoModelForCausalLM.from_pretrained("unsloth/Qwen2.5-Coder-32B-Instruct")
33
+ model = PeftModel.from_pretrained(base_model, "Jake5/Qwen2.5-Coder-32B-Instruct-WMX", subfolder="adapters_v0.9")
34
+ tokenizer = AutoTokenizer.from_pretrained("Jake5/Qwen2.5-Coder-32B-Instruct-WMX", subfolder="adapters_v0.9")
35
  ```
36
 
37
  ## vLLM Serving
38
  ```bash
39
  python -m vllm.entrypoints.openai.api_server \
40
  --model unsloth/Qwen2.5-Coder-32B-Instruct \
41
+ --lora-modules my-lora=Jake5/Qwen2.5-Coder-32B-Instruct-WMX/adapters_v0.9 \
42
  --dtype bfloat16 \
43
  --port 8000
44
  ```