Update README.md
Browse files
README.md
CHANGED
@@ -42,7 +42,7 @@ model = AutoModelForCausalLM.from_pretrained(
|
|
42 |
).to("cuda" if torch.cuda.is_available() else "cpu") # Send model to GPU if available
|
43 |
|
44 |
# 🛠 **Define Inference Function**
|
45 |
-
def generate_response(model, tokenizer, prompt, max_new_tokens=
|
46 |
|
47 |
# Tokenize input
|
48 |
inputs = tokenizer(prompt, return_tensors="pt", padding=True, truncation=True).to(model.device)
|
|
|
42 |
).to("cuda" if torch.cuda.is_available() else "cpu") # Send model to GPU if available
|
43 |
|
44 |
# 🛠 **Define Inference Function**
|
45 |
+
def generate_response(model, tokenizer, prompt, max_new_tokens=4096, temperature=0.7):
|
46 |
|
47 |
# Tokenize input
|
48 |
inputs = tokenizer(prompt, return_tensors="pt", padding=True, truncation=True).to(model.device)
|