Update config.json
#5
by
Corellios
- opened
Hi, first, thank you for making this model :)
When using LLM Compressor to quantize in FP8, I got the following error when first loading the model:
model = AutoModelForCausalLM.from_pretrained(MODEL_ID, dtype=torch.bfloat16, low_cpu_mem_usage=True, device_map="cuda", local_files_only=True)
rope_scaling's beta_fast field must be a float, got 32rope_scaling's beta_slow field must be a float, got 1
It might be my local environment only (my transformers version is "4.57.3"), but just in case you observe the same.
Nope, your fix is indeed required to suppress the warning (though the warning is harmless). Thanks again, merged!
baileyk
changed pull request status to
merged