LLaMmlein2Vec_1B / README.md
Julia287's picture
Create README.md
9ec5904 verified
---
datasets:
- togethercomputer/RedPajama-Data-V2
language:
- de
library_name: transformers
license: other
pipeline_tag: feature-extraction
tags:
- masked-lm
- long-context
base_model:
- LSX-UniWue/LLaMmlein_1B
---
# LLäMmlein2Vec 1B
LLäMmlein2Vec 1B is a German encoder language model derived from our German decoder-only model [LLäMmlein 1B](https://huggingface.co/LSX-UniWue/LLaMmlein_1B) via [LLM2Vec](https://github.com/McGill-NLP/llm2vec).
Find more details in our [preprint](https://arxiv.org/abs/2505.13136)!
We provide three transformed models:
* [LLäMmlein 7B](https://huggingface.co/LSX-UniWue/LLaMmlein2Vec_7B)
* [LLäMmlein 1B](https://huggingface.co/LSX-UniWue/LLaMmlein2Vec_1B) ← You are here
* [LLäMmlein 120M](https://huggingface.co/LSX-UniWue/LLaMmlein2Vec_120M)
### Usage
You can use LLäMmlein2Vec with the `llm2vec` library.
```python
import torch
from llm2vec import LLM2Vec
model_id = "LSX-UniWue/LLaMmlein2Vec_1B"
l2v = LLM2Vec.from_pretrained(
model_id,
device_map="cuda" if torch.cuda.is_available() else "cpu",
torch_dtype=torch.bfloat16,
)
```
### License
We release the ModernGBERT models under a research-only RAIL-M license. See [license.md](./license.md) for details.