Hugging Face
Models
Datasets
Spaces
Community
Docs
Enterprise
Pricing
Log In
Sign Up
50
63
188
Nick Doiron
monsoon-nlp
Follow
mmhamdy's profile picture
pitagoras-alves's profile picture
zhaoxu98's profile picture
37 followers
·
72 following
https://mapmeld.com/plant-based-llms/
mapmeld
mapmeld.bsky.social
AI & ML interests
biology and multilingual models
Recent Activity
liked
a model
about 2 hours ago
jhu-clsp/mmBERT-base
upvoted
an
article
about 2 hours ago
mmBERT: ModernBERT goes Multilingual
reacted
to
tomaarsen
's
post
with ❤️
about 2 hours ago
ModernBERT goes MULTILINGUAL! One of the most requested models I've seen, The Johns Hopkins University's CLSP has trained state-of-the-art massively multilingual encoders using the ModernBERT architecture: mmBERT. Model details: - 2 model sizes: - https://huggingface.co/jhu-clsp/mmBERT-small - https://huggingface.co/jhu-clsp/mmBERT-base - Uses the ModernBERT architecture, but with the Gemma2 multilingual tokenizer (so: flash attention, alternating global/local attention, unpadding/sequence packing, etc.) - Maximum sequence length of 8192 tokens, on the high end for encoders - Trained on 1833 languages using DCLM, FineWeb2, and many more sources - 3 training phases: 2.3T tokens pretraining on 60 languages, 600B tokens mid-training on 110 languages, and 100B tokens decay training on all 1833 languages. - Both models are MIT Licensed, and the full datasets and intermediary checkpoints are also publicly released Evaluation details: - Very competitive with ModernBERT at equivalent sizes on English (GLUE, MTEB v2 English after finetuning) - Consistently outperforms equivalently sized models on all Multilingual tasks (XTREME, classification, MTEB v2 Multilingual after finetuning) - In short: beats commonly used multilingual base models like mDistilBERT, XLM-R (multilingual RoBERTa), multilingual MiniLM, etc. - Additionally: the ModernBERT-based mmBERT is much faster than the alternatives due to its architectural benefits. Easily up to 2x throughput in common scenarios. Check out the full blogpost with more details. It's super dense & gets straight to the point: https://huggingface.co/blog/mmbert Based on these results, mmBERT should be the new go-to multilingual encoder base models at 300M and below. Do note that the mmBERT models are "base" models, i.e. they're currently only trained to perform Mask Filling. They'll need to be finetuned for downstream tasks like semantic search, classification, clustering, etc.
View all activity
Organizations
monsoon-nlp
's datasets
11
Sort: Recently updated
monsoon-nlp/genetic-counselor-freeform-questions
Viewer
•
Updated
Jun 27
•
44
•
11
monsoon-nlp/papaya-plant
Updated
Jun 24
•
2
monsoon-nlp/wheat-bees
Updated
Apr 23
•
16
monsoon-nlp/relive-qa
Preview
•
Updated
Mar 23
•
379
•
2
monsoon-nlp/genetic-counselor-multiple-choice
Viewer
•
Updated
Mar 14
•
225
•
70
•
1
monsoon-nlp/redcode-hf
Viewer
•
Updated
Jan 15
•
630
•
158
monsoon-nlp/greenbeing-binary
Viewer
•
Updated
Jul 16, 2024
•
4.46k
•
8
monsoon-nlp/greenbeing-proteins
Viewer
•
Updated
May 20, 2024
•
3.95M
•
13
•
2
monsoon-nlp/asknyc-chatassistant-format
Viewer
•
Updated
Mar 30, 2024
•
13.4k
•
11
monsoon-nlp/primate-proteins
Viewer
•
Updated
Mar 28, 2024
•
27.2k
•
6
monsoon-nlp/protein-pairs-uniprot-swissprot
Viewer
•
Updated
Mar 23, 2024
•
2.15M
•
29
•
2