mmBERT is trained on 3T tokens from over 1800 languages, showing SoTA scores on benchmarks and exceptional low-resource performance
AI & ML interests
None defined yet.
Recent Activity
View all activity
Papers
Genomic Next-Token Predictors are In-Context Learners
Controlled Generation for Private Synthetic Text
mmBERT is trained on 3T tokens from over 1800 languages, showing SoTA scores on benchmarks and exceptional low-resource performance
A collection of SOTA, open-data, paired encoder-only and decoder only models ranging from 17M params to 1B. See the paper at https://arxiv.org/abs/250
rank1 is the first test-time compute reasoning model in IR