Ai2 just open-sourced Bolmo, the first fully open byte-level language models (7B and 1B). Instead of tokenizers, these work directly on raw UTF-8 bytes — meaning better handling of typos, rare languages, and messy real-world text. Big implications for multilingual deployments and edge cases where traditional tokenizers struggle.
Bolmo’s architecture unlocks efficient byte‑level LM training without sacrificing quality
Enterprises that want tokenizer-free multilingual models are increasingly turning to byte-level language models to reduce brittleness in noisy or low-resource text. To tap into that niche — and make it practical at scale — the Allen Institute of AI (Ai2) introduced Bolmo, a new family of models that leverage its Olmo 3 models by “bytefiying” them and reusing their backbone and capabilities. The company launched two versions, Bolmo 7B and Bolmo 1B, which are “the first fully open byte-l
Like
1
0 Commentaires 0 Parts 19 Vue
Zubnet https://www.zubnet.ca