AI-ML
Transformers v5.0.0
RESUMEN
What's Changed This release candidate was focused mostly on `quantization` support with the new dynamic weight loader, and a few notable 🚨 breaking changes🚨: 1. Default dtype for any model when using `from_pretrained` is now `auto`! * Default auto 🚨 🚨 by @ArthurZucker in
Descripción Detallada
What's Changed This release candidate was focused mostly on `quantization` support with the new dynamic weight loader, and a few notable 🚨 breaking changes🚨: 1. Default dtype for any model when using `frompretrained` is now `auto`! Default auto 🚨 🚨 by @ArthurZucker in 2. Default shard size when saving a model is now 50GB: 🚨🚨 [saving] Default to 50GB shards, and remove non-safe serialization by @Cyrilvallez in This is now as fast as before thanks to xet, and is just more convenient on the hub. 3. Kwargs. They are fundamental to enable integration with vllm and other toosl: Every model forward() should have kwargs by @Rocketknight1 in Dynamic weight loader updates: Mostly QOL and fixed + support back CPU offloading. mark params as ishfinitialized with DS Zero3 from weight conversion by @winglian in [loading] Allow loading to happen without threading by @Cyrilvallez in [loading] Correctly load params during offloading & careful memory considerations by @Cyrilvallez in * allow registration of custom checkpoint conversion mappings by @winglian in
Transformers v5.0.0 introduce cambios importantes y rompe compatibilidad.
- El dtype por defecto al usar from_pretrained ahora es auto.
- El tamaño de shard por defecto al guardar un modelo es ahora 50GB.
- Se requiere kwargs en cada forward() de modelo para integración con vllm y otras herramientas.
A quién le importa
Solo si usas from_pretrained o guardas modelos en Transformers.
Generado por IA · puede contener errores
Releases Relacionados
AI-ML
Transformers v5.3.0
## New Model additions ### EuroBERT <img width="1080" height="1080" alt="image" src="https://github.com/user-attachments/assets/33603f42-5435-421a-9641-baf72faacb22" /> EuroBERT is a multilingual encoder model based on a refreshed transformer architecture, akin to Llama but with bidirection
AI-ML
Transformers v5.2.0
## New Model additions ### VoxtralRealtime <img width="1920" height="1080" alt="image" src="https://github.com/user-attachments/assets/80e37670-6d70-402b-8c8e-ccfb8c32df2d" /> VoxtralRealtime is a streaming speech-to-text model from [Mistral AI](https://mistral.ai), designed for real-time a