FineGrainedFP8Config(dequantize=true) materializes full BF16 model per rank before sharding
May 1, 2026 ยท #2114
Python
Difficulty: Medium
Parent Repository
NVIDIA-NeMo/Automodel
Python repository
503 150
NVIDIA-NeMo/Automodel
Python repository
Sign in required
Authenticate to use favourites & bookmarks