FineGrainedFP8Config(dequantize=true) materializes full BF16 model per rank before sharding

May 1, 2026 ยท #2114
View on GitHub
Python Difficulty: Medium

Sign in required

Authenticate to use favourites & bookmarks

5