[Bug] Gemma4 MoE backend missing vision-aware attention mask for use_bidirectional_attention="vision" models
April 17, 2026 ยท #1891
Python
Difficulty: Medium
Labels
bug
Parent Repository
NVIDIA-NeMo/Automodel
Python repository
447 133