Gemma-4-21B-REAP model released, performs well on inference tasks

ME News Report, April 6 (UTC+8), Hugging Face recently released the Gemma-4-21B-REAP model. According to the publisher, the model performs well on inference tasks and even achieves improved accuracy. In terms of VRAM requirements, when provided with some context, the model can run on as little as 12GB of VRAM; for full context, 16GB of VRAM is needed. The publisher calls on MLX and GGUF community members to try it out. (Source: InFoQ)

View Original
This page may contain third-party content, which is provided for information purposes only (not representations/warranties) and should not be considered as an endorsement of its views by Gate, nor as financial or professional advice. See Disclaimer for details.
  • Reward
  • Comment
  • Repost
  • Share
Comment
Add a comment
Add a comment
No comments
  • Pin