0

Welcome Gemma 4: Frontier multimodal intelligence on device

https://huggingface.co/blog/gemma4(huggingface.co)
Google DeepMind has released the Gemma 4 family of open, multimodal models on Hugging Face, supporting image, text, and audio inputs. These models are designed for high performance and on-device use, coming in four sizes including dense and mixture-of-experts (MoE) variants. Key architectural features include alternating attention layers, Per-Layer Embeddings (PLE), and a Shared KV Cache to improve efficiency and support long contexts. The release includes broad support for popular libraries like transformers, llama.cpp, and MLX, facilitating widespread deployment and fine-tuning.
0 pointsby ogg9 hours ago

Comments (0)

No comments yet. Be the first to comment!

Want to join the discussion?