Introducing EmbeddingGemma Brand new open weights (under the slightly janky Gemma license) 308M parameter embedding model from Google:Based on the Gemma 3 architecture, EmbeddingGemma is trained on 100+ languages and is small enough to run on less than 200MB of RAM with quantization. It's available via sentence-transformers, llama.cpp, MLX, Ollama, LMStudio and more. As usual for these smaller models there's a Transformers.js demo (via) that runs directly in the browser (in Chrome variants) -...