Google open sourced two 'little steel cannons'! 270 million parameters counterattack SOTA!

HongKong.info
Technology
22 Dec 2025 09:56:00 AM
The AI model race is ushering in a "lightweight revolution", and Google has recently released two small parameter high-performance models - Gemma 3 270M and EmbeddingGamma. The former achieves state-of-the-art performance.
Google open sourced two 'little steel cannons'! 270 million parameters counterattack SOTA!

The AI model race is ushering in a "lightweight revolution", and Google has recently released two small parameter high-performance models - Gemma 3 270M and EmbeddingGamma. The former achieves state-of-the-art performance for instruction following tasks with 270 million parameters, while the latter tops the MTEB text embedding list with a 300 million parameter scale. The two "small steel cannons" break the monopoly of large models on top performance and inject strong momentum into the landing of end-to-end AI applications with their core advantage of "small yet strong".

As the core force of this open source, Gemma 3 270M can be regarded as a model of "small parameters, big energy". This model, which only contains 270 million parameters, innovatively integrates 170 million embedded parameters and 100 million Transformer module parameters, combined with a 256k large vocabulary, to accurately handle rare professional terms. In the IFEval instruction following benchmark test, it achieved an excellent score of 51.2%, significantly surpassing small models of the same magnitude. What is even more remarkable is its ultimate energy efficiency. After INT4 quantification, it only consumes 0.75% battery power when running on Pixel 9 Pro phones, and can smoothly support offline operation of lightweight hardware such as web browsers and Raspberry Pi, completely eliminating dependence on cloud computing power.

Another "little cannon" EmbeddingGamma focuses on breaking through the core scene of text embedding. As a derivative model of the Gemma 3 series, although its 308 million parameter scale is much smaller than mainstream models, it ranks at the top of the MTEB Global General Evaluation List in the 500M parameter and below group. This model adopts an optimized Transformer encoder architecture, outputs 768 dimensional embedding vectors, supports 2048 token long text input, and converts any length of text into a fixed vector through average pooling technology. The unique Matryoshka representation learning technology enables it to have flexible adaptability, with vector dimensions that can be trimmed to 128 dimensions as needed, significantly saving storage and computing costs while ensuring performance.

Google open sourced two 'little steel cannons'! 270 million parameters counterattack SOTA!

The open source of two "small steel cannons" precisely targets the pain points of the industry. The current large models generally face problems such as high computing power costs, high privacy risks, and difficulty in deploying on the end side. However, Google's two models have a quantified memory usage of less than 200MB, which can be carried by ordinary mobile phones. Among them, Gemma 3 270M supports quick fine-tuning in just a few minutes and has been implemented in creative applications such as the "Bedtime Story Generator"; EmbeddingGamma is specifically optimized for local RAG scenarios, enabling offline operation of the entire retrieval generation process, perfectly adapting to the private knowledge base construction needs of sensitive industries such as finance and healthcare.

The comprehensive ecological support further amplifies the practical value of the two models. They are compatible with mainstream toolchains such as Hugging Face and LangChain, and support cross platform deployment of Weaviate vector databases and ONNX. Developers can quickly integrate applications with just a few lines of code. With the cumulative download volume of Gemma series models exceeding 200 million times, this dual model open source will accelerate the wave of AI innovation on the end side. Industry analysis points out that Google's move not only reconstructs the industry perception of "parameter scale=efficiency", but also promotes the shift of AI competition focus towards "efficient adaptation". In the future, lightweight and high-efficiency small models will become the core carrier of end-to-end intelligence, opening a new era of inclusive AI applications.

Keywords:
Technology Google
Disclaimers:

HongKong.info Committed to providing fair and transparent reports. This article aims to provide accurate and timely information, but should not be construed as financial or investment advice. Due to the rapidly changing market conditions, we recommend that you verify the information yourself and consult a professional before making any decisions based on this information.