by meetpateltech on 5/20/25, 6:03 PM with 163 comments
by nolist_policy on 5/20/25, 7:54 PM
Download the Edge Gallery apk from github: https://github.com/google-ai-edge/gallery/releases/tag/1.0.0
Download one of the .task files from huggingface: https://huggingface.co/collections/google/gemma-3n-preview-6...
Import the .task file in Edge Gallery with the + bottom right.
You can take pictures right from the app. The model is indeed pretty fast.
by onlyrealcuzzo on 5/20/25, 6:53 PM
Gemma 3n is a model utilizing Per-Layer Embeddings to achieve an on-device memory footprint of a 2-4B parameter model.
At the same time, it performs nearly as well as Claude 3.7 Sonnet in Chatbot Arena.
by IceWreck on 5/20/25, 7:39 PM
E4B has a score of 44.4 in the Aider polyglot dashboard. Which means its on-par with gemini-2.5-flash (not the latest preview but the version used for the bench on aider's website), gpt4o and gpt4.5.
Thats sounds very good - imagine what a coding focused version of this could do if this is a "generic" embedded only model.
On the other hand - this does have a much lower score for livecodebench.
by ljosifov on 5/20/25, 8:14 PM
https://huggingface.co/collections/google/gemma-3n-preview-6...
Gemma 3n Preview
google/gemma-3n-E4B-it-litert-preview
google/gemma-3n-E2B-it-litert-preview
Interesting, hope it comes on LMStudio as MLX or GGUF. Sparse and or MoE models make a difference when running on localhost. MoE Qwen3-30B-A3B most recent game changer for me. Activating only 3b weights on the gpu cores of sparse Qwen3-30B-A3B, rather than comparable ~30b of dense models (Qwen3-32B, Gemma3-27b, GLM-{4,Z1}-32B, older QwQ-32B), is a huge speedup for me: MoE A3B achieves 20-60 tps on my oldish M2 in LMStudio, versus only 4-5 tps for the dense models.
Looking forward to trying gemma-3n. Kudos to Google for open sourcing their Gemmas. Would not have predicted that the lab with "open" in the name has yet to release even v1 (atm at 0; disregarding gpt-2), while other labs, more commercial labs, are are at versions 3, 4 etc already.
by jeroenhd on 5/20/25, 10:59 PM
The picture interpretation seems to work fine, as does the OCR capability. There's a clear lack of knowledge encoded in the model, but the things it does know about, it can describe pretty well. Impressive for a model only a bit larger than a DVD.
by lxgr on 5/20/25, 7:25 PM
On the other hand, I'm really not looking forward to app sizes ballooning even more – there's no reasonable way to share them across apps at least on iOS, and I can absolutely imagine random corporate apps to start including LLMs, just because it's possible.
by android521 on 5/21/25, 4:47 AM
by krackers on 5/20/25, 7:02 PM
And for that matter, what is
>mix’n’match capability in Gemma 3n to dynamically create submodels
It seems like mixture-of-experts taken to the extreme, where you actually create an entire submodel instead of routing per token?
by rvnx on 5/21/25, 10:28 AM
# Is Eiffel Tower or a soccer ball bigger ?
> A soccer ball is bigger than the Eiffel Tower! Here's a breakdown:
> Eiffel Tower: Approximately 330 meters (1,083 feet) tall.
> Soccer Ball: A standard soccer ball has a circumference of about 68-70 cm (27-28 inches).
> While the Eiffel Tower is very tall, its base is relatively small compared to its height. A soccer ball, though much smaller in height, has a significant diameter, making it physically larger in terms of volume.
by barnas2 on 5/20/25, 7:38 PM
by mltsd on 5/20/25, 11:10 PM
by impure on 5/20/25, 9:04 PM
Also it's funny that they are saying that Llama 4 Maverick performs about the same as GPT-4.1 Nano.
by turnsout on 5/20/25, 7:35 PM
by mmaunder on 5/21/25, 1:05 PM
Wow!!
by sujayk_33 on 5/21/25, 7:32 AM
in the video they've added in announcement, they are showing some live interaction with the model(which is quite fast as compared to AI Edge gallery app), how's it built, how can I use it like this?
by angst on 5/21/25, 2:13 AM
loads pretty fast. starts to reply near-instant (text chat mode).
doesn't answer questions like "when is your cutoff date"
apparently answers "may 15 2024" as today date so probably explains why it answered joe biden as answer to who is US president
by TOMDM on 5/20/25, 10:44 PM
Giving Gemini and other apps the ability to interact with each other feels like it has potential.
by einpoklum on 5/21/25, 9:04 AM
by devcb on 5/25/25, 8:15 AM
by username135 on 5/21/25, 1:07 PM
by jakemanger on 5/21/25, 4:37 AM
by adityakusupati on 5/20/25, 8:36 PM
by sandowsh on 5/21/25, 12:32 AM
by quaintdev on 5/20/25, 8:45 PM
Seems like we will not be able to run this with Llama and friends.
by cmcconomy on 5/20/25, 7:32 PM
by happy_one on 5/21/25, 2:38 PM
by bionhoward on 5/20/25, 11:50 PM
by jonplackett on 5/21/25, 10:28 AM
That’s all I really want for Christmas.