HN
Today

Gemma 4 on iPhone

Google's AI Edge Gallery app brings the Gemma 4 LLM to iPhones, enabling powerful, fully offline AI capabilities like agent skills and multimodal interactions. This release excites developers by offering local, private inference and new avenues for experimenting with 'unaligned' models. The community is buzzing about the implications for on-device privacy, performance on NPUs, and the potential for a new era of client-side AI applications.

92
Score
21
Comments
#1
Highest Rank
26h
on Front Page
First Seen
Apr 5, 7:00 PM
Last Seen
Apr 6, 8:00 PM
Rank Over Time
5111121323323232544681119192121

The Lowdown

Google has launched the AI Edge Gallery, an innovative app that allows users to run the newly released Gemma 4 family of Large Language Models (LLMs) directly on their iPhone. This initiative emphasizes fully offline, private, and high-performance Generative AI experiences, leveraging the device's hardware without sending data to external servers.

Key features and capabilities highlighted in the app include:

  • On-Device Gemma 4: Experience cutting-edge LLM performance entirely offline.
  • Agent Skills: Extend LLM functionality with modular tools, such as Wikipedia integration and interactive maps, with the ability to load custom skills from URLs or GitHub.
  • AI Chat with Thinking Mode: Engage in multi-turn conversations and visualize the model's step-by-step reasoning process.
  • Ask Image: Utilize multimodal AI for object identification, visual puzzle-solving, and detailed descriptions using device camera or photo gallery.
  • Audio Scribe: Real-time, on-device transcription and translation of voice recordings.
  • Prompt Lab: A dedicated environment for testing prompts and fine-tuning model parameters.
  • Mobile Actions: Offline device controls and automated tasks powered by a finetuned FunctionGemma 270m.
  • Model Management & Benchmark: Download and manage open-source models, and run benchmarks on specific hardware.
  • 100% On-Device Privacy: Guarantees total privacy as all model inferences occur locally without internet connectivity.
  • Community-Built: An open-source project designed for developers and AI enthusiasts to contribute.

This app represents a significant step towards powerful, private, and localized AI, empowering users with advanced generative capabilities directly on their mobile devices.

The Gossip

On-Device Optimism & Performance

Users are highly enthusiastic about the app's ability to run advanced LLMs like Gemma 4 directly on an iPhone, highlighting the significant privacy benefits and the innovative 'mobile actions' feature. There's considerable discussion about the potential for future performance gains as dedicated Neural Processing Units (NPUs) in newer chips mature, promising even faster, more efficient on-device AI.

Censorship Circumvention & Model Modding

A notable portion of the discussion revolves around the freedom offered by local models to bypass content moderation or 'nannies' present in cloud-based AI. Commenters explore tools and methods like 'abliterating' models to achieve 'unaligned' or 'heretical' outputs, sparking debate on the ethical justification for such practices and whether this empowers more good actors than bad.

App Store Aesthetic & Browser Blips

Some users reported visual glitches and perceived low quality on the App Store listing page, particularly for the Dutch version, describing pixelated text and flickering elements. This prompted a troubleshooting discussion, with others confirming similar issues depending on their browser and OS (e.g., Firefox on Windows), while many reported no problems, suggesting browser rendering quirks rather than a fundamental flaw.