On-device AI

Explore our entire collection of insights, tutorials, and industry news.

  • Model Reviews

    Implementing Transformers.js in a Chrome Extension

    A comprehensive technical guide on integrating on-device machine learning within Chrome Extensions using Transformers.js and Manifest V3, featuring hybrid strategies with cloud APIs.
    Read more
  • AI Tutorials

    Running 400B Parameter AI Models on a Smartphone

    A technical breakdown of how Flash-MoE and Apple's 'LLM in a Flash' research enabled a 400-billion parameter model to run on an iPhone, and what it means for the future of hybrid AI applications.
    Read more
  • AI Tutorials

    How to Run a 400B Parameter LLM on a Phone

    Discover the engineering breakthroughs behind running massive 400B models like DeepSeek-V3 or Llama 3 on mobile hardware using flash offloading and quantization.
    Read more