• AI Engineering
  • Posts
  • Apple Releases FastVLM: Blazing-Fast Vision-Language Model

Apple Releases FastVLM: Blazing-Fast Vision-Language Model

.. PLUS: Eigent – 100% Open Source Multi-Agent Workforce

In today’s newsletter:

  • Glean:LIVE – AMA + engineer-led demos (Sept 25)

  • Apple’s FastVLM – Blazing-Fast Vision-Language Model

  • Eigent – Local, Open Source Multi-Agent Workforces

Reading time: 3 minutes.

FREE LIVE SESSION — TOGETHER WITH GLEAN

The Glean team is organizing a live technical deep dive with their engineering team on September 25 at 9:30 AM PT.

The session will cover:

  • Internals of the new Assistant and how it handles personalization at query time

  • How to program agents with “vibe coding” (rapid agent composition without boilerplate)

  • Latency and accuracy benchmarks from real-world workflows

  • Under-the-hood demos of ranking pipelines and orchestration, followed by a live AMA with engineers

If you are exploring large-scale personalization, relevance modeling, or multi-agent workflows, this session provides implementation details and performance results you can apply to your own systems.

Now, let’s get back into the newsletter!

Apple released FastVLM, a vision-language model optimized for speed and on-device performance. Think lightning-fast multimodal reasoning without depending on cloud infrastructure.

Why it matters:

  • Blazing Speed - Hybrid vision encoder (FastViTHD) delivers up to 85× faster response and 7.9× lower TTFT than LLaVA-OneVision

  • On-Device First - Runs locally on laptops and smartphones, boosting privacy and responsiveness

  • Compact + Capable - Smaller model variants make deployment practical without losing performance

  • High-Resolution Support - Processes images up to 1152×1152 pixels for detailed understanding

Apple also shipped WebGPU support and released models on Hugging Face, signaling they want developers to experiment openly and not just on iPhones.

🔗 Check out the Github repo and Hugging Face demo

Imagine building your own specialized workforce of AI agents completely local, fully open source.

Eigent is a desktop application that lets you build, manage, and deploy custom agent workforces to turn complex workflows into automated tasks.

Key Features:

  • Multi-Agent by Default – Developer, searcher, document analyzer, and multimodal agents work in parallel.

  • Zero Setup – Lightweight desktop app, quick install.

  • Customizable – Add your own models, plugins, or API integrations (Slack, Notion, browsing).

  • Human in the Loop – Requests feedback when ambiguity arises.

  • Prebuilt Templates – Research assistance, travel planning, reporting, and more.

  • MCP Integration – Native support for multi-tool coordination.

  • Drag & Drop Workflows – Visual workflow editor makes automation more intuitive.

This is one of the first fully-local open-source multi-agent solutions. It’s like AutoGPT, but designed as a sustainable desktop platform with transparency and user control.

That’s a Wrap

That’s all for today. Thank you for reading today’s edition. See you in the next issue with more AI Engineering insights.

PS: We curate this AI Engineering content for free, and your support means everything. If you find value in what you read, consider sharing it with a friend or two.

Your feedback is valuable: If there’s a topic you’re stuck on or curious about, reply to this email. We’re building this for you, and your feedback helps shape what we send.

WORK WITH US

Looking to promote your company, product, or service to 150K+ AI developers? Get in touch today by replying to this email.