Tekmono
  • News
  • Guides
  • Lists
  • Reviews
  • Deals
No Result
View All Result
Tekmono
No Result
View All Result
Home News
Apple’s FastVLM AI Model Now Runs in Browser

Apple’s FastVLM AI Model Now Runs in Browser

by Tekmono Editorial Team
02/09/2025
in News
Share on FacebookShare on Twitter

Apple has made its FastVLM (Visual Language Model) available for testing directly in a web browser, allowing users with Apple Silicon-powered Macs to experience its near-instant high-resolution image processing capabilities.

The model, initially released a few months ago, leverages MLX, Apple’s open machine learning framework optimized for Apple Silicon, to achieve significantly faster video captioning and a smaller model size compared to its counterparts. FastVLM is reported to be up to 85 times faster in video captioning and more than three times smaller than comparable models.

Apple has expanded the project’s availability, making it accessible on Hugging Face in addition to GitHub. Users can now load the lightweight FastVLM-0.5B model directly in their browser to evaluate its performance. According to tests, loading the model takes a couple of minutes on a 16GB M2 Pro MacBook Pro.

Related Reads

OpenAI Launches Customizable Skills for Codex Coding Agent

Amazon’s Alexa+ to Integrate with Four New Services

EA Investigated for AI-Generated Content in Battlefield 6

Apple to Start iPhone 18 Production in January

Once loaded, the model accurately describes the user’s appearance, the surrounding environment, expressions, and objects in view. Users can interact with the model by adjusting the prompt or selecting from preset options such as “Describe what you see in one sentence,” “What is the color of my shirt?”, “Identify any text or written content visible,” “What emotions or actions are being portrayed?” and “Name the object I am holding in my hand.” This allows for a customized and interactive experience.

Furthermore, users can employ a virtual camera app to feed live video to the tool, enabling it to instantly describe multiple scenes in detail. This highlights the model’s speed and accuracy. A key feature of this implementation is that it runs locally in the browser, ensuring data privacy as no data leaves the device.

The model can also operate offline, making it suitable for applications in wearables and assistive technology, where low latency and efficiency are crucial. The browser-based demo utilizes the 0.5-billion-parameter version of FastVLM. The FastVLM family includes larger variants with 1.5 billion and 7 billion parameters.

While these larger models could offer improved performance and speed, running them directly in the browser is unlikely due to resource constraints. Apple encourages users to test the model and share their feedback.

ShareTweet

You Might Be Interested

OpenAI Launches Customizable Skills for Codex Coding Agent
News

OpenAI Launches Customizable Skills for Codex Coding Agent

24/12/2025
Amazon’s Alexa+ to Integrate with Four New Services
News

Amazon’s Alexa+ to Integrate with Four New Services

24/12/2025
EA Investigated for AI-Generated Content in Battlefield 6
News

EA Investigated for AI-Generated Content in Battlefield 6

24/12/2025
Apple to Start iPhone 18 Production in January
News

Apple to Start iPhone 18 Production in January

24/12/2025
Please login to join discussion

Recent Posts

  • OpenAI Launches Customizable Skills for Codex Coding Agent
  • Amazon’s Alexa+ to Integrate with Four New Services
  • EA Investigated for AI-Generated Content in Battlefield 6
  • Apple to Start iPhone 18 Production in January
  • Connect Your Phone to Wi-Fi Easily

Recent Comments

No comments to show.
  • News
  • Guides
  • Lists
  • Reviews
  • Deals
Tekmono is a Linkmedya brand. © 2015.

No Result
View All Result
  • News
  • Guides
  • Lists
  • Reviews
  • Deals