ESEthan Steininger
••7 min read🎯 Multimodal Monday #2 — From Tiny VLMs to 10M‑Token Titans
📢 Quick Take (TL;DR) * Major multimodal model releases: Meta unveiled Llama 4 Scout & Maverick – open Mixture-of-Experts models with native text+image (and even video/audio) support – and Microsoft introduced Phi-4-Multimodal, a compact 3.8B-param model integrating vision, text, and spee (Today is the start of a new era of natively multimodal AI… | AI at Meta | 190 comments) ([2503.01743] Phi-4-Mini Technical Report: Compact yet Powerful Multimodal Language Models via Mixture-of-LoRAs)7】. Bot

Multimodal Monday
📢 Quick Take (TL;DR)
- Major multimodal model releases: Meta unveiled Llama 4 Scout & Maverick – open Mixture-of-Experts models with native text+image (and even video/audio) support – and Microsoft introduced Phi-4-Multimodal, a compact 3.8B-param model integrating vision, text, and spee (Today is the start of a new era of natively multimodal AI… | AI at Meta | 190 comments) ([2503.01743] Phi-4-Mini Technical Report: Compact yet Powerful Multimodal Language Models via Mixture-of-LoRAs)7】. Both push state-of-the-art capabilities, from 10M-token long contexts to top-tier multimodal reasoning, widening access to advanced AI.
- Efficiency & accessibility lead the week: Hugging Face’s new SmolVLM-256M (the smallest vision-language model yet) runs on <1 GB R (Hugging Face claims world’s smallest vision language models)6】, and novel training techniques (e.g. synthetic data for retrieval models) achieved state-of-the-art results without giant model siz (GME: Improving Universal Multimodal Retrieval by Multimodal LLMs)9】. In short, multimodal AI is getting both stronger and easier to deploy.
- Trust & safety get attention: Researchers showed a single malicious image can “poison” a retrieval-augmented generator, causing it to return misleading answers to many queri (One Pic is All it Takes: Poisoning Visual Document Retrieval Augmented Generation with a Single Image)4】. This highlights growing efforts to evaluate and bolster the robustness of multimodal systems as they become more widely used.
🧠 Research Highlights
- Phi-4-Multimodal (Microsoft) – Introduces a unified multimodal language model (3.8B params) that uses “mixture-of-LoRA” adapters to handle text, image, and audio inputs without retraining the whole model. Impressively, it outperforms much larger vision-language and speech models on many ta ([2503.01743] Phi-4-Mini Technical Report: Compact yet Powerful Multimodal Language Models via Mixture-of-LoRAs)27】. Why it matters: Shows that carefully adding modular adapters can yield small models with big-model skills (even topping an OpenASR speech leaderboard).
- GME for Universal Retrieval – Proposes a multimodal LLM-based retriever that works across text, images, and image+text documents. By training on a balanced mix of single-, cross-, and fused-modal data, GME achieves state-of-the-art results on a new benchmark (UMRB) covering diverse retrieval ta (GME: Improving Universal Multimodal Retrieval by Multimodal LLMs)89】. Takeaway: One model can now match task-specific retrievers in domains like visual documents, pointing toward more unified retrieval syst (GME: Improving Universal Multimodal Retrieval by Multimodal LLMs)89】.
- ImageScope (WWW ’25) – A training-free, three-stage framework that uses LLM reasoning (chain-of-thought) to unify various language-guided image retrieval tasks. It reframes tasks (text queries, image+text queries, dialog-based queries, etc.) into a general text-to-image retrieval problem and then uses a multimodal model to verify and rerank resu (ImageScope: Unifying Language-Guided Image Retrieval via Large Multimodal Model Collective Reasoning)L4】. On 6 diverse datasets, ImageScope outperformed dedicated task-specific baseli (ImageScope: Unifying Language-Guided Image Retrieval via Large Multimodal Model Collective Reasoning)72】. Why it matters: Clever use of “collective reasoning” lets one system handle many image search flavors without specialized training.
- ReflectiVA (Knowledge-Augmented VQA) – Introduces “self-reflective tokens” for multimodal LLMs to decide when to fetch external knowledge. Built on LLaVA, ReflectiVA learns to dynamically trigger text retrieval from a database and weigh the results, rather than relying only on baked-in training d (Augmenting Multimodal LLMs with Self-Reflective Tokens for Knowledge-based Visual Question Answering)48】. It set new highs on knowledge-intensive visual QA ta (Augmenting Multimodal LLMs with Self-Reflective Tokens for Knowledge-based Visual Question Answering)48】. Takeaway: Bridging LLMs with external information on the fly can greatly improve accuracy on detailed or long-tail visual questions (and the code is open-source for developers).
🛠️ Tools & Techniques
- Hugging Face SmolVLM – New SmolVLM-256M and 500M models claim to be the smallest vision-language models to (Hugging Face claims world’s smallest vision language models)L18】. The 256M version (256 million params) can even run on a laptop with <1 GB RAM, yet handles tasks like image captioning and document Q&A with performance comparable to much larger models from a year (Hugging Face claims world’s smallest vision language models)L26】. How you might use it: Deploy multimodal AI on edge devices or in cost-sensitive pipelines – e.g. running basic vision-language analysis on smartphones or batch-processing millions of images on a budget.
- Meta Llama 4 (Scout & Maverick) – Meta’s latest open models (17B active params, MoE architecture) arrived on Hugging Face this (Welcome Llama 4 Maverick & Scout on Hugging Face)L83】. They are natively multimodal (accept text and images out-of-the-box) and boast massive context windows (up to 10 million tok (Today is the start of a new era of natively multimodal AI… | AI at Meta | 190 comments)-L4】. Notably, Llama 4 Scout runs on a single H100 GPU (thanks to INT4 quantization) and both models outperform many 70B+ dense models in reasoning and coding benchm (Today is the start of a new era of natively multimodal AI… | AI at Meta | 190 comments)-L4】. How you might use it: Fine-tune Llama 4 on your own multimodal dataset or use it via API for tasks like long-document understanding with images – it’s readily available in the Transformers library and inference (Welcome Llama 4 Maverick & Scout on Hugging Face)L83】.
🏗️ Real-World Applications
- Healthcare QA gets visual: Google Cloud’s Vertex AI Search (Healthcare) just added a Visual Q&A feature, enabling doctors to query medical records using images like charts or x-rays dir (Google Cloud unveils new genAI and visual search tech at HIMSS25 | Healthcare IT News)-L77】. Instead of relying only on text, the system analyzes visual inputs (e.g. a diagram from a patient intake form) to find relevant info. This multimodal search promises more comprehensive clinical answers – a boon when ~90% of healthcare data is imagery. (Powered under the hood by Google’s Gemini 2.0 multimodal model.)
- Assistive Tech Retrieval: A new assistive technology search dataset (from an ECIR 2025 paper) showed that representing products with both text and image features helps find the best disability aids for (Multimodal Feature Extraction for Assistive Technology: Evaluation and Dataset | SpringerLink)1-L4】. In tests, a multimodal model beat text-only or image-only approaches at predicting which assistive products meet certain user “goals.” Interestingly, for abstract criteria (like a user’s end-goal), single modalities sufficed, but for detailed specs the multimodal richness wo (Multimodal Feature Extraction for Assistive Technology: Evaluation and Dataset | SpringerLink)1-L4】. Why it matters: This real-world study (built from a production AT database) underscores how multimodal retrieval can improve recommendations for special-needs equipment – an area aligned with inclusive design and UN accessibility goals.
📈 Trends & Predictions
- Small & Modular is Big: Rather than simply scaling up parameter counts, the community is embracing efficient architectures. Meta’s Llama 4 uses mixture-of-experts to boost capacity by activating subsets of weights, and Phi-4-Multimodal keeps a base model frozen and adds LoRA adapters per modality – a technique that outperforms more brute-force appr ([2503.01743] Phi-4-Mini Technical Report: Compact yet Powerful Multimodal Language Models via Mixture-of-LoRAs)L1-L4】. Expect more such modular tricks (MoE, adapters, distillation) to deliver multimodal prowess without requiring trillion-parameter models.
- Edge AI and Accessibility: A clear push is on to make multimodal AI run anywhere. From “smol” models by Hugging Face to optimized encoders, new releases aim to be **accessible on lower-spec hard (Hugging Face claims world’s smallest vision language models)3-L28】. This trend means we’ll see multimodal assistants in mobile apps, IoT cameras, AR glasses, etc., bringing vision-and-language understanding to devices that previously couldn’t handle large models.
- Robustness & Evaluation Focus: As multimodal systems deploy in high-stakes settings, there’s rising concern with their reliability. Recent work exposing poisoning vulnerabi (One Pic is All it Takes: Poisoning Visual Document Retrieval Augmented Generation with a Single Image)2-L47】 and studies on how well multimodal models truly “understand” (e.g. aligning with human brain semantics) are spurring efforts to harden these models. We anticipate more benchmarks and defenses around factuality, security, and bias in multimodal AI – not just raw performance.
- Enterprise Adoption Accelerates: Big players are investing in domain-specific multimodal AI. This week’s example: Google’s visual search in healthcare showing strong industry (Google Cloud unveils new genAI and visual search tech at HIMSS25 | Healthcare IT News)1-L69】. We foresee finance, security, education, and other sectors following suit with custom multimodal solutions (often retrieval-augmented). The presence of multimodal features in cloud platforms and enterprise software signals that “multimodal everything” could be the next competitive differentiator in AI products.
🧩 Community + Shoutouts
- Meta announces LlamaCon: Mark your calendars – LlamaCon, Meta’s first developer conference for generative AI (named after the Llama model series), is set for April 2 (This Week in AI: Maybe we should ignore AI benchmarks for now | TechCrunch)6-L204】. This reflects how large AI providers are actively building developer communities around their models. We expect to see demos, fine-tuning tutorials, and perhaps some Llama 4 surprises at the event. 🦙🎉
- MAGMaR Workshop CFP: The 1st Workshop on Multimodal Augmented Generation via Retrieval (MAGMaR) is calling f ( MAGMaR Workshop )†L53-L61】. Slated for ACL later this year, it focuses on exactly our newsletter’s theme: combining text, images, video, etc. with retrieval and generation. There’s even a shared task on video retrieval. Shoutout to scholars – submissions are open till May 1, 2025, for those pushing the envelope in multimodal RAG.
- Llama 4 hype: The open-release of Llama 4 caused an eruption on social media. Meta’s announcement (“a new era of natively multimodal AI”) racked up over 6,500 likes and hundreds of comments wit (Today is the start of a new era of natively multimodal AI… | AI at Meta | 190 comments)†L50-L58】. Developers across Reddit/Twitter immediately started experimenting – e.g. loading Llama 4 into chatbots and even running decentralized instances. 🚀 The enthusiasm highlights the growing impact of community-driven AI: when a powerful multimodal model is freely available, creators jump in to build cool demos and applications almost overnight.