LaMI: Augmenting Large Language Models
via Late Multi-Image Fusion

ACL 2026

1The Hebrew University of Jerusalem 2Bar-Ilan University

* Equal contribution

TL;DR

LaMI augments text-only LLMs with visual commonsense by generating multiple images from the input text at inference time and integrating them via a lightweight late-fusion layer, without any multimodal retraining. It improves visual reasoning while enhancing text-based performance.

Abstract

Commonsense reasoning often requires both textual and visual knowledge, yet Large Language Models (LLMs) trained solely on text lack visual grounding (e.g., "what color is an emperor penguin's belly?"). Visual Language Models (VLMs) perform better on visually grounded tasks but face two limitations: (i) often reduced performance on text-only commonsense reasoning compared to text-trained LLMs, and (ii) adapting newly released LLMs to vision input typically requires costly multimodal training. An alternative augments LLMs with test-time visual signals, improving visual commonsense without harming textual reasoning, but prior designs often rely on early fusion and a single image, which can be suboptimal.

We propose a late multi-image fusion method: multiple images are generated from the text prompt with a lightweight parallel sampling, and their prediction probabilities are combined with those of a text-only LLM through a late-fusion layer that integrates projected visual features just before the final prediction. Across visual commonsense and NLP benchmarks, our method significantly outperforms augmented LLMs on visual reasoning, matches VLMs on vision-based tasks, and, when applied to strong LLMs such as LLaMA 3, also improves NLP performance while adding only modest test-time overhead.

Method

LaMI enhances language models with visual cues to improve object and visual commonsense reasoning while preserving text-only performance. The architecture consists of four components: a frozen pre-trained LLM, a frozen pre-trained vision encoder, a trainable Visual Token Projector (VTP), and a trainable Late Fusion Attention Layer (LFAL).

Overview of the LaMI architecture. Multiple images are generated from the input text and independently encoded by a frozen vision encoder, then projected to pseudo-text tokens. In parallel, the text is processed by a frozen pre-trained LLM. A trainable late-fusion attention layer allows the LLM's final text representations to attend to the projected visual tokens, combining both modalities before the prediction head.

Figure 2. Overview of LaMI. Multiple images are generated from the input text and independently encoded by a frozen vision encoder, then projected to pseudo-text tokens. In parallel, the text is processed by a frozen pre-trained LLM. A trainable late-fusion attention layer allows the LLM's final text representations to attend to the projected visual tokens, combining both modalities before the prediction head. Blue: frozen; orange: trainable.

Late Fusion Architecture. The vision encoder extracts patch features, which the VTP maps to pseudo-text embeddings. The LFAL fuses these with text embeddings by allowing text tokens to attend once to visual tokens immediately before projecting to vocabulary logits. This design keeps the LLM focused on language while enabling access to visual information when helpful.

Multi-Image Evidence at Inference. Since paired images are unavailable at test time, k images are generated from the input text using a distilled text-to-image generator with batched, parallel sampling. Each generated image is processed through the late-fusion module to produce a probability distribution, which is then aggregated with the text-only distribution using entropy-aware, CLIP-based weighting.

Results

Object Commonsense

LaMI substantially improves performance across all object commonsense tasks compared to prior visually-augmented language models, while using equal or less multimodal training data.

Model Mem. Color Color Terms Obj. Shape Rel. Size
Masked Language Models
BERT31.630.728.138.1
BERT (FT)33.931.521.535.7
Vokenization*14.220.043.272.4
X-adapter*64.160.0--
LaMI74.572.567.378.4
Causal Language Models
GPT-232.434.644.543.1
GPT-2 (FT)33.334.939.338.2
LiVE49.646.741.566.7
Z-LaVI*50.449.264.476.8
VaLM* (k=4)54.052.762.885.0
VaLM* (k=8)58.650.259.462.4
LaMI72.569.266.885.5

* retrieves images; zero-shot; generates images during inference.

Ablation: Fusion Strategy and Multi-Image Generation

Both late fusion and multi-image generation independently contribute to performance. Their combination achieves the best overall results.

Method Mem. Color Color Terms Obj. Shape Rel. Size
GPT-2 (Base)32.434.644.543.1
Early Fusion49.145.340.370.1
Early Fusion + Multi55.552.141.275.5
Intermediate Fusion62.859.360.077.2
Intermediate Fusion + Multi69.767.863.081.1
Late Fusion65.162.263.580.2
Late Fusion + Multi (Ours)72.569.266.885.5

Visual Commonsense, Commonsense Reasoning, and Reading Comprehension

LaMI consistently improves LMs across all scales. Unlike VLMs, which often improve visual commonsense at the cost of text-task performance, LaMI enhances visual commonsense while maintaining or improving text-based results.

Model Base VC CR RC Avg.
Small-Scale Models
GPT-2-30.346.130.535.6
LaMIGPT-238.646.732.239.2
Mid-Scale Models
Gemma-2B-45.663.848.852.7
LaMIGemma-2B50.165.148.954.7
Large-Scale Models
Vicuna-7B-45.157.657.553.4
InstructBLIP*Vicuna-7B50.152.653.652.1
Llava-Next*Vicuna-7B50.354.554.753.1
LaMIVicuna-7B48.658.857.955.1
Llama3-8B-52.072.057.960.6
LaMILlama3-8B55.072.958.062.0
Large-Scale Instruct Models
Llama3-8B-Instruct-53.071.659.261.2
Llava-Next*Llama3-8B-Inst.56.570.854.860.7
LaMILlama3-8B-Inst.55.671.760.962.7
Qwen-2.5-7B-Instruct-54.875.663.064.4
Qwen2-VL-7B-Instruct*Qwen-2.5-7B-Inst.59.074.457.963.7
LaMIQwen-2.5-7B-Inst.57.875.964.466.0

* VLM trained on large-scale image-text data. VC = Visual Commonsense, CR = Commonsense Reasoning, RC = Reading Comprehension.

Qualitative Analysis

Representative examples using Llama-3 illustrating the behavior of LaMI.

Success Case

"How many humps does a Bactrian camel have?"

Llama-3 predicts one, confusing Dromedary and Bactrian species. LaMI generates images of two-humped camels, correcting the prediction to two.

Robustness to Negation

"Which color is not on a stop sign?"

The generator produces a red stop sign, which is misleading under negation. However, the low CLIP alignment score suppresses the visual path, and LaMI falls back to the text-only prediction, correctly outputting blue.

Failure Case

"What material holds the Sword of Damocles?"

Llama-3 predicts a thin rope. LaMI generates images depicting a metal chain—a visually plausible but incorrect depiction (the correct answer is a single horse hair). The high alignment score causes the visual path to override the text prior. Such failures are more likely for abstract or legendary concepts where text-to-image generators lack faithful grounding.

BibTeX

@inproceedings{yariv2026lami,
  title={LaMI: Augmenting Large Language Models via Late Multi-Image Fusion},
  author={Yariv, Guy and Schwartz, Idan and Adi, Yossi and Benaim, Sagie},
  booktitle={Proceedings of the 64th Annual Meeting of the Association for Computational Linguistics (ACL)},
  year={2026}
}