|国家预印本平台
首页|Beyond the Linear Separability Ceiling

Beyond the Linear Separability Ceiling

Beyond the Linear Separability Ceiling

来源:Arxiv_logoArxiv
英文摘要

Most state-of-the-art Visual-Language Models (VLMs) are seemingly limited by the linear separabilty of their visual embeddings on abstract reasoning tasks. This work investigates this "linear reasoning bottleneck" by introducing the Linear Separability Ceiling (LSC), the performance of a simple linear classifier on a VLM's visual embeddings. We find this bottleneck is widespread and stems not from poor perception, but from failures in the language model's reasoning pathways. We demonstrate this is a solvable alignment issue. The required intervention, however, is task-dependent: activating existing pathways suffices for semantic concepts, while complex relational reasoning requires adapting core model weights. Using postfix tuning as a methodological control, we find strong evidence for powerful, dormant reasoning pathways within VLMs. However, for complex relational tasks requiring deeper adaptation, explicitly improving representation quality causes the model to fail on new prompt formats despite its embeddings remaining well separated. Ultimately, this work provides a new lens for VLM analysis, showing that robust reasoning is a matter of targeted alignment, not simply improved representation learning.

Enrico Vompa、Tanel Tammet、Mohit Vaishnav

计算技术、计算机技术

Enrico Vompa,Tanel Tammet,Mohit Vaishnav.Beyond the Linear Separability Ceiling[EB/OL].(2025-07-10)[2025-07-21].https://arxiv.org/abs/2507.07574.点此复制

评论