Hyunin Lee

An Orthogonal Alignment Phenomenon in Cross-Attention

Left view of vector alignment
(a) Residual Alignment
Right view of vector alignment
(b) Orthogonal Alignment
Right view of vector alignment
(c) Cross-attention
Figure 1: Conceptual illustration of Orthogonal Alignment.
Given a source representation vector Y from domain B, suppose the algorithm progressively updates target representation vector X from domain A throughout training iterations {X₁, X₂, ⋯, X′}.
(a) Residual alignment: The prevailing view of cross-attention is that it refines X by reducing irrelevant and preserving relevant information by referring Y to update X′.
(b) Orthogonal Alignment: We observe a complement-discovery phenomenon where X′ becomes increasingly orthogonal to X as model performance improves. We show that this orthogonality emerges because cross-attention enables parameter-efficient scaling by extracting complementary information from an orthogonal manifold T(X), thus enhancing performance without a proportional increase in parameters.
(c) X′ is the output of cross-attention, with X as the query and Y as the key and value.

Preface

I’m excited to share a somewhat counterintuitive phenomenon- Orthogonal Alignment—with the open-world research community (see Figure 1(b)). Before diving in, a brief disclaimer: this phenomenon has so far been observed only in multi-domain recommendation data, so I remain cautious about generalizing it to vision-language models (or more broadly, to multi-modal learning).

That said, I’m optimistic that Orthogonal Alignment may also appear in vision-language settings, given that our study is grounded in transformer architectures with gated cross-attention—a core component of many modern fusion models. Still, as a researcher, I want to avoid overgeneralization and therefore frame this observation strictly within the recommendation domain until further studies confirm its presence in vision-language models.

Ultimately, my hope is that this discovery inspires new ways of thinking about algorithmic design and sheds light on how to achieve better scaling law in multi-modal models. In this post, I want to highlight one simple message:

When a multi-modal model exhibits the Orthogonal Alignment phenomenon, it tends to improve scaling law.

I’ve attempted to clarify why this phenomenon naturally occurs and identified one possible explanation during my internship at Meta: parameter-efficient scaling. However, the underlying mechanism behind why on earth this phenomenon naturally emerges still remains largely a black box, presenting opportunities for deeper investigation. I would be to discuss or further explore why this phenomenon arises — please feel free to reach out.


The Rise of Multi-Modal Recommendation Systems

Imagine you have a dataset D₁=(X₁,Z) and want to build a model that predicts a binary label Z from input X₁. In many real-world cases, the label Z is extremely sparse — meaning that most of the values are just zeros.

When I worked on the Ranking AI Research team at Meta, one of my main tasks was building recommendation models that display sponsored posts (ads) on Instagram and Facebook (If user clicks an ad, that’s how Meta earns money💰). The key challenge was data sparsity—users rarely clicked on ads, often engaging with only one out of ten sponsored posts, or sometimes none at all. In practice, even though the model continuously recommended posts over every minites – say, X₁(14:00), X₁(14:01), X₁(14:02), X₁(14:03)…, – most of these interactions resulted in no clicks, leaving almost all corresponding outcomes as Z = 0.

Since high-quality recommendations rely on accurately modeling user engagement, this extreme sparsity made it difficult to infer user intent. In other words, simply training on D₁ was not enough to build a truly effective recommendation system.

One effective way to address this problem is to incorporate richer signals from other domains D₂=(X₂,Z) — for example, how long a user stays on which type of post or whether they leave a comment or have shared with others. These additional behavioral cues provide valuable context about user interests and help reduce the impact of sparse labels of other domain D₁.

This observation motivates a central research problem in multi-modal learning – developing architectural principles that enable the effective fusion of heterogeneous behavioral modalities.

A widely adopted solution is the cross-attention mechanism, which learns to align and project information from different domains into a shared latent space. This allows the model to combine diverse signals and better capture a user’s overall intent — even when direct click data is scarce.


What Cross-Attention do?: Residual Alignment View

Despite its popularity, the internal mechanisms of cross-attention across domains remain poorly understood and are largely explored through empirical studies.

So far, current research views cross-attention as enabling one domain (X in Figure 1c) to query another (Y in Figure 1c) and integrate only the most relevant information (X’ as a weighted sum of Y in Figure 1c).

A growing body of empirical evidence supports this view, especially in various multi-modal models:

Therefore, understanding cross-attention as a “residual alignment” mechanism is the prevalent interpretation within the research community.

Current research interprets cross-attention as primarily a residual alignment mechanism, where the output (X′) is generated by removing redundant information and preserveing relevant content from the input (X) by referencing another domain (Y).

Orthogonal Alignment

This work challenges this conventional view and uncovers a new, counter-intuitive mechanism of cross-attention.

A Co-Existence Observation in Multi-Modal Learning
We argue that two contrasting alignment mechanisms are able to co-exist in cross attention:

1. Residual Alignment (conventional view)
2. Orthogonal Alignment (our discovery)

We define an Orthogonal Alignment Phenomenon as follows.

An Orthogonal Alignment is a phenomenon where the input query (X) and the output (X') of the cross-attention are orthogonal, rather than simply reinforcing the existing pre-aligned features of X when updating to X'

Please refer to Figure 1 for a visual illustration of this phenomenon, contrasted with the conventional residual-alignment perspective.

What is role of Y in Orthogonal Alignment?

After reading the above definition of Orthogonal Alignment, a natural question arises: “Then, what is the role of Y?”

My interpretation is that the query Y functions as a guide that identifies which directions on the tangent space of X correspond to positive transfer signals. More concretely, consider the tangent space of X. Within this space, there exist multiple orthogonal directions—some leading to negative transfer, others contributing to positive transfer. In principle, all of these directions could serve as candidates for X′, since they are orthogonal to the original X. Then, the introduction of Y provides the crucial signal that distinguishes among these directions—indicating which orthogonal components are constructive (positive transfer) and should therefore be incorporated into X′.

Intuitively, Y acts as a directional filter that orients the orthogonal updates toward beneficial regions of the feature manifold, enabling cross-attention to expand representational capacity without amplifying redundant correlations.

Experiment results.

Empirically, we observe that the Gated Cross-Attention (GCA) module enhances recommendation performance by generating outputs that are not merely filtered versions of the input query(See Figure 2). In simple terms, GCA introduces a learnable gating mechanism that combines the input and the cross-attention output as X + αX’ where X’ is the output of cross attention and α is a learnable parameter. This formulation allows the model to produce complementary representations—capturing aspects of the input query that were previously underrepresented or unseen.

We evaluated this effect using three recent Cross-Domain Sequential Recommendation (CDSR) models: LLM4CDSR¹, CDSRNP², and ABXI³ – all of which are transformer-based architectures reported as state-of-the-art in their respective papers. In Figure 2, the evaluation metric NDCG@10 (Normalized Discounted Cumulative Gain at rank 10) measures how accurately each model ranks the top 10 items compared with the ground-truth order—that is, how well it predicts both which items should appear and in what order. The x-axis represents the absolute value of the cosine similarity between X and X′ for both Domain A and Domain B, where the blue dots correspond to Domain A and the red dots correspond to Domain B.

To ensure robustness, we conducted experiments with multiple random initializations, several GCA architectural variants, and different datasets. Each configuration corresponds to one datapoint in the three subfigures below, and each point represents the best test results from its train process.

Overall, the results consistently demonstrate that the Orthogonal Alignment effect—induced by GCA—leads to model performance increases: Lower cosine similarity indicates stronger orthogonal alignment, which tends to correlate with higher NDCG@10.

cdsrnp
(a)CDSRNP
abxi
(b) ABXI
llm4cdsr
(b) LLM4CDSR
Figure2. We observed that the gated cross-attention module introduces an unseen, orthogonal feature representation: as the input query X and its cross-attended output X' (conditioned on key and value Y) become more orthogonal, the ranking performance improves. Blue color dots are domain A and red color dots are domain B

Orthogonal Alignment improves scaling law

Crucially, we classify orthogonal alignment as a phenomenon because we empirically show that it emerges naturally, without requiring ANY explicit orthogonality regularization in either: Loss formulation or Model architecture. So why this pheomena just naturally happens? This is where this works’ main contribution comes from.

We argue that this phenomenon improves scaling law in multi-modal model:

Hypothesis: Orthogonal Alignment improves scaling law in multi-modal model.

By ensuring that updates occupy subspace orthogonal to the input query, the model gains new representational capacity without needing more parameters.

We compare two approaches:

  1. Baseline + GCA module
  2. Parameter-augmented baseline (simply increasing parameters)

For instance, suppose the baseline model has 2 M parameters and the GCA module adds 0.5 M. To make the comparison fair, we also evaluate a parameter-augmented baseline with 2.5 M parameters—matching the total parameter count of the GCA-enhanced model.

We observed that the Baseline + GCA consistently outperformed the parameter-augmented baseline, demonstrating that the performance gain comes from orthogonal alignment rather than mere model scaling (see Figure 3).

In Figure 3, Baseline + GCAearly refers to inserting a single GCA module at the early stage of the model, while Baseline + GCAstack denotes stacking multiple GCA modules vertically throughout the network—from early to later layers.

cdsrnp
(a) CDSRNP
abxi
(b) ABXI
abxi
(c) ABXI
llm4cdsr
(d) LLM4CDSR
llm4cdsr
(e) LLM4CDSR
Figure 3: NDCG@10 comparison between baseline and baseline + gated cross attention model

First, our results show that across all five experimental cases, the addition of baseline with GCA_early consistently yields higher single-domain ranking performance (Domain A’s NDCG@10) compared to parameter-matched baselines, while Domain B’s NDCG@10 also shows general improvement.

Moreover, in both LLM4CDSR settings, GCAearly demonstrates the strongest parameter efficiency. We attribute this advantage to the fixed hidden dimensionality of the initial embedding vectors inherited from the pretrained LLM, which constrains the representational capacity of the baseline model. As a result, simply scaling up the baseline parameters eventually leads to performance saturation—and in some cases, degradation—as model size increases.

In contrast, introducing orthogonal alignment through GCA enables more effective information extraction under limited representational capacity. This property allows GCA to achieve a superior accuracy-per-parameter trade-off, demonstrating a more efficient use of model capacity.

Concluding Remark: Toward Vision–Language Generalization

We remain cautious about generalizing our findings to vision–language models, since all of our experiments on Orthogonal Alignment were conducted exclusively with recommendation data. Nonetheless, we are optimistic that similar phenomena could emerge in vision–language settings, given that our study also relies on transformer architectures with gated cross-attention—a core component in many multi-modal models.

The key distinctions between our setting and typical vision–language architectures are as follows:

This difference matters because most vision–language encoders are trained using self-contrastive objectives, which explicitly encourage high cosine similarity between matching image–text pairs and low similarity between mismatched ones. As a result, their latent representations are already well-aligned before cross-attention is applied—potentially making orthogonal alignment less pronounced or more difficult to observe directly.

Therefore, while we expect Orthogonal Alignment to exist in vision–language models, it may manifest under more subtle and nuanced conditions, reflecting the already pre-aligned nature of their learned embeddings.

Fore more information, please check paper📗


References:

¹ LLM4CDSR: Liu, Qidong, et al. “Bridge the Domains: Large Language Models Enhanced Cross-domain Sequential Recommendation.” Proceedings of the 48th International ACM SIGIR Conference on Research and Development in Information Retrieval. 2025.

² CDSRNP: Li, Haipeng, et al. “Cross-Domain Sequential Recommendation via Neural Process.” arXiv preprint arXiv:2410.13588 (2024).

³ ABXI: Bian, Qingtian, et al. “ABXI: invariant interest adaptation for task-guided cross-domain sequential recommendation.” Proceedings of the ACM on Web Conference. 2025.