Modality Alignment across Trees on Heterogeneous Hyperbolic Manifolds

1 Beijing Key Laboratory of Intelligent Information Technology, School of Computer Science & Technology,
Beijing Institute of Technology
2 Guangdong Laboratory of Machine Perception and Intelligent Computing, Shenzhen MSU-BIT University
3 Department of Electrical and Computer System Engineering, Monash University

ICLR 2026 (Poster)

*Indicates Equal Contribution

Abstract

Motivation figure

Modality alignment is critical for vision-language models (VLMs) to effectively integrate information across modalities. However, existing methods extract hierarchical features from text while representing each image with a single feature, leading to asymmetric and suboptimal alignment. To address this, we propose Alignment across Trees, a method that constructs and aligns tree-like hierarchical features for both image and text modalities. Specifically, we introduce a semantic-aware visual feature extraction framework that applies a cross-attention mechanism to visual class tokens from intermediate Transformer layers, guided by textual cues to extract visual features with coarse-to-fine semantics. We then embed the feature trees of the two modalities into hyperbolic manifolds with distinct curvatures to effectively model their hierarchical structures.

To align across the heterogeneous hyperbolic manifolds with different curvatures, we formulate a KL distance measure between distributions on heterogeneous manifolds, and learn an intermediate manifold for manifold alignment by minimizing the distance. We prove the existence and uniqueness of the optimal intermediate manifold. Experiments on taxonomic open-set classification tasks across multiple image datasets demonstrate that our method consistently outperforms strong baselines under few-shot and cross-domain settings.

Our Method

Pipeline overview

Experimental Results

TOS classification results on the 1-shot and 16-shot settings. We bold the best results.

K-Shot Base Method Variant Cifar100 SUN ImageNet Rare Species
LAHCAMTA LAHCAMTA LAHCAMTA LAHCAMTA
1 MaPLe Vanilla 68.754.6550.60 63.9825.1550.31 68.912.9748.16 41.555.0944.75
1 MaPLe +ProTeCt 69.3348.1083.36 64.2950.4576.73 66.1620.4485.18 39.9213.2270.04
1 MaPLe +Ours 71.3753.1985.29 67.5757.9280.55 66.3325.5685.98 46.7720.9476.83
1 PromptSRC Vanilla 72.4814.3651.91 70.5842.1457.19 68.824.4654.10 45.396.7244.72
1 PromptSRC +ProTeCt 73.0749.5485.16 70.6155.5278.73 68.4321.5885.63 44.5620.3674.42
1 PromptSRC +Ours 73.5451.9185.76 70.6457.7979.94 68.8625.1386.45 46.9823.0377.32
16 MaPLe Vanilla 75.0117.5452.21 71.8633.2554.29 70.704.1548.16 50.945.3040.41
16 MaPLe +ProTeCt 75.3461.1588.04 72.1759.7182.27 69.5231.2487.87 48.1424.8278.79
16 MaPLe +Ours 77.9269.3890.89 75.4768.6786.02 71.4143.7988.78 69.9653.6587.27
16 PromptSRC Vanilla 77.7115.0756.86 75.7545.2359.42 71.502.4846.71 59.2011.6455.82
16 PromptSRC +ProTeCt 78.7666.7490.79 75.5466.0184.75 70.9832.8988.31 56.4033.9282.47
16 PromptSRC +Ours 78.9068.4791.12 76.5469.1886.20 71.6742.2689.64 67.3850.7787.60

Learned Representations

Our method demonstrates improved feature separability across taxonomic categories.

Learned feature representations with improved separability across taxonomic categories

Attention Visualization

Our model adaptively generates semantic-aware visual features by attending to different regions corresponding to each taxonomic granularity (from coarse to fine, left to right).

Baseline

Baseline attention visualization across taxonomic granularity

Ours

Our model's semantic-aware attention visualization across taxonomic granularity

Citation & BibTeX

If you find our work useful, please cite:

@inproceedings{wu2026hypmodalalign,
  title={Modality Alignment across Trees on Heterogeneous Hyperbolic Manifolds},
  author={Wu, Wei and Fan, Xiaomeng and Wu, Yuwei and Gao, Zhi and Li, Pengxiang and Jia, Yunde and Harandi, Mehrtash},
  booktitle={International Conference on Learning Representations (ICLR)},
  year={2026}
}