Back to blog

Better AI Vision: Uncertainty Guides Hierarchical Understanding

Based on research by Hayeon Kim, Ji Ha Jang, Junghun James Kim, Se Young Chun

Better AI Vision: Uncertainty Guides Hierarchical Understanding

New models finally stop treating every image part as equal, unlocking accurate "part-to-whole" reasoning previously impossible for vision systems.

Researchers introduced UNCHA, a framework that assigns variable uncertainty weights to image regions based on how semantically representative they are of the entire scene. While current models struggle with complex compositions, this approach uses hyperbolic geometry—a mathematical structure designed for nested hierarchies—to learn which parts define an object and which are merely background noise.

The system works by assigning lower uncertainty values to critical components, allowing the model to prioritize them during training through a novel contrastive objective. This calibration is further refined using an entailment loss that balances entropy, ensuring the AI builds a precise internal map of how fragments relate to the whole.

The result is a Vision-Language Model that grasps multi-object scenes with state-of-the-art precision, significantly outperforming Euclidean baselines in zero-shot classification and retrieval tasks.

Source: Uncertainty-guided Compositional Alignment with Part-to-Whole Semantic Representativeness in Hyperbolic Vision-Language Models by Hayeon Kim et al., arXiv:2603.22042

Source: arXiv:2603.22042

This post was generated by staik AI based on the academic publication above.