skip to content
figure
Xu  Pan 潘旭 M.Sc. Student at Wuhan University

Hello World!

Hi, I am a M.Sc. student at Wuhan University logoWuhan University, working at the LIESMARS logoState Key Laboratory of Information Engineering in Surveying, Mapping and Remote Sensing(LIESMARS), under the guidance of Prof. Xianwei Zheng. My research focuses on embodied intelligence and 3D visual perception, with an emphasis on how spatial representations support generalizable decision-making and agent-centric policy learning.

My current work lies at the intersection of computer vision, reinforcement learning, and generative modeling, where I study how 2D and 3D representations can be unified to enable robust perception–action coupling. I am particularly interested in structure-aware visual representations that support cross-view understanding, generalization across environments, and interaction-driven learning in embodied settings.

Previously, I explored generative AI for image and video synthesis during my internship at Baidu logoBaidu. I am currently a remote research intern at the A*STAR logoCentre for Frontier AI Research (CFAR), Agency for Science, Technology and Research (A*STAR), supervised by Dr. Xingrui Yu, where I work on generalizable reinforcement learning for embodied agents, with a focus on agent-centric formulations and transferable policies grounded in implicit spatial representations that support generalization across tasks and scenes.

More broadly, my goal is to develop spatially grounded learning frameworks that bridge perception, geometry, and control, advancing the next generation of embodied systems that can reason about and act within complex real-world environments.

Find me on

News

Experiences


Acknowledgements:
I’m grateful to my collaborators and mentors for their guidance and support, especially
Prof. Xianwei Zheng, Prof. Hanjiang Xiong, Dr. Xingrui Yu (A*STAR), Dr. Zimin Xia (EPFL), Dr. Yan Zhang (Baidu),
and my colleagues/peers including
Zhenglin Wan (NUS), Jiashen Huang (NTU), Qiyuan Ma, Jintao Zhang, Chenyu Zhao, Ziqong Lu (HKU)
and others I’ve had the pleasure to work with.

Publications

Scale-aware Co-visible Region Detection for Image Matching

Xu Pan, Zimin Xia, Xianwei Zheng*

ISPRS Journal of Photogrammetry and Remote Sensing 2025
Matching images with significant scale differences remains a persistent challenge in photogrammetry and remote sensing. The scale discrepancy often degrades appearance consistency and introduces uncertainty in keypoint localization. While existing methods address scale variation through scale pyramids or scale-aware training, matching under significant scale differences remains an open challenge. To overcome this, we address the scale difference issue by detecting co-visible regions between image pairs and propose SCoDe (Scale-aware Co-visible region Detector), which both identifies co-visible regions and aligns their scales for highly robust, hierarchical point correspondence matching. Specifically, SCoDe employs a novel Scale Head Attention mechanism to map and correlate features across multiple scale subspaces, and uses a learnable query to aggregate scale-aware information of both images for co-visible region detection. In this way, correspondences can be established in a coarse-to-fine hierarchy, thereby mitigating semantic and localization uncertainties. Extensive experiments on three challenging datasets demonstrate that SCoDe outperforms state-of-the-art methods, improving the precision of a modern local feature matcher by 8.41%. Notably, SCoDe shows a clear advantage when handling images with drastic scale variations.

SAMatcher: Segment Anything Co-visible for Robust Feature Matching

Xu Pan, Qiyuan Ma, Jintao Zhang, Xianwei Zheng*

(In Preparation) 2026

SG-VLA: Spatially Grounded Vision-Language-Action Learning via Dense Flow Policy Optimization

Xu Pan, Zhenglin Wan, Xingrui Yu*

(In Preparation) 2026

Research on Large-Scale Disparity Image Matching Method Guided by Co-Visible Region

Xu Pan, Xianwei Zheng*

Master's Thesis 2026

The Institutional Filter: How Trust Shapes Inequalities Between Domestic and Global AI Models

Jiashen Huang, Xu Pan

(Under Review) 2026
Artificial intelligence is increasingly woven into the way people communicate, think, and make decisions. Yet trust in AI does not grow evenly across contexts; it carries traces of national identity, institutional credibility, and emotional attachment. This study examines how institutional trust shapes user trust in domestic (DeepSeek) and global (ChatGPT) large language models (LLMs) in China. Specifically, it distinguishes between cognitive and affective dimensions of trust. Using survey data from 405 participants, we found that higher institutional trust strengthens emotional confidence in domestic AI models, while at low levels of institutional trust, this domestic advantage in perceived competence disappears. By examining the relationship between institutional trust and AI adoption, this study deepens theoretical insights into global communication inequalities in the digital era. The findings suggest that institutional trust operates as a social resource, channeling legitimacy into technological trust, thus contributing to the uneven distribution of trust in AI technologies across different societal groups. The findings offer policy insights for inclusive AI governance and the promotion of global technological equity.

Personal Philosophy

I follow Stoic philosophy. Life is a joyful ascent: a true mountaineer delights in the climb itself, not just the summit.

“Thou sufferest this justly: for thou choosest rather to become good to-morrow than to be good to-day.”
— Marcus Aurelius, Meditations 8.22

I resonate with the spirit of Slow Science.

We live in an age tyrannized by efficiency, outcomes, and speed, to the point that nothing lasts and nothing leaves a deep impression. In the midst of noisy bubbles and short-lived hype, I hope to take time to think carefully, to doubt, to refine, and to do research that is genuinely meaningful and worth remembering.