Lecturer | Researcher in Multimodal Data Fusion School of Computer Science, Guangdong Polytechnic Normal University
I am currently a Lecturer at the School of Computer Science, Guangdong Polytechnic Normal University. I received my Ph.D. degree in Electronics Information from Sun Yat-sen University.
My primary research interests lie at the intersection of Computer Vision and Multimodal Data Fusion, with specific applications deployed in autonomous driving, remote sensing, and vehicle-infrastructure cooperation (V2X). I am dedicated to developing robust perception systems for complex, real-world open traffic scenarios.
I maintain active and longβterm scientific collaborations with researchers from Sun Yat-sen University, the University of Cambridge, Peng Cheng Laboratory, and Robert Gordon University.
Note: My research philosophy strongly supports open science. Datasets and codebases associated with my publications are made publicly available where possible. For a comprehensive and up-to-date list of my publications, please refer to my Google Scholar Profile.
-
[CVPR '26] SceneBench: Evaluating and Enhancing Long Video Understanding via Scene-Level Context. Seng Nam Chen, Hao Chen, Chenglam Ho, Xinyu Mao, Jinping Wang, Yu Zhang, Chao Li.
Introduces SceneBench to evaluate scene-level long video understanding in VLMs, and proposes Scene-RAG to effectively mitigate long-context forgetting via dynamic scene memory. π Paper
-
[IEEE TMM '26] Cognidrive: Cognitive Autonomous Driving Understanding with Multistep Multimodal Chain-of-Thought Reasoning. Xiangyi Qin, Xiaofei Zhang, Shuai Wang, Yuzhen Wei, Jinping Wang*, Xiaojun Tan.
A large-scale dataset featuring multi-position LiDARs in a real-world setting, addressing occlusion challenges within I2I perception systems. π Accepted
-
[Information Fusion '26] Inscope: A new real-world 3d infrastructure-side collaborative perception dataset for open traffic scenarios, Xiaofei Zhang#, Yining Li#, Jinping Wang#, Xiangyi Qin, Ying Shen, Zhengping Fan, Xiaojun Tan
A large-scale dataset featuring multi-position LiDARs in a real-world setting, addressing occlusion challenges within I2I perception systems. π Paper π Dataset & Code
-
[ADVEI '26] Confidence-V2X: Confidence-driven sparse communication for efficient V2X cooperative perception. Xiaojun Tan, Rui Wang, Jinping Wang*, Shuai Wang, Xu Wang, Dongsheng Wu.
A confidenceβaware cooperative perception framework designed to jointly optimize object detection performance and communication efficiency in V2X systems. π Paper π Code
-
[IEEE TGRS '25] FusDreamer: Label-efficient remote sensing world model for multimodal data classification. Jinping Wang, Weiwei Song, Hao Chen, Jinchang Ren, Huimin Zhao.
A label-efficient remote sensing world model for multimodal data fusion, exploring the potential of the world model in the RS field. π Paper π Code
-
[IEEE GRSL '25] CaPaT: Cross-Aware Paired-Affine Transformation for Multimodal Data Fusion Network. Jinping Wang, Hao Chen, Xiaofei Zhang, Weiwei Song.
Introduces a direct feature interaction paradigm to improve the transfer efficiency of feature fusion while significantly reducing model parameters. π Paper
-
[ICASSP '24] BEVLOC: End-to-end 6-dof localization via cross-modality correlation under birdβs eye view. Nanjie Chen, Jinping Wang, Hao Chen, Ying Shen, Shuai Wang, Xiaojun Tan
An end-to-end approach for vehicle localization that fuses monocular image and LiDAR map features in the BEV space via optical flow-based cross-modality correlation. π Paper
-
[IEEE TCSVT '23] Mutually beneficial transformer for multimodal data fusion. Jinping Wang, Xiaojun Tan.
Introduces dynamic region-aware convolution for spatial guide mask generation and elevation salience agent guidance. π Paper
-
[IEEE TCSVT '22] AMΒ³Net: Adaptive mutual-learning-based multimodal data fusion network. Jinping Wang, Jun Li, Yanli Shi, Jianhuang Lai, Xiaojun Tan.
Collaborative feature transmission focusing on the specificity of HSI spectral channels and the complementarity of HSI and LiDAR spatial information. π Paper π Code
-
[IEEE ICASSP '22] Spectral-spatial symmetrical aggregation cross-linking multi-modal data fusion network. Jinping Wang, Jun Li, Xiaojun Tan.
Develops a SACLNet utilizing involution operations and pyramid feature fusion for robust multi-modal data classification. π Paper
-
[Neurocomputing '22] ASPCNet: Deep adaptive spatial pattern capsule network for hyperspectral image classification. Jinping Wang, Xiaojun Tan, Jianhuang Lai, Jun Li.
An adaptive spatial pattern capsule network architecture based on an enlarged, semantically-adaptive receptive field. π Paper π Code
-
[Electronics '21] A simulated annealing algorithm and grid map-based UAV coverage path planning method for 3D reconstruction. Sichen Xiao, Xiaojun Tan, Jinping Wang.
Proposes a UAV CPP framework considering both image overlapping and energy efficiency, validated through site experiments. π Paper
-
[IEEE TGRS '19] Spatial density peak clustering for hyperspectral image classification with noisy labels. Bing Tu, Xiaofei Zhang, Xudong Kang, Jinping Wang, JΓ³n Atli Benediktsson.
Proposes a spatial density peak (SDP) clustering-based method to detect and handle mislabeled samples in HSI training sets. π Paper π Code
-
[IEEE JSTARS '19] Texture pattern separation for hyperspectral image classification. Bing Tu#, Jinping Wang#, Guoyun Zhang, Xiaofei Zhang, Wei He.
Addresses the layer-separation problem in HSI via a novel TPS feature extraction method. π Paper π Code
-
[IEEE JSTARS '18] KNN-based representation of superpixels for hyperspectral image classification. Bing Tu#, Jinping Wang#, Xudong Kang, Guoyun Zhang, Xianfeng Ou, Longyuan Guo.
Explores optimal representations of superpixels using two k-selection rules to find the most representative samples. π Paper π Code
-
[IEEE GRSL '18] Hyperspectral image classification via fusing correlation coefficient and joint sparse representation. Bing Tu, Xiaofei Zhang, Xudong Kang, Guoyun Zhang, Jinping Wang, Jianhui Wu.
A hyperspectral image classification method via fusing correlation coefficient and joint sparse representation. π Paper π Code
- Cimy_PPtools: A comprehensive Python toolbox designed for data preprocessing in hyperspectral classification and fusion tasks, supporting model serialization and result visualization.
Core Technical Stack:
Python | MATLAB | Shell | LaTeX
βοΈ Open to scientific cooperation and academic inquiries. Please feel free to reach out.

