| SpatialBot: Precise Spatial Understanding with Vision Language Models | Munir | 2026/4/13 |
| Dur360BEV: A Real-world 360-degree Single Camera Dataset and Benchmark for Bird-Eye View Mapping in Autonomous Driving | 桂茹 | 2026/4/13 |
| Domain Adaptation-Based Crossmodal Knowledge Distillation for 3D Semantic Segmentation | 旻璇 | 2026/4/20 |
| Towards Robust Autonomous Driving: Conditional Multimodal Large Language Models for Fine-Grained Perception | 新哲 | 2026/4/20 |
| DP-Habitat: Bridging the Gap Between Simulation and Reality for Visual Navigation in Dynamic Pedestrian Environments | 宇廷 | 2026/4/27 |
| A Comprehensive LLM-powered Framework for Driving Intelligence Evaluation | 汶璇 | 2026/4/27 |
| Safety and Naturalness Perceptions of Robot-to-Human Handovers Performed by Data-Driven Robotic Mimicry of Human Givers | 粲儒 | 2026/5/04 |
| Vision Transformers for End-to-End Vision-Based Quadrotor Obstacle Avoidance | 騏緯 | 2026/5/04 |
| VLN-KHVR: Knowledge-and-History Aware Visual Representation for Continuous Vision-and-Language Navigation | 淳傑 | 2026/5/11 |
| GND: Global Navigation Dataset with Multi-Modal Perception and Multi-Category Traversability in Outdoor Campus Environments | 樂業 | 2026/5/11 |