ViTHL
ViTHL is the first hybrid research entry in the repository.
The research record owns canonical metadata such as title, venue, and authors. This writeup provides the narrative context needed for a readable public page.
A hybrid localization approach combining vision transformers with traditional geometric methods for robust self-localization of humanoid robots in RoboCup competition environments.
RoboCup Symposium 2025
S. Khatibi, A. Rahmani, A. Azadfar, V. P. da Fonseca, T. E. A. Oliveira
Robust self-localization is a critical capability for autonomous humanoid robots competing in dynamic environments. We present ViTHL, a hybrid localization architecture that combines the representational power of vision transformers with geometric constraints for accurate and reliable robot pose estimation. Our approach leverages pre-trained ViT models for feature extraction from onboard camera feeds, fused with traditional odometry signals through a probabilistic filter. Evaluated on humanoid robots in RoboCup competition scenarios, ViTHL demonstrates improved localization accuracy over baseline methods under challenging lighting and field conditions.
ViTHL is the first hybrid research entry in the repository.
The research record owns canonical metadata such as title, venue, and authors. This writeup provides the narrative context needed for a readable public page.