Video-based person recognition achieves robust identification by integrating face, body, and gait. However, current systems waste computational resources by processing all modalities with fixed heavyweight ensembles regardless of input complexity. To address these limitations, we propose IDSelect, a reinforcement learning-based cost-aware selector that chooses one pre-trained model per modality per-sequence to optimize the accuracy-efficiency trade-off. Our key insight is that an input-conditioned selector can discover complementary model choices that surpass fixed ensembles while using substantially fewer resources. IDSelect trains a lightweight agent end-to-end using actor-critic reinforcement learning with budget-aware optimization. The reward balances recognition accuracy with computational cost, while entropy regularization prevents premature convergence. At inference, the policy selects the most probable model per modality and fuses modality-specific similarities for the final score. Extensive experiments on challenging video-based datasets demonstrate IDSelect's superior efficiency: on CCVID, it achieves 95.9% Rank-1 accuracy with 92.4% less computation than strong baselines while improving accuracy by 1.8%; on MEVID, it reduces computation by 41.3% while maintaining competitive performance.
IDSelect introduces an RL-based cost-aware agent that dynamically selects one pre-trained model per modality for each input, discovering complementary model combinations that outperform static ensembles while using significantly fewer computational resources.
Selection frequency of each model across face, gait, and body modalities on the CCVID dataset. IDSelect learns to select complementary models across modalities, demonstrating input-adaptive selection behavior.
@article{ji2026idselect,
title={IDSelect: An RL-based Cost-Aware Selection Agent for Video-based Person Recognition},
author={Ji, Yuyang and Shen, Yixuan and Nguyen, Kien and Zhou, Lifeng and Liu, Feng},
journal={arXiv preprint},
year={2026}
}