ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2506.12815
17
0

TrojanTO: Action-Level Backdoor Attacks against Trajectory Optimization Models

15 June 2025
Yang Dai
Oubo Ma
Longfei Zhang
Xingxing Liang
Xiaochun Cao
Shouling Ji
J. Zhang
Jincai Huang
Li Shen
ArXiv (abs)PDFHTML
Main:9 Pages
6 Figures
Bibliography:3 Pages
20 Tables
Appendix:11 Pages
Abstract

Recent advances in Trajectory Optimization (TO) models have achieved remarkable success in offline reinforcement learning. However, their vulnerabilities against backdoor attacks are poorly understood. We find that existing backdoor attacks in reinforcement learning are based on reward manipulation, which are largely ineffective against the TO model due to its inherent sequence modeling nature. Moreover, the complexities introduced by high-dimensional action spaces further compound the challenge of action manipulation. To address these gaps, we propose TrojanTO, the first action-level backdoor attack against TO models. TrojanTO employs alternating training to enhance the connection between triggers and target actions for attack effectiveness. To improve attack stealth, it utilizes precise poisoning via trajectory filtering for normal performance and batch poisoning for trigger consistency. Extensive evaluations demonstrate that TrojanTO effectively implants backdoor attacks across diverse tasks and attack objectives with a low attack budget (0.3\% of trajectories). Furthermore, TrojanTO exhibits broad applicability to DT, GDT, and DC, underscoring its scalability across diverse TO model architectures.

View on arXiv
@article{dai2025_2506.12815,
  title={ TrojanTO: Action-Level Backdoor Attacks against Trajectory Optimization Models },
  author={ Yang Dai and Oubo Ma and Longfei Zhang and Xingxing Liang and Xiaochun Cao and Shouling Ji and Jiaheng Zhang and Jincai Huang and Li Shen },
  journal={arXiv preprint arXiv:2506.12815},
  year={ 2025 }
}
Comments on this paper