학과 공지

  • 정보 센터
  • 학과 공지

자율주행 기술 관련 인공지능분야 최고 권위 학술지 T-PAMI 논문게재 승인_이유철 교수

  • AI자율주행시스템공학
  • 2022-10-14

이유철 교수의 Robotics and AI Navigation (RAIN) 실험실에서는 한국전자통신연구원(ETRI)과 미국 스토니브룩대학교(Stony Brook University)와 협업을 통하여 강화학습과 트랜스포머 모델을 이용한 종단간 모방학습 기반의 자율주행 기술에 대한 연구를 진행하였습니다.

그 결과 교신저자로 참여한 "StARformer: Transformer with State-Action-Reward Representations for Robot Learning" 논문은 인공지능분야 최고 권위 학술지인 IEEE Transactions on Pattern Analysis and Machine Intelligence(T-PAMI)(JCR 0.54%, Impact Factor 24.314)에 게재 승인을 받았으며, 12월에 출간 예정입니다.

  

 
 

 


 

 


Reinforcement Learning (RL) can be considered as a sequence modeling task, where an agent employs a sequence of past state-action-reward experiences to predict a sequence of future actions. In this work, we propose St ate- A ction- R eward Transformer ( StAR former), a Transformer architecture for robot learning with image inputs, which explicitly models short-term state-action-reward representations (StAR-representations), essentially introducing a Markovian-like inductive bias to improve long-term modeling. StARformer first extracts StAR-representations using self-attending patches of image states, action, and reward tokens within a short temporal window. These StAR-representations are combined with pure image state representations, extracted as convolutional features, to perform self-attention over the whole sequence. Our experimental results show that StARformer outperforms the state-of-the-art Transformer-based method on image-based Atari and DeepMind Control Suite benchmarks, under both offline-RL and imitation learning settings. We find that models can benefit from our combination of patch-wise and convolutional image embeddings. StARformer is also more compliant with longer sequences of inputs than the baseline method. Finally, we demonstrate how StARformer can be successfully applied to a real-world robot imitation learning setting via a human-following task.