Inverse Reinforcement Learning (IRL) aims to retrieve reward function from expert demonstrations.
Recently, Optimal Transport (OT) methods have been effective but come with complexities.
A new study challenges the necessity of OT in IRL with two simpler alternatives: Minimum-Distance Reward and Segment-Matching Reward.
Extensive evaluations show these simple methods match or surpass recent OT-based approaches, indicating a reevaluation of complexity in future IRL design.