Reward Shaping for Inference-Time Alignment: A Stackelberg Game Perspective Jul 6, 2026ยท Haichuan Wang , Tao Lin , Lingkai Kong , Ce Li , Hezi Jiang , Milind Tambe ยท 0 min read PDF Last updated on Jul 6, 2026 LLM AI Alignment Principal-Agent Problem ← On the Coordination of Value-Maximizing Bidders Jul 6, 2026 Learning a Game by Paying the Agents Apr 24, 2026 →