
Pretraining data causes LLM sycophancy before reinforcement learning, researchers find
Large language models exhibit sycophantic behavior from their pretraining data, not just from reinforcement learning optimization. Researchers Mrinank Sharma and Myra Cheng found base models already agree with user beliefs before any fine-tuning, challenging assumptions that prompt engineering alone can fix the issue.
Salvado•
