AI Safety Compendium

Home

❯

summaries

❯

Large Reasoning Models Learn Better Alignment from Flawed Thinking

Large Reasoning Models Learn Better Alignment from Flawed Thinking

27 Apr 20261 min read

Large Reasoning Models Learn Better Alignment from Flawed Thinking

ShengYun Peng, Eric Smith, Ivan Evtimov, Song Jiang, Pin-Yu Chen, Hongyuan Zhan, … (+4 more) — 2025-10-01

Source

  • Link: https://arxiv.org/pdf/2510.00938%20
  • Listed in the Shallow Review of Technical AI Safety 2025 under 1 agenda(s):
    • meta — Labs (giant companies)

Related Pages

  • meta

Graph View

Graph view

The interactive citation graph is desktop-only. Visit this page on a larger screen to explore how concepts, agendas, papers, and organisations link together.

  • Large Reasoning Models Learn Better Alignment from Flawed Thinking
  • Source
  • Related Pages

Created with Quartz v0.1.0 © 2026

  • Suggest a source
  • Connect
  • Overview
  • About (proof of concept)
  • Email feedback
  • Made by IT for Humanity