This is bad, but the report demonstrates a common, fundamental misunderstanding of these systems. The LLM did not commit "fraud"; it can no more commit fraud than a spreadsheet can commit fraud. It did not "know it was lying" at the time it output a series of tokens; it doesn't "know" anything.
add a skeleton here at some point
about 2 months ago