Discussion about this post

User's avatar
Lakshmi Narasimhan's avatar

the Priya example nails it. the finance friend evaluated the email in isolation. the experienced coworker simulated how it would land in Priya's inbox, against her triage heuristics, under deadline pressure

this is the gap between LLMs writing code and LLMs building systems. code that compiles isn't code that survives contact with users, adversaries, edge cases

been running production systems solo for 20 years. the best operators aren't the ones who know the most commands — they're the ones who can simulate what will break next. "if I do X, the cache invalidates, which triggers Y, which overloads Z." that's a world model

the poker vs chess analogy is perfect. hidden state + adversarial adaptation = can't just pattern match

curious where you see the fix coming from. is it more training on game theory scenarios? or do we need fundamentally different architectures to track hidden state?

Graham's avatar

This is a great framework. LLMs can regurgitate game theory but can’t yet ‘feel’ the move/counter move process. They are great at predicating the next word but perhaps not yet predicting the next emotional state. Makes me wonder if the next frontier is capturing a better understanding of human behavioral psychology and being able to relate to adversarial circumstance.

Do you think we can get there with text alone? You might make the case that art could be useful in this sense: drama, etc.

What is Shakespeare if not a picture of clashing wills.. type of the thing.

2 more comments...

No posts

Ready for more?