Can LLMs SAT?

7 points by aiono


gignico

The simple answer is of course "no" because SAT is NP-complete and the inferring process of LLMs does not take exponential time over the length of the prompt. So, unless P = NP, any LLM can only either guess the results blindly (i.e. not with a specific probability of success greater than 0.5, because SAT is not known to be in BPP either) or tell the correct answer only for a fixed subset of learned examples.

Funnily enough this is what is happening with the different LLMs tested in the post. GPT 5.2 is randomly guessing correctly half of the inputs, and GPT 5.2 mini works under 100 clauses.