https://jameswillia.ms/posts/chatgpt-rot13.html [ ] [ ] James Williams [ ] * Home * Contact * Blog + Short URLs + ChatGPT vs a cryptic crossword + ChatGPT, rot13, and Daniel Kahneman * Guitar Tabs Menu ChatGPT, rot13, and Daniel Kahneman Table of Contents ChatGPT, rot13, and Daniel Kahneman I thought it would be interesting to see whether ChatGPT could solve some basic ciphers (Caesar, Vignere, etc.). I decided to start at the bottom, with perhaps the easiest possible cipher: rot13. I asked it to decode the rot-13 encoding of "Why did the chicken cross the road": [chatgpt-ro] So the first thing to note here is that ChatGPT is not able to solve the task, even for very small cases. But nonetheless, I found this exchange interesting. ChatGPT's approach to solving this brought to mind Daniel Kahneman's book Thinking, Fast and Slow. To me, it feels like ChatGPT is attempting to apply System 1 thinking (fast, instinctual) to a problem that requires a System 2 approach, and ends up getting exactly the results you'd expect from that: answers that looks vaguely sensible, but are actually way off base. As a human, when performing rot13 decoding, you apply System 2 thinking: working through each letter, doing calculations in your head to find the right letter - perhaps remembering common letter mappings as you go. But in this case, it feels like ChatGPT is approaching the problem like a language learner put on spot: going wordwise, hazily recognising words, and filling in the gaps from there. I wonder whether the distinction between the tasks ChatGPT is good at, and those that it isn't, is whether the task is more amenable to System 1 or System 2 thinking? When I think of things that ChatGPT has been observed to be poor at, for example: * evaluating complex mathematical equations * writing mathematical proofs * solving cryptic crossword clues all of them (generally) require some degree of System 2 thinking as a human. On the other hand, many of the tasks it is good at are things that humans use System 1 thinking to do: * translating languages (once fluent) * writing jokes * summarising text When it makes mistakes, ChatGPT also displays the kind of overconfidence that System 1 thinking results in. Interestingly, ChatGPT can kind of follow along with our System 2 reasoning (e.g: where I explained that it answering plaintext with a different number of letters from the ciphertext could not make sense, because rot13 is a one-to-one mapping). But it seems incapable of taking that reasoning and applying it again, even when the next application is very similar. James Williams * 2020