

Against my better judgement I got into an argument with a promptfan on Bluesky. To his credit, aside from the usual boring arguments (“models are getting better, and better”, “have you tried model xyz”, “everyone not using chatbots will be left in the dust” he provided an actual example.
https://github.com/dfed/SafeDI/issues/183 It’s a bug that’s supposedly easy to test, but hard to reason about. Took the chatbot half an hour while it would take him several (allegedly).
Now, my first thought was: “If a clanker could do it (something that famously can’t reason) then it couldn’t be that hard to reason about.”
But I was curious so I looked. Unfortunately it is an area I’m not familiar with and in a language (Swift) I don’t know at all.
Probably should file the claim under “not true or false” and touch grass or something, but it’s bugging me.
Any one y’all who could say if there’s something interesting in there?


Thanks so much! Now I can waste my life on more interesting things…