

0·
3 months agoDon’t take this as a defense of him, but rather a statement that there are better, more credible arguments to push. From what I can gather the entirety of the evidence “linking” is a phone call making the claim. It’s not even clear the baby was real at all. If it’s true, this isn’t evidence enough, and is effectively a red herring and a self made straw man. Don’t take the bait, use better arguments.

If you’re honestly asking, LLMs are much better at coding than any other skill right now. On one hand there’s a ton of high quality open source training data that appropriated, on the other code is structured language so is very well suited for what models “are”. Plus, code is mechanically verifiable. If you have a bunch of tests, or have the model write tests, it can check its work as it goes.
Practically, the new high end models, GPT 5.4 or Claude Opus 4.6, can write better code faster than most people can type. It’s not like 2 years ago when the code mostly wouldn’t build, rather they can write hundreds or thousands of lines of code that works first try. I’m no blind supporter of AI, and it’s very emotionally complicated watching it after years honing the craft, but for most tasks it’s simple reality that you can do more with AI than without it. Whether it’s higher quality, higher volume, or integrating knowledge you don’t have.
Professionally I don’t feel like I have a choice, if I want to stay employed in the field at least.