Language is really powerful, I think it's a huge part of our intelligence.
The interesting part of the article to me is the focus on fluency. I have not seen anything that LLMs do well that isn't related to powerful utilization of fluency.
I don't think changeable code is the number one priority. The goal is to solve a problem and code that solves a problem without needing to change is sufficient.
Code that doesn't need to change is a really good sign that you've got something good.
The real world moves. If your code didn't change it must be generating value against something that is very standard but I'd be very surprised that you're not modifying, adding, etc based on usage to derive even more value somehow.
> Code that doesn't need to change is a really good sign that you've got something good.
Not really, but you're thinking in terms of "my code nailed it because I'm not touching it" and that has NOTHING to do with it being easy to change.
In the article the author mentions wanting to benchmark a GPU and using ChatGPT to write CUDA. Benchmarks are easy to mess up and to interpret incorrectly without understanding. I see this as an example where a subtly-wrong idea could cause cascading problems.