Yes, thar is the same feelingg I have.
Giving it some json and describe how a website should look? Super fast results and amazing capabilities.
Trying to get it to translate my unit tests from Xunit to Tunit, where the latter is new and does not have a ton of blog posts? Forget about it. The process is purely mechanical and it is easy after RTFM, but it falls flat on its face
Although I think if you asked people 20 years ago to describe a test for something AGI would do, they would be more likely to say “writing a poem” or “making art” than “turning Xunit code to Tunit”
IMO I think if you said to someone in the 90s “well we invented something that can tell jokes, make unique art, write stories and hold engaging conversations, although we haven’t yet reached AGI because it can’t transpile code accurately - I mean it can write full applications if you give it some vague requirements, but they have to be reasonably basic, like only the sort of thing a junior dev could write in a day it can write in 20 seconds, so not AGI” they would say “of course you have invented AGI, are you insane!!!”.
LLMs to me are still a technology of pure science fiction come to life before our eyes!
Tell them humans need to babysit it and doublecheck its answers to do anything since it isn't as reliable as a human then no they wouldn't call it an AGI back then either.
The whole point about AGI is that it is general like a human, if it has such glaring weaknesses as the current AI has it isn't AGI, it was the same back then. That an AGI can write a poem doesn't mean being able to write a poem makes it an AGI, its just an example the AI couldn't do 20 years ago.
Why do human programmers need code review then if they are intelligent?
And why can’t expert programmers deploy code without testing it? Surely they should just be able to write it perfectly first time without errors if they were actually intelligent.
> Why do human programmers need code review then if they are intelligent?
Human programmers don't need code reviews, they can test things themselves. Code reviews is just an optimization to scale up it isn't a requirement to make programs.
Also the AGI is allowed to let another AGI code review it, the point is there shouldn't be a human in the loop.
> And why can’t expert programmers deploy code without testing it?
Testing it can be done by themselves, the AGI model is allowed to test its own things as well.
Well AGI can write unit tests, write application code then run the tests and iterate - agents in cursor are doing this already.
Just not for more complex applications.
Code review does often find bugs in code…
Put another way, I’m not a strong dev but good LLMs can write lots of code with less bugs than me!
I also think it’s quite a “programmer mentality” that most of the tests in this forum about if something is/isn’t AGI ultimately boils down to if it can write bug-free code, rather than if it can negotiate or sympathise or be humerous or write an engaging screen play… I’m not saying AGI is good at those things yet, but it’s interesting that we talk about the test of AGI being transpiling code rather than understanding philosophy.
> Put another way, I’m not a strong dev but good LLMs can write lots of code with less bugs than me!
But the AI still can't replace you, it doesn't learn as it go and therefore fail to navigate long term tasks the way humans do. When a human writes a big program he learns how to write it as he writes it, these current AI cannot do that.