Yes, the goal posts keep moving, but they do so for a rather solid reason: We humans are famously bad at understanding intelligence and at understanding the differences between human and computer intelligence.
100 years ago, doing complex calculations was seen as something very complex that only reasonably smart humans could do. Computers could easily outcompete humans, because calculations are inherently easy for computers while very difficult for humans.
30 years ago we thought that high-level chess was something reserved only to the smartest of humans, and that it was a decent benchmark for intelligence. Turns out, playing chess is something that benefits greatly from large memory and fast computations, so again, it was easy for computers while really hard for humans.
Nowadays AI can do a lot of things we thought would be really hard to do, but that computers can actually do. But there’s hardly any task performed by LLMs where it’s actually better than a moderately proficient human being. (Apart from tasks like “Do homework task X”, where again LLMs benefit from large memory since they can just regurgitate stuff from the training set.)
Yes, the goal posts keep moving, but they do so for a rather solid reason: We humans are famously bad at understanding intelligence and at understanding the differences between human and computer intelligence.
100 years ago, doing complex calculations was seen as something very complex that only reasonably smart humans could do. Computers could easily outcompete humans, because calculations are inherently easy for computers while very difficult for humans.
30 years ago we thought that high-level chess was something reserved only to the smartest of humans, and that it was a decent benchmark for intelligence. Turns out, playing chess is something that benefits greatly from large memory and fast computations, so again, it was easy for computers while really hard for humans.
Nowadays AI can do a lot of things we thought would be really hard to do, but that computers can actually do. But there’s hardly any task performed by LLMs where it’s actually better than a moderately proficient human being. (Apart from tasks like “Do homework task X”, where again LLMs benefit from large memory since they can just regurgitate stuff from the training set.)