The architectures we know of so far have not been sufficient to achieve AGI with just text and image data. Humans and higher animals learn with much richer modalities than those two and probably would not be nearly as intelligent if forced to learn with just text and images. There are already ongoing efforts to train models with other modalities. Latest foundation models already go beyond pure LLMs.
Your reasoning above doesn’t mean some improvements to the current architecture(s) coupled with richer data would not be sufficient to achieve AGI.
There’s also a possibility OpenAI has recently achieved a yet undisclosed breakthrough.
Sam Altman at the APEC Summit yesterday:
"4 times now in the history of OpenAI — the most recent time was just in the last couple of weeks — I’ve gotten to be in the room when we push the veil of ignorance back and the frontier of discovery forward”
Your reasoning above doesn’t mean some improvements to the current architecture(s) coupled with richer data would not be sufficient to achieve AGI.
There’s also a possibility OpenAI has recently achieved a yet undisclosed breakthrough.
Sam Altman at the APEC Summit yesterday:
"4 times now in the history of OpenAI — the most recent time was just in the last couple of weeks — I’ve gotten to be in the room when we push the veil of ignorance back and the frontier of discovery forward”
https://twitter.com/SpencerKSchiff/status/172564613068224524...