Hazzard ,

I don't necessarily disagree that we may figure out AGI, and even that LLM research may help us get there, but frankly, I don't think an LLM will actually be any part of an AGI system.

Because fundamentally it doesn't understand the words it's writing. The more I play with and learn about it, the more it feels like a glorified autocomplete/autocorrect. I suspect issues like hallucination and "Waluigis" or "jailbreaks" are fundamental issues for a language model trying to complete a story, compared to an actual intelligence with a purpose.

  • All
  • Subscribed
  • Moderated
  • Favorites
  • [email protected]
  • kbinchat
  • All magazines