Evidence that LLMs are reaching a point of diminishing returns - and what that might mean - eviltoast
  • huginn@feddit.it
    link
    fedilink
    arrow-up
    4
    arrow-down
    2
    ·
    7 months ago

    I think increasingly specialized models and analog systems that run them will be increasingly prevalent.

    LLMs at their current scales don’t do enough to be worth their enormous cost… And adding more data is increasingly difficult.

    That said: the gains on LLMs have always been linear based on recent research. Emergence was always illusory.

    • ericjmorey@discuss.online
      link
      fedilink
      English
      arrow-up
      2
      arrow-down
      1
      ·
      7 months ago

      I’d like to read the research you alluded to. What research specifically did you have in mind?

      • huginn@feddit.it
        link
        fedilink
        arrow-up
        2
        ·
        7 months ago

        Sure: here’s the article.

        https://arxiv.org/abs/2304.15004

        The basics are that:

        1. LLM “emergent behavior” has never been consistent, it has always been specific to some types of testing. Like taking the SAT saw emergent behavior when it got above a certain number of parameters because it went from missing most questions to missing fewer.

        2. They looked at the emergent behavior of the LLM compared to all the other ways it only grew linearly and found a pattern: emergence was only being displayed in nonlinear metrics. If your metric didn’t have a smooth t transition between wrong, less wrong, sorta right, and right then the LLM would appear emergent without actually being so.