• Ephera@lemmy.ml
    link
    fedilink
    English
    arrow-up
    24
    arrow-down
    1
    ·
    17 hours ago

    Okay, but just to be clear, the problem is not that it can’t do a timer. The problem is that it claims to be able to and even produces a result which looks plausible. It means, you cannot trust it to do anything that you can’t easily verify. If they could fix that overconfidence in a year, it would be much better.

    • fox [comrade/them]@hexbear.net
      link
      fedilink
      English
      arrow-up
      14
      ·
      17 hours ago

      The overconfident tone is baked in. LLMs don’t have knowledge or world models, and all text they produce is nothing more than statistical relation of input to output based on frequency of appearance and semantic closeness. Therefore you can train the things to lean towards doubtfulness (nobody will use them) or confidence (wow, it must be true if it’s this certain). It’s abusing the human tendency to anthropomorphize to sell a really shitty product.

      • wheezy@lemmy.ml
        link
        fedilink
        English
        arrow-up
        7
        ·
        16 hours ago

        What if we just, idk, handled those corner cases with something like a human created control system that follows a set of very specific instructions that always produce the same result.

        Stick with me here. I know this is a radical idea. But, say you were able to parse the input from the user and map it to the same resulting, let’s call it, function.

        So, the user says something like “start a timer for 60 seconds” or “60 second timer please”. Using a basic word mapping we could infer the confidence of English sentences and produce results.

        We could even improve our results through automatic user feedback based on behavior and popularity of their mapping choices. Yes.

        We could even do this for like multiple “features”. Like have one “function” that maps requests to timers, another to setting an alarm, maybe even something radical like doing mathematical computations.

        But, again, instead of throwing the input into a block box that burns massive compute power that we have no control of. We just. Write the box ourselves for very common tasks.

        Idk, maybe I’m crazy. It probably wouldn’t work. I’m probably just oversimplifying it.