• skuzz@discuss.tchncs.de
    link
    fedilink
    arrow-up
    3
    ·
    1 day ago

    “Thinking” mode is just sending wave upon wave of GPUs at the problem until the killbots hit their pre-set kill count. One could roughly simulate that by not using thinking mode and just feeding the answer and question back to the LLM repeatedly until it eventually gets an answer that might be “right”. These companies have hit a technological wall with LLMs and will do anything to attempt to look like they still have forward inertia.

    • jj4211@lemmy.world
      link
      fedilink
      arrow-up
      4
      ·
      1 day ago

      Well, not quite, because they don’t have criteria for ‘right’.

      They do basically say ‘generate 10x more content than usual, then dispose of 90% of it’, and that surprisingly seems to largely improve results, but at no point is it ‘grading’ the result.

      Some people have bothered to provide ‘chain of thought’ examples and even when it’s largely ‘correct’, you may see a middle step be utterly flubbed in a way that should have fouled the whole thing, but the error is oddly isolated and doesn’t carry forward into the subsequent content, as would be the case in actual ‘reasoning’.