• Sundray@lemmus.org
    link
    fedilink
    English
    arrow-up
    74
    ·
    1 day ago

    If AI does not yet work for a particular task, he said, employees should “use it until it does”

    Uh, using the AI doesn’t train the AI, bud.

    • Jhex@lemmy.world
      link
      fedilink
      arrow-up
      5
      ·
      8 hours ago

      I think the lesson Jensen is pushing here is “use it until you learn to stop complaining about it”

    • llama@lemmy.zip
      link
      fedilink
      English
      arrow-up
      8
      ·
      14 hours ago

      Well that logic perfectly explains why some people keep making AI generated content even when others say they don’t want it, keep pushing slop until we do!

    • WallsToTheBalls@lemmynsfw.com
      link
      fedilink
      arrow-up
      21
      ·
      20 hours ago

      Genuinely, this is the driving misconception people have about AIs right now. That somehow everybody using them is making them smarter, when really it’s leading to model collapse

      • moonshadow@slrpnk.net
        link
        fedilink
        arrow-up
        2
        ·
        17 hours ago

        Can you help correct this for me? Don’t you feed them valuable training data and exposure to real world problems in the process of using them?

        • freddydunningkruger@lemmy.world
          link
          fedilink
          arrow-up
          9
          arrow-down
          1
          ·
          17 hours ago

          No. AI models are pre-trained, they do not learn on the fly. They are hoping to discover General Artificial Intelligence, which is what you are describing. The problem is that they don’t even understand exactly how training even works. While engineers understand the overall architecture, the specific “reasoning” or decision-making pathways within the model are too complex to fully interpret, leading to a gap between how it works and why it makes a particular decision.

          • moonshadow@slrpnk.net
            link
            fedilink
            arrow-up
            9
            ·
            16 hours ago

            My assumption wasn’t that they learned on the fly, it was that they were trained on previous interactions. Eg the team developing them would use data collected from interaction with model v3 to train model v4. Seems like juicy relevant data they wouldn’t even have to go steal and sort

            • selfAwareCoder@programming.dev
              link
              fedilink
              arrow-up
              4
              ·
              14 hours ago

              That’s true to an extend, but the interactions are only useful for training if you can mark it as good / bad etc (which is why sometimes apps will ask you if they were useful). But the ‘best’ training data like professional programming etc is usually sold at a premium tier with a promise not to use your data for training (since corporations don’t want their secrets getting out).

            • TRBoom@lemmy.zip
              link
              fedilink
              arrow-up
              2
              arrow-down
              2
              ·
              15 hours ago

              You can’t train ai on ai output. It causes degradation on the newly trained model.

              • FooBarrington@lemmy.world
                link
                fedilink
                arrow-up
                4
                arrow-down
                1
                ·
                15 hours ago

                First: that’s wrong, every big LLM uses some data cleaned/synthesized by previous LLMs. You can’t solely train on such data without degradation, but that’s not the claim.

                Second: AI providers very explicitly use user data for training, both prompts and response feedback. There’s a reason businesses pay extra to NOT have their data used for training.

                  • FooBarrington@lemmy.world
                    link
                    fedilink
                    arrow-up
                    1
                    ·
                    8 hours ago

                    I mean - yeah, it is? This is a well-researched part of the data pipelines for any big model. Some companies even got into trouble because their models identified as other models, whose outputs they were trained on.

                    It seems you have a specific bone to pick that you attribute to such training, but it’s just such a weird approach to deny pretty broadly understood results…