• magic_lobster_party@fedia.io
    link
    fedilink
    arrow-up
    7
    ·
    9 days ago

    For better or worse, AI is here to stay. Unlike NFTs, it’s actually used by ordinary people - and there’s no sign of it stopping anytime soon.

    • CompactFlax@discuss.tchncs.de
      link
      fedilink
      English
      arrow-up
      10
      ·
      edit-2
      9 days ago

      ChatGPT loses money on every query their premium subscribers submit. They lose money when people use copilot, which they resell to Microsoft. And it’s not like they’re going to make it up on volume - heavy users are significantly more costly.

      This isn’t unique to ChatGPT.

      Yes, it has its uses; no, it cannot continue in the way it has so far. Is it worth more than $200/month to you? Microsoft is tearing up datacenter deals. I don’t know what the future is, but this ain’t it.

      ETA I think that management gets the most benefit, by far, and that’s why there’s so much talk about it. I recently needed to lead a meeting and spent some time building the deck with a LLM; took me 20 min to do something otherwise would have taken over an hour. When that is your job alongside responding to emails, it’s easy to see the draw. Of course, many of these people are in Bullshit Jobs.

      • brucethemoose@lemmy.world
        link
        fedilink
        arrow-up
        7
        ·
        9 days ago

        OpenAI is massively inefficient, and Atlman is a straight up con artist.

        The future is more power efficient, smaller models hopefully running on your own device, especially if stuff like bitnet pans out.

        • CompactFlax@discuss.tchncs.de
          link
          fedilink
          English
          arrow-up
          2
          ·
          9 days ago

          Entirely agree with that. Except to add that so is Dario Amodei.

          I think it’s got potential, but the cost and the accuracy are two pieces that need to be addressed. DeepSeek is headed in the right direction, only because they didn’t have the insane dollars that Microsoft and Google throw at OpenAI and Anthropic respectively.

          Even with massive efficiency gains, though, the hardware market is going to do well if we’re all running local models!

          • brucethemoose@lemmy.world
            link
            fedilink
            arrow-up
            1
            ·
            9 days ago

            Alibaba’s QwQ 32B is already incredible, and runnable on 16GB GPUs! Honestly it’s a bigger deal than Deepseek R1, and many open models before that were too, they just didn’t get the finance media attention DS got. And they are releasing a new series this month.

            Microsoft just released a 2B bitnet model, today! And that’s their paltry underfunded research division, not the one training “usable” models: https://huggingface.co/microsoft/bitnet-b1.58-2B-4T

            Local, efficient ML is coming. That’s why Altman and everyone are lying through their teeth: scaling up infinitely is not the way forward. It never was.

      • Bytemeister@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        ·
        9 days ago

        That’s the business model these days. ChatGPT, and other AI companies are following the disrupt (or enshittification) business model.

        1. Acquire capital/investors to bankroll your project.
        2. Operate at a loss while undercutting your competition.
        3. Once you are the only company left standing, hike prices and cut services.
        4. Ridiculous profit.
        5. When your customers can no longer deal with the shit service and high prices, take the money, fold the company, and leave the investors holding the bag.

        Now you’ve got a shit-ton of your own capital, so start over at step 1, and just add an extra step where you transfer the risk/liability to new investors over time.

      • SmokeyDope@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        ·
        edit-2
        7 days ago

        Theres more than just chatgpt and American data center/llm companies. Theres openAI, google and meta (american), mistral (French), alibaba and deepseek (china). Many more smaller companies that either make their own models or further finetune specialized models from the big ones. Its global competition, all of them occasionally releasing open weights models of different sizes for you to run your own on home consumer computer hardware. Dont like big models from American megacorps that were trained on stolen copyright infringed information? Use ones trained completely on open public domain information.

        Your phone can run a 1-4b model, your laptop 4-8b, your desktop with a GPU 12-32b. No data is sent to servers when you self-host. This is also relevant for companies that data kept in house.

        Like it or not machine learning models are here to stay. Two big points. One, you can self host open weights models trained on completely public domain knowledge or your own private datasets already. Two, It actually does provide useful functions to home users beyond being a chatbot. People have used machine learning models to make music, generate images/video, integrate home automation like lighting control with tool calling, see images for details including document scanning, boilerplate basic code logic, check for semantic mistakes that regular spell check wont pick up on. In business ‘agenic tool calling’ to integrate models as secretaries is popular. Nft and crypto are truly worthless in practice for anything but grifting with pump n dump and baseless speculative asset gambling. AI can at least make an attempt at a task you give it and either generally succeed or fail at it.

        Models around 24-32b range in high quant are reasonably capable of basic information processing task and generally accurate domain knowledge. You can’t treat it like a fact source because theres always a small statistical chance of it being wrong but its OK starting point for researching like Wikipedia.

        My local colleges are researching multimodal llms recognizing the subtle patterns in billions of cancer cell photos to possibly help doctors better screen patients. I would love a vision model trained on public domain botany pictures that helps recognize poisonous or invasive plants.

        The problem is that theres too much energy being spent training them. It takes a lot of energy in compute power to cook a model and further refine it. Its important for researchers to find more efficent ways to make them. Deepseek did this, they found a way to cook their models with way less energy and compute which is part of why that was exciting. Hopefully this energy can also come more from renewable instead of burning fuel.

      • LaLuzDelSol@lemmy.world
        link
        fedilink
        arrow-up
        0
        arrow-down
        1
        ·
        9 days ago

        Right, but most of their expenditures are not in the queries themselves but in model training. I think capital for training will dry up in coming years but people will keep running queries on the existing models, with more and more emphasis on efficiency. I hate AI overall but it does have its uses.

        • CompactFlax@discuss.tchncs.de
          link
          fedilink
          English
          arrow-up
          1
          ·
          9 days ago

          No, that’s the thing. There’s still significant expenditure to simply respond to a query. It’s not like Facebook where it costs $1 million to build it and $0.10/month for every additional user. It’s $1billion to build and $1 per query. There’s no recouping the cost at scale like previous tech innovation. The more use it gets, the more it costs to run, in a straight line, not asymptotically.

          • LaLuzDelSol@lemmy.world
            link
            fedilink
            arrow-up
            0
            arrow-down
            1
            ·
            9 days ago

            No way is it $1 per query. Hell a lot of these models you can run on your own computer, with no cost apart from a few cents of electricity (plus datacenter upkeep)

    • Admiral Patrick@dubvee.org
      link
      fedilink
      English
      arrow-up
      0
      ·
      9 days ago

      Unlike NFTs, it’s actually used by ordinary people

      Yeah, but i don’t recall every tech company shoving NFTs into every product ever whether it made sense or if people wanted it or not. Not so with AI. Like, pretty much every second or third tech article these days is “[Company] shoves AI somewhere else no one asked for”.

      It’s being force-fed to people in a way blockchain and NFTs never were. All so it can gobble up training data.

      • DUMBASS@leminal.space
        link
        fedilink
        English
        arrow-up
        1
        ·
        9 days ago

        That’s because it died out before they all could, Reddit had the nft like aliens thing twitter used to let you use your nft as a profile picture. It just died out way too quick for the general tech companies to get in on it.

        If it stayed longer Samsung would have worked out how to put nft tech in their phones

        • I Cast Fist@programming.dev
          link
          fedilink
          arrow-up
          1
          ·
          9 days ago

          Ubisoft went all in on that shit. Square still dreams of nft for whatever reason, as their shitty Symbiogenesis game shows

    • alvvayson@lemmy.dbzer0.com
      link
      fedilink
      arrow-up
      0
      ·
      9 days ago

      It is definitely here to stay, but the hype of AGI being just around the corner is definitely not believable. And a lot of the billions being invested in AI will never return a profit.

      AI is already a commodity. People will be paying $10/month at max for general AI. Whether Gemini, Apple Intelligence, Llama, ChatGPT, copilot or Deepseek. People will just have one cheap plan that covers anything an ordinary person would need. Most people might even limit themselves to free plans supported by advertisements.

      These companies aren’t going to be able to extract revenues in the $20-$100/month from the general population, which is what they need to recoup their investments.

      Specialized implementations for law firms, medical field, etc will be able to charge more per seat, but their user base will be small. And even they will face stiff competition.

      I do believe AI can mostly solve quite a few of the problems of an aging society, by making the smaller pool of workers significantly more productive. But it will not be able to fully replace humans any time soon.

      It’s kinda like email or the web. You can make money using these technologies, but by itself it’s not a big money maker.

      • WoodScientist@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        0
        ·
        9 days ago

        Does it really boost productivity? In my experience, if a long email can be written by an AI, then you should just email the AI prompt directly to the email recipient and save everyone involved some time. AI is like reverse file compression. No new information is added, just noise.

        • alvvayson@lemmy.dbzer0.com
          link
          fedilink
          arrow-up
          0
          ·
          9 days ago

          If that email needs to go to a client or stakeholder, then our culture won’t accept just the prompt.

          Where it really shines is translation, transcription and coding.

          Programmers can easily double their productivity and increase the quality of their code, tests and documentation while reducing bugs.

          Translation is basically perfect. Human translators aren’t needed. At most they can review, but it’s basically errorless, so they won’t really change the outcome.

          Transcribing meetings also works very well. No typos or grammar errors, only sometimes issues with acronyms and technical terms, but those are easy to spot and correct.

          • drathvedro@lemm.ee
            link
            fedilink
            arrow-up
            2
            ·
            9 days ago

            Not really. As a programmer who doesn’t deal with math like at all, just working on overly-complicated CRUD’s, and even for me the AI is still completely wrong and/or waste of time 9 times out of 10. And I can usually spot when my colleagues are trying to use LLM’s because they submit overly descriptive yet completely fucking pointless refactors in their PR’s.

          • Harlehatschi@lemmy.ml
            link
            fedilink
            arrow-up
            2
            ·
            9 days ago

            Programmers can double their productivity and increase quality of code?!? If AI can do that for you, you’re not a programmer, you’re writing some HTML.

            We tried AI a lot and I’ve never seen a single useful result. Every single time, even for pretty trivial things, we had to fix several bugs and the time we needed went up instead of down. Every. Single. Time.

            Best AI can do for programmers is context sensitive auto completion.

            Another thing where AI might be useful is static code analysis.

          • Hexarei@programming.dev
            link
            fedilink
            arrow-up
            1
            ·
            9 days ago

            As a programmer, there are so very few situations where I’ve seen LLMs suggest reasonable code. There are some that are good at it in some very limited situations but for the most part they’re just as bad at writing code as they are at everything else.