• Blue_Morpho@lemmy.world
    link
    fedilink
    arrow-up
    1
    arrow-down
    1
    ·
    2 days ago

    Yeah its in the news that AI companies are working to add ads. And while SEO’s are trying, its not like Google’s algorithm which can be easily gamed. Google used number of links to an url as a measure of quality. AI’s train by injesting the entire contents of the entire internet. They don’t care what is popular or what keywords are in the html title. It’s only a chain of text based on probability of the next token. It’s much harder to game a system where everything is read, not just hyperlinks and keywords.

    • very_well_lost@lemmy.world
      link
      fedilink
      English
      arrow-up
      4
      ·
      2 days ago

      I think you’re misunderstanding how AI search actually works. When you ask it to do something timely like “find me a good place to eat”, it’s not looking through its training data for the answer. There might be restaurant reviews in the training data, sure, but that stuff goes stale extremely quickly, and it’s way too expensive to train new versions of the model frequently enough to keep up with that shifting data.

      What they do instead is a technique called RAG — retrieval assisted generation. With RAG, data from some other system (a database, a search engine, etc) is pushed into the LLM’s context window (basically it’s short-term memory) so that it can use that data when crafting a response. When you ask AI for restaurant reviews of whatever, it’s just RAGing in Yelp or Google data and summarizing that. And because that’s all it’s doing, the same SEO techniques (and paid advertising deals) that push stuff to the top of a Google search will also push that same stuff to the front of the AI’s working memory. The model’s own training data guides it through the process of synthesizing a response out of that RAG data, but if the RAG data is crap, the LLMs response will still be crap.

      • ZDL@lazysoci.al
        link
        fedilink
        arrow-up
        2
        ·
        1 day ago

        Further, you can inject more text into the LLMbecile’s hidden prompt to cause some things to show up more often. Think Grok’s weird period where it was attaching the supposed plight of white people in South Africa into every query, but more subtle.