• very_well_lost@lemmy.world
    link
    fedilink
    English
    arrow-up
    9
    ·
    2 days ago

    Using AI as a search engine has become almost a necessity because Google and Bing have destroyed the usefulness of search engines with ads.

    What? Using AI for search is even worse than using a conventional search engine. All the LLM is doing is summarizing data that it did a Google search to get, and its summarization obscures the obvious ads and astroturfing that’s easy to spot when you’re doing the search yourself.

    AI is complete garbage for search unless you know that all of the data you’re searching through is accurate and trustworthy. Data from the public Internet is very much not that.

    • Blue_Morpho@lemmy.world
      link
      fedilink
      arrow-up
      1
      arrow-down
      4
      ·
      2 days ago

      Llm’s don’t read the SEO keywords and then give you a result filtered through Google’s adsense. Llm’s read absolutely everything and the results are (as of now) not filtered by who paid the most to show you a particular result.

      • very_well_lost@lemmy.world
        link
        fedilink
        English
        arrow-up
        5
        ·
        2 days ago

        Llm’s don’t read the SEO keywords and then give you a result filtered through Google’s adsense.

        Maybe not, but if you don’t think people are already doing “AI optimization” to get AI search tools to prefer their shitty content, then I have a trillion dollar data center I’d like to sell you.

        • Blue_Morpho@lemmy.world
          link
          fedilink
          arrow-up
          1
          arrow-down
          1
          ·
          2 days ago

          Yeah its in the news that AI companies are working to add ads. And while SEO’s are trying, its not like Google’s algorithm which can be easily gamed. Google used number of links to an url as a measure of quality. AI’s train by injesting the entire contents of the entire internet. They don’t care what is popular or what keywords are in the html title. It’s only a chain of text based on probability of the next token. It’s much harder to game a system where everything is read, not just hyperlinks and keywords.

          • very_well_lost@lemmy.world
            link
            fedilink
            English
            arrow-up
            4
            ·
            2 days ago

            I think you’re misunderstanding how AI search actually works. When you ask it to do something timely like “find me a good place to eat”, it’s not looking through its training data for the answer. There might be restaurant reviews in the training data, sure, but that stuff goes stale extremely quickly, and it’s way too expensive to train new versions of the model frequently enough to keep up with that shifting data.

            What they do instead is a technique called RAG — retrieval assisted generation. With RAG, data from some other system (a database, a search engine, etc) is pushed into the LLM’s context window (basically it’s short-term memory) so that it can use that data when crafting a response. When you ask AI for restaurant reviews of whatever, it’s just RAGing in Yelp or Google data and summarizing that. And because that’s all it’s doing, the same SEO techniques (and paid advertising deals) that push stuff to the top of a Google search will also push that same stuff to the front of the AI’s working memory. The model’s own training data guides it through the process of synthesizing a response out of that RAG data, but if the RAG data is crap, the LLMs response will still be crap.

            • ZDL@lazysoci.al
              link
              fedilink
              arrow-up
              2
              ·
              1 day ago

              Further, you can inject more text into the LLMbecile’s hidden prompt to cause some things to show up more often. Think Grok’s weird period where it was attaching the supposed plight of white people in South Africa into every query, but more subtle.