Mine attempts to lie whenever it can if it doesn’t know something. I will call it out and say that is a lie and it will say “you are absolutely correct” tf.

I was reading into sleeper agents placed inside local LLMs and this is increasing the chance I’ll delete it forever. Which is a shame because it is the new search engine seeing how they ruined search engines

  • Bob Robertson IX @discuss.tchncs.de
    link
    fedilink
    English
    arrow-up
    16
    ·
    2 days ago

    Think about the data that the models were trained on… pretty much all of it was based on sites like Reddit and Stack Overflow.

    If you look at the conversations that occur on those sites, it is very rare for someone to ask a question and then someone else replies with “I don’t know”, or even an “I don’t know, but I think this is how you could find out”. Instead, the vast majority of replies are someone confidently stating what they believe is the truth.

    These models are just mimicking the data they’ve been trained on, and they have not really been trained to be unsure. It’s up to us as the users to not rely on an LLM as a source of truth.