(_____(_____________(#)~~~~~~

  • 1 Post
  • 73 Comments
Joined 3 年前
cake
Cake day: 2022年4月11日

help-circle


  • But most importantly, it won’t work in the end. These scraping tech companies have much deeper pockets and can use specialized hardware that is much more efficient at solving these challenges than a normal web browser.

    A lot of people don’t seem to be able to comprehend this. Even the most basic Server Hardware that these companies have access to is many times more powerful than the best Gaming PC you can get right now. And if things get too slow they can always just spin up more nodes, which is trivial to them. If anything, they could use this as an excuse to justify higher production costs, which would make resulting datasets and models more valuable.

    If this PoW crap becomes widespread it will only make the Internet more shitty and less usable for the average person in the long term. I despise the idea of running completely arbitrary computations just so some Web Admin somewhere can be relieved to know that the CPU spikes they see coming from their shitty NodeJS/Python Framework that generates all the HTML+CSS on-the-fly, does a couple of roundtrips and adds tens of lines of log on every single request, are maybe, hopefully caused by a real human and not a sophisticated web crawler.

    My theory is people like to glaze Anubis because it’s linked to the general “Anti-AI” sentiment now (thanks to tech journalism), and also probably because its mascot character is an anime girl and the Developer/CEO of Techaro is a streamer/vtuber.




  • AI? Look, I helped a friend fix a new install. It wasn’t Linux fault, it was a setting in the bios that needed to be changed. But the AI had them trying all sorts of things that were unrelated, and was never going to help. Use with a grain of salt.

    I have the same experience but sometimes it was even worse; Sometimes the AI would confidently recommend doing things that might lead to breakage. Personally I recommend against using AI to learn Linux. It’s just not worth it and will only give new users a false impression of how things work on Linux. People are much better off reading documentation (actual documentation, not SEO slop on random websites) or asking for help in forums.







  • One time I asked DeepSeek for guidance on a more complex problem involving a linked list and I wanted to know how a simple implementation of that would look like in practice. The most high level I go is C and they claim it knows C, so I asked it to write in the C language. It literally started writing code like this:

    void important_function() {
        // important_function code goes here
    }
    
    void black_magic() {
        // Code that performs black magic goes here.
    }
    

    I tried at least 2 more times after that and while it did actually write code this time, the code it wrote made no sense whatsoever. For example one time it started writing literal C# in the middle of a C function for some reason. Another time it wrongly assumed that I’m asking for C++ (despite me explicitly stating otherwise) and the C++ it produced was horrifying and didn’t even work. Yet another time it acted like the average redditor and hyper focused on a very specific part of my prompt and then only responded to that while ignoring my actual request.

    I tried to “massage” it a lot in hopes of getting some useful information out of it but in the end I found that some random people’s Git repos and Stackexchange questions were way more helpful to my problem. All of my experiences with LLMs have been like this thus far and I’ve been messing with them for 1+ years now. People claim they’re very useful for writing repetitive or boiler plate code but I am never in a position where I’d want or need that. Maybe my use cases are just too niche lol.


  • A few days ago I was configuring some software where it’s difficult to find good documentation about so I decided to ask DeepSeek. I described what I’m trying to do and asked if it could give me an example setup so I can get a better understanding. All it did was confidently make shit up and told me things that I already knew. And that’s only the most recent example. I have yet to find LLMs be a useful tool.