• 5 Posts
  • 276 Comments
Joined 10 months ago
cake
Cake day: April 27th, 2024

help-circle




  • No. I am not saying that to put man and machine in two boxes. I am saying that because it is a huge difference, and yes, a practical one.

    An LLM can talk about a topic for however long you wish, but it does not know what it is talking about, it has no understanding or concept of the topic. And that shines through the instance you hit a spot where training data was lacking and it starts hallucinating. LLMs have “read” an unimaginable amount of texts on computer science, and yet as soon as I ask something that is niche, it spouts bullshit. Not it’s fault, it’s not lying; it’s just doing what it always does, putting statistically likely token after statistically liken token, only in this case, the training data was insufficient.

    But it does not understand or know that either; it just keeps talking. I go “that is absolutely not right, remember that <…> is <…,>” and whether or not what I said was true, it will go "Yes, you are right! I see now, <continues to hallucinate> ".

    There’s no ghost in the machine. Just fancy text prediction.



  • A begruntled Culture citizen entasked with a seemingly straight-forward matter in a coincidentally-very-earth-like civ where capitalism reigns supreme in all the worst ways. A society on the brink of collapse due to leapfrogging technological advancements while ignoring, and possibly suppressing, the societal changes this necessitates, benefitting only those already in power. Not-entirely-clear motivations of unseen Culture Minds, presumably plotting beyond what is apparent to our protagonist. Getting to re-evaluate this strange society by experiencing first its thrills, then slowly but surely its horrors through the eyes of that protagonist, whose view of the world and themselves can never be the same afterwards.

    All the while, a member of this fledgling society, finding themselves aboard the Culture vessel “I’m sorry, I thought there was still money on that card; here, try this one”, a state-of-the-art warshipnope we don’t do those, a state-of-the-art Very Fast Picket, earth-bound with engines pushing their limits, is motivated by the fleeting hope of maybe, just maybe, getting there in time to share a crucial piece of information with the protagonist, but - oh, too late. This storyline lead nowhere, and you are still glad to have read it, for the possibly best parts of the book where the witty ship-banter that had you laughing with tears.

    Oh we also end the book with the cold-blooded murder of a BezosMusk look-alike at the hands of SC, and thanks to the book you are left with the clear impression that in the grand schemes of things, this was not simply necessary, it was just.

    …something like that?


  • Yeah, with seniors it’s even more clear how little LMs can help.

    I feel you on the AI tools being pushed thing. My company is too small to have a dedicated team for something like that, buuuut… As of last week, we’re wasting resources on an internal server hosting Deepseek on absurd hardware. Like, far more capable than our prod server.

    Oh, an we pride ourselves on being soooo environmentally friendly 😊🎉






  • Even with LMs supposedly specialising in the areas that I am knowledgable (but by no means an expert) in, it’s the same. Drill down even slightly beyond surface-level, and it’s either plain wrong, or halucinated when not immediately disprovable.

    And why wouldn’t it be? These things do not possess knowledge, they possess the ability to generate texts about things we’d like them to be knowledgable in, and that is a crucial difference.


  • I’m a programmer as well. When ChatGPT & Co initially came out, I was pretty excited tbh and attempted to integrate it into my workflow, which kinda worked-ish? But was also a lot of me being amazed by the novelty, and forgiving of the shortcomings.

    Did not take me long to phase them out again though. (And no, it’s not the models I used; I have tried again now and then with the new, supposedly perfect-for-programming models, same results). The only edgecase where they are generally useful (to me at least) are simple tasks that I have some general knowledge of (to double theck the LM’s work) but not have any interest in learning anything further than I already know. Which does occur here and there, but rarely.

    For everything else programming-related, it’s flat out shit.I do not beleive they are a time saver for even moderately difficult programs. Bu the time you’ve run around in enough circles, explaining “now, this does not do what you say it does”, “that’s the same wring answer you gave me two responses ago”, “you have hallucinated that function”, and found out the framework in use dropped that general structure in version 5, you may as well do it yourself, and actually learn how to do it at the same time.

    For work, I eventually found that it took me longer to describe the business logic (and do the above dance) than to just… do the work. I also have more confidence in the code, and understand it completely.

    In terms of programming aids, a linter, formatter and LSP are, IMHO, a million times more useful than any LM.