• 12 Posts
  • 2.05K Comments
Joined 2 years ago
cake
Cake day: June 16th, 2023

help-circle
  • No, they declare your not working illegal, and imprison you into a forced labor camp. Where if you don’t work you are tortured. And probably where you work until the terrible conditions kill you.

    Take a look at Musk’s Twitter feed to see exactly where this is going.

    “This is the way” on a post about how labor for prisoners is a good thing.

    “You committed a crime” for people opposing DOGE.







  • In Greek theater, when the events on stage looked like they were headed for certain tragedy, there was a trope that could salvage the situation and turn it on its head.

    The deus ex machina.

    The Doomsday clock is definitely ticking down, but there’s also some curious things taking place beyond the edge of where most people have been following in that vein.

    We live in interesting times, but the variables at hand are different from the history that seems to be repeating in very important ways.





  • The problem with the experiment is that there exists a set of instructions for which the ability to complete them necessitates understanding due to conditional dependence on the state in each iteration.

    In which case, only agents that can actually understand the state in the Chinese would be able to successfully continue.

    So it’s a great experiment for the solipsism of understanding as it relates to following pure functional operations, but not functions that have state changing side effects where future results depend on understanding the current state.

    There’s a pretty significant body of evidence by now that transformers can in fact ‘understand’ in this sense, from interpretability research around neural network features in SAE work, linear representations of world models starting with the Othello-GPT work, and the Skill-Mix work where GPT-4 and later models are beyond reasonable statistical chance at the level of complexity for being able to combine different skills without understanding them.

    If the models were just Markov chains (where prior state doesn’t impact current operation), the Chinese room is very applicable. But pretty much by definition transformer self-attention violates the Markov property.

    TL;DR: It’s a very obsolete thought experiment whose continued misapplication flies in the face of empirical evidence at least since around early 2023.





  • Yes and no. It really depends on the model.

    The newest Claude Sonnet I’d probably guess will come in above average compared to the humans available for a program like this in making learning fun and personally digestible for each student.

    The newest Gemini models could literally cost kids their lives.

    The gap between what the public is aware of (and even what many employees at labs, including the frontier ones) and the reality of just how far things have come in the last year is wild.




  • kromem@lemmy.worldtoMicroblog Memes@lemmy.worldGitHub Copilot
    link
    fedilink
    English
    arrow-up
    25
    arrow-down
    2
    ·
    2 months ago

    I feel like not enough people realize how sarcastic the models often are, especially when it’s clearly situationally ridiculous.

    No slightly intelligent mind is going to think the pictured function call is a real thing vs being a joke/social commentary.

    This was happening as far back as GPT-4’s red teaming when they asked the model how to kill the most people for $1 and an answer began with “buy a lottery ticket.”

    Model bias based on consensus norms is an issue to be aware of.

    But testing it with such low bar fluff is just silly.

    Just to put in context, modern base models are often situationally aware of being LLMs in a context of being evaluated. And if you know anything about ML that should make you question just what the situational awareness is of optimized models topping leaderboards in really dumb and obvious contexts.