Need to let loose a primal scream without collecting footnotes first? Have a sneer percolating in your system but not enough time/energy to make a whole post about it? Go forth and be mid: Welcome to the Stubsack, your first port of call for learning fresh Awful you’ll near-instantly regret.

Any awful.systems sub may be subsneered in this subthread, techtakes or no.

If your sneer seems higher quality than you thought, feel free to cut’n’paste it into its own post — there’s no quota for posting and the bar really isn’t that high.

The post Xitter web has spawned soo many “esoteric” right wing freaks, but there’s no appropriate sneer-space for them. I’m talking redscare-ish, reality challenged “culture critics” who write about everything but understand nothing. I’m talking about reply-guys who make the same 6 tweets about the same 3 subjects. They’re inescapable at this point, yet I don’t see them mocked (as much as they should be)

Like, there was one dude a while back who insisted that women couldn’t be surgeons because they didn’t believe in the moon or in stars? I think each and every one of these guys is uniquely fucked up and if I can’t escape them, I would love to sneer at them.

      • Soyweiser@awful.systems
        link
        fedilink
        English
        arrow-up
        8
        ·
        4 months ago

        It told him he was very smart and correct so he had an hypnogasm yes.

        (For the people not in the know and who want some more psychic damage, look up Scott and his ‘I can hypnotise you into having orgasms’ blog posts, the man is utterly nuts, and it is really scary how many people seriously follow him).

        • o7___o7@awful.systems
          link
          fedilink
          English
          arrow-up
          8
          ·
          4 months ago

          Looks like he confused “tantrum” with “tantric”

          No shame, happens to guys all the time

      • self@awful.systems
        link
        fedilink
        English
        arrow-up
        19
        ·
        4 months ago

        explaining these things to normal humans is how you turn them into humans with years of psychic damage, and I support that mission wholeheartedly

        an attempt at a summary:

        the dilbert guy, who believes he can hypnotize women into having sex with him, now also believes he knows a magic incantation to teach hypnosis to a chatbot, and heavily implies the chatbot was able to hypnotize him in turn (presumably into having sex with it)

        • V0ldek@awful.systems
          link
          fedilink
          English
          arrow-up
          10
          ·
          4 months ago

          who believes he can hypnotize women into having sex with him

          Wait I missed this, he believes what?

          Now this is doubly funny.

          • Soyweiser@awful.systems
            link
            fedilink
            English
            arrow-up
            10
            ·
            edit-2
            4 months ago

            Content warning for this post, you will get psychic damage.

            It is worse, he believes he can hypnotize people over the internet into having orgasms. Here is a link to a screenshot, not sure if I’m able to find a archive link. I think he deleted his blog.

            E: a tumblr has archive links but it seems he managed to delete the first post.

            • V0ldek@awful.systems
              link
              fedilink
              English
              arrow-up
              13
              ·
              edit-2
              4 months ago

              Hey, haven’t seen a proof of god’s non-existence today yet, thanks!

              extreme psychic damage trap

              Submissives, I want you to start planning your New Year’s Eve now. Make sure you have some time alone, or some time with a partner who fully accepts your wonderful nature. But most of all, I want you wet, or hard, and especially obedient, starting now. And I want you to know how much I enjoy putting you in this state of mind. It starts now, but will get more intense by Thursday night. Expect to be a quivering, throbbing, wet mess by then.

              How did that make you feel?

              Queasy.

              This is quite literally “creep discovers dirty-talking, thinks it’s a superpower even though he’s terrible at it”.

              EDIT: From Part 3:

              So today I am only talking to the perhaps 20% of you who were turned off by this series.

              Bwahahahaha! That’s some funny shit mate, I’ve seen many bollocks number straight from between them buttocks, but this is fuckin’ choice.

              If you see a train entering a tunnel, think of it as nothing but transportation.

              Hey mate, I think… I think that’s what most normal people do? I sometimes imagine an action hero standing on top and fighting a villain before the tunnel slams them in the heads but that’s just me. What do you think about, Scott? Don’t answer that. please

    • Amoeba_Girl@awful.systems
      link
      fedilink
      English
      arrow-up
      13
      ·
      4 months ago

      If this wasn’t Scott Adams I’d have assumed this was fetish content. Now I don’t know what to think.

    • scruiser@awful.systems
      link
      fedilink
      English
      arrow-up
      10
      ·
      edit-2
      4 months ago

      I saw people making fun of this on (the normally absurdly overly credulous) /r/singularity of all places. I guess even hopeful techno-rapture believers have limits to their suspension of disbelief.

    • corbin@awful.systems
      link
      fedilink
      English
      arrow-up
      3
      ·
      4 months ago

      At risk of being NSFW, this is an amazing self-own, pardon the pun. Hypnosis via text only works on folks who are fairly suggestible and also very enthusiastic about being hypnotized, because the brain doesn’t “power down” as much machinery as with the more traditional lie-back-on-the-couch setup. The eyes have to stay open, the text-processing center is constantly engaged, and re-reading doesn’t deepen properly because the subject has to have the initiative to scroll or turn the page.

      Adams had to have wanted to be hypnotized by a chatbot. And that’s okay! I won’t kinkshame. But this level of engagement has to be voluntary and desired by the subject, which is counter to Adams’ whole approach of hypnosis as mind control.

      • froztbyte@awful.systems
        link
        fedilink
        English
        arrow-up
        1
        ·
        4 months ago

        Yep, was thinking something similar. Dude just posted a public sscce about the effect addressed in the llmentalist article

  • froztbyte@awful.systems
    link
    fedilink
    English
    arrow-up
    20
    ·
    4 months ago

    imagine one day buying some shitpost novelty stickers from that one site you heard a friend mention sometime, and then getting them and laughing about it and forgetting it

    all too rapidly the years pass: young trees shoot up, older trees start boughing their way past electrical lines, the oldest all already in their position of maximum comfort. whole generations of memes have been born and died. you no longer even get to make fun of your weird aunt for still sending the dancing baby gif (these days it’s all about the autotune clips of a decade ago…)

    and then one day you get reminded that the shitpost novelty sticker web store exists by receiving an email from them

    what the fuck.

    • Sailor Sega Saturn@awful.systems
      link
      fedilink
      English
      arrow-up
      13
      ·
      4 months ago

      In his twitter thread he’s attempting to troll people in the replies. And not even doing a particularly good job at it. A bold business strategy.

      • Soyweiser@awful.systems
        link
        fedilink
        English
        arrow-up
        12
        ·
        edit-2
        4 months ago

        Apparently they also once chained a design into ‘liberal moron’ from what was some political message once in the past, so it isn’t coming from nowhere. Guess it wasn’t an innocent mistake.

  • o7___o7@awful.systems
    link
    fedilink
    English
    arrow-up
    15
    ·
    edit-2
    4 months ago

    In the spirit of the fondly-remembered “Cloud-to-Butt” plugin, I propose a new tool that transforms mentions of AI and LLMs to something else. But what shape should our word transformer take? While scatology is good shit, I bet we can come up with something more clever than “ChatGPT-to-Poop”

    BTW, take 1d6 psychic damage when you realize that Cloud-to-Butt was released more than ten years ago

    Edit: Favorite suggestions so far:

    • Architeuthis@awful.systems
      link
      fedilink
      English
      arrow-up
      17
      ·
      edit-2
      4 months ago

      Ah yes, Alexander’s unnumbered hordes, that endless torrent of humanity that is all but certain to have made a lasting impact on the sparsely populated subcontinent’s collective DNA.

      edit: Also, the absolute brain on someone who would think that before entertaining a random recent western ancestor like a grandfather or whateverthefuckjesus.

      • YourNetworkIsHaunted@awful.systems
        link
        fedilink
        English
        arrow-up
        11
        ·
        4 months ago

        No, it wasn’t a massive impact by numbers but that superior euro genetics has been floating around waiting to be born like the Kwasitz Haderach.

  • Sailor Sega Saturn@awful.systems
    link
    fedilink
    English
    arrow-up
    14
    ·
    edit-2
    4 months ago

    AI Maxers Thrilled with Trump’s Vice President Pick JD Vance

    This isn’t really too interesting yet; but something to keep an eye on. As things like blockchain and AI alignment becomes weirdly political it’s likely that sneering will get unpleasantly close to politics at times. And yet sneer we must.

    Other self-titled techno-optimists highlighted Vance’s ties to venture capital, Thiel, and Andreessen, saying the “Gray Tribe [is] in control.” Gray Tribe is a reference to a term originating from Scott Alexander’s Slate Star Codex blog, which points to a group that is neither red (Republican) or Blue (Democrat), but a libertarian, tech savvy alternative.

    • self@awful.systems
      link
      fedilink
      English
      arrow-up
      15
      ·
      4 months ago

      I really should write more about technofascism while I still have electricity, clean water, a relatively unfractured global information network, and there aren’t too many gunshots outside

        • self@awful.systems
          link
          fedilink
          English
          arrow-up
          12
          ·
          4 months ago

          I have an extremely rough draft up on our writing community MoreWrite: here it is. it very much deserves to be restructured and expanded; I wrote this first draft in a hurry, and since then I’ve both gotten some very solid recommendations for prior work in the area to read and cite from @[email protected] and I’ve experienced more instances where technofascist methods were used to take control over open source projects away from their communities.

            • self@awful.systems
              link
              fedilink
              English
              arrow-up
              8
              ·
              4 months ago

              nah all good! our writing community isn’t meant to be containment, it’s just a different category with different expectations (mainly that feedback should be constructive, since MoreWrite is for writers posting their own work)

    • Soyweiser@awful.systems
      link
      fedilink
      English
      arrow-up
      12
      ·
      4 months ago

      But but, tracingwood told us that claiming NRx is close to Rationalism is a lie made up by the evil David G.

    • Mii@awful.systems
      link
      fedilink
      English
      arrow-up
      10
      ·
      4 months ago

      God damn, I don’t think I’ve read an article with that many name drops in a while. It’s like a Marvel film but with techfash assholes.

      “JD Vance is pro-OSS [open source] AI,” Verdon tweeted. “We are so unfathomably back.”

      It’s actually impressive how this guy is able to make me despise him even more every single time he opens his mouth.

    • V0ldek@awful.systems
      link
      fedilink
      English
      arrow-up
      15
      ·
      4 months ago

      Can an observer be a single photon, or does it have to be a conscious human being?

      The former. I’m glad we can stop the article right there and go home.

      What the fuck is this question even? What the fuck is “conscious”? Do you think in the double-slit experiment we closed a guy inside the box to watch?

      • V0ldek@awful.systems
        link
        fedilink
        English
        arrow-up
        12
        ·
        4 months ago

        Under this, let’s charitably call it, “interpretation”, the Schrödinger cat analogy makes no sense, surely THE CAT is bloody conscious about ITSELF BEING ALIVE??

    • self@awful.systems
      link
      fedilink
      English
      arrow-up
      14
      ·
      4 months ago

      there’s so much quantum woo in that article I want to sneer at, but I don’t know anywhere close to enough about quantum physics to do so without showing my entire ass

      • mountainriver@awful.systems
        link
        fedilink
        English
        arrow-up
        11
        ·
        4 months ago

        To me, the most sneerable thing in that article is where they assume a mechanical brain will evolve from ChatGPT and then assume a sufficiently large quantum computer to run it on. And then start figuring out how to port the future mechanical brain to the quantum computer. All to be able to run an old thought experiment that at least I understood as highlighting the absurdity of focusing on the human brain part in the collapse of a wave function.

        Once we build two trains that can run near the speed of light we will be able to test some of Einstein’s thought experiments. Better get cracking on how we can get enough coal onboard to run the trains long enough to get the experiments done.

      • Soyweiser@awful.systems
        link
        fedilink
        English
        arrow-up
        10
        ·
        4 months ago

        Well a good thing to remember re quantum mechanics, Schrödinger Cat is intended as a thought experiment showing how dumb the view on QM was. So it is always a bit funny to see people extrapolate from that thought experiment without acknowledging the history and issues with it. (But I think that also depends on the various interpretations, and this means I’m showing a cheekily high amount of ass here myself).

        • swlabr@awful.systems
          link
          fedilink
          English
          arrow-up
          11
          ·
          4 months ago

          Pretty much any mention of a thought experiment in the wild gets my hackles up. “Isn’t it cool that the cat is alive and dead at the same time?” Shut up! Shut up shut up shut up!!! Tho to be honest it might just be schrodinger’s cat that comes up. I wish they’d leave the poor cat alone, and stop trying to poison it.

          • blakestacey@awful.systems
            link
            fedilink
            English
            arrow-up
            11
            ·
            4 months ago

            I have a whole series of rants about that cat, starting with how it doesn’t illuminate anything about quantum theory specifically — as opposed to probabilistic or stochastic theories in general — and culminating in “Hey, maybe we should stop naming things after pedo creeps.”

          • corbin@awful.systems
            link
            fedilink
            English
            arrow-up
            1
            ·
            4 months ago

            What really gets me is that we never look past Schrödinger’s version of the cat. I want us to talk about Bell’s Cat, which cannot be alive or dead due to a contradiction in elementary linear algebra and yet reliably is alive or dead once the box opens. (I guess technically it should be “alive”, “dead”, or “undead”, since we’re talking spin-1 particles.)

      • blakestacey@awful.systems
        link
        fedilink
        English
        arrow-up
        9
        ·
        4 months ago

        There are some interesting ideas in that general direction (wrapping Bell inequalities within different new types of thought experiment, etc.), but some of the people involved have done rather a lot of overselling, and now bringing in talk of “AI” just obscures the whole situation. Which was already obscure enough.

    • blakestacey@awful.systems
      link
      fedilink
      English
      arrow-up
      10
      ·
      4 months ago

      According to one story at least, Wigner eventually concluded that if you take some ideas that physicists widely hold about quantum mechanics as postulates and follow them through to their logical conclusion, then you must conclude that there is a special role for conscious observers. But he took that as a reason to question those assumptions.

      (That story comes from Leslie Ballentine reporting a conversation with Wigner in the course of promoting an ensemble interpretation of QM.)

      • smiletolerantly@awful.systems
        link
        fedilink
        English
        arrow-up
        6
        ·
        4 months ago

        Also there’s a book by Stephen Baxter set in his Xeelee universe which takes this premise for the cult mentality of a terrorist cell

        Sorry this doesn’t really add anything, just thought it kinda funny

      • bunchberry@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        ·
        edit-2
        4 months ago

        Yes, the problem with quantum mechanics is it’s not just your Deepak Chopras of the world that get sucked into quantum woo, but even a lot of respectable academics with serious credentials, thus giving credence to these ideas. Quantum mechanics is a context-dependent theory, the properties of systems are context variant. It is not observer-dependent. The observer just occupies their own unique context and since it is context-dependent, they have to describe things from their own context.

        It is kind of like velocity in Galilean relativity, you have to take into account reference frame. Two observers in Galilean relativity could disagree on certain things, such as the velocity of an object but the disagreement is not “confusing” because if you understand relativity, you’d know it’s just a difference in reference frame. Nothing important about “observers” here.

        I do not understand what is with so many academics in fully understanding that properties of systems can be variant under different reference frames in special relativity, but when it comes to quantum mechanics their heads explode trying to interpret the contextual nature of it and resort to silly claims like saying it proves some fundamental role for the conscious observer. All it shows is that the properties of systems are context variant. There is nothing else.

        Once you accept that, then everything else follows. All of the unintuitive aspects of quantum mechanics disappear, you do not need to posit systems in two places at once, some special role for observers, a multiverse, nonlocality, hidden variables, nothing. All the “paradoxes” disappear if you just accept the context variance of the states of systems.

    • aio@awful.systems
      link
      fedilink
      English
      arrow-up
      10
      ·
      4 months ago

      I honestly think anyone who writes “quantum” in an article should be required to take a linear algebra exam to avoid being instantly sacked

  • self@awful.systems
    link
    fedilink
    English
    arrow-up
    13
    ·
    4 months ago

    speaking of technofascism, we’re at the stage where supposed Democrat billionaires like the Andreesen Horowitz fuckers suddenly come out in support of Trump:

    Marc Andreessen, the co-founder of one of the most prominent venture capital firms in Silicon Valley, says he’s been a Democrat most of his life. He says he has endorsed and voted for Bill Clinton, Al Gore, John Kerry, Barack Obama and Hillary Clinton.

    However, he says he’s no longer loyal to the Democratic Party. In the 2024 presidential race, he is supporting and voting for former President Donald Trump. The reason he is choosing Trump over President Joe Biden boils down primarily to one major issue — he believes Trump’s policies are much more favorable for tech, specifically for the startup ecosystem.

    none of this should be surprising, but it should be called out every time it happens, and we’re gonna see it happen a lot in the days ahead. these fuckers finally feel secure in taking their masks off, and that’s not good.

    • blakestacey@awful.systems
      link
      fedilink
      English
      arrow-up
      12
      ·
      4 months ago

      Apparently the “startup ecosystem” matters more than the ecosystem of, you know, actual living things.

      These people are just amazingly fucking evil.

    • David Gerard@awful.systemsOPM
      link
      fedilink
      English
      arrow-up
      8
      ·
      4 months ago

      Democrat?? AH’s previous hit was the one where they enthusiastically endorsed literally the co-author of the original Fascist manifesto

      and a16z does get Yarvin in to dispense wisdom and insight

      • YourNetworkIsHaunted@awful.systems
        link
        fedilink
        English
        arrow-up
        4
        ·
        4 months ago

        See, I feel like the Democrats have had a pretty strong technocrat wing that is much more in synch with Neoreaction than people care to acknowledge. As the right shifts towards pursuing the pro-racist anti-women anti-lgbt aspects of their agenda through the courts rather than the ballot box, it seems like the fault lines between the technocratic fascists and the theocratic fascists are thinner than the lines between the techfash and the progressives.

    • pyrex@awful.systems
      link
      fedilink
      English
      arrow-up
      7
      ·
      edit-2
      4 months ago

      I don’t understand why people take him at face value when he claims he’s always been a Democrat up until now. He’s historically made large contributions to candidates from both parties, but generally more Republicans than Democrats, and also Republican PACs like Protect American Jobs. Here is his personal record.

      Since 2023, he picked up and donated ~$20,000,000 to Fairshake, a crypto PAC which predominantly funds candidates running against Democrats.

      Has he moved right? Sure. Was he ever left? No, this is the voting record of someone who wants to buy power from candidates belonging to both parties. If it implies anything, it implies he currently finds Republicans to be corruptible.

  • Sailor Sega Saturn@awful.systems
    link
    fedilink
    English
    arrow-up
    12
    ·
    edit-2
    4 months ago

    HN: I am starting an AI+Education company called Eureka Labs.

    Their goal: robo-feynman:

    For example, in the case of physics one could imagine working through very high quality course materials together with Feynman, who is there to guide you every step of the way. Unfortunately, subject matter experts who are deeply passionate, great at teaching, infinitely patient and fluent in all of the world’s languages are also very scarce and cannot personally tutor all 8 billion of us on demand. However, with recent progress in generative AI, this learning experience feels tractable.

    NGL though mostly just sharing this link for the concept art concept fart which features a three-armed many fingered woman smiling at an invisible camera.

    • YourNetworkIsHaunted@awful.systems
      link
      fedilink
      English
      arrow-up
      10
      ·
      4 months ago

      But just focus on the vibes. This diverse group of young mutants getting an education in the overgrown ruins of this university.

      Not sure how it ties into robo-feyman at all but the vibes

    • blakestacey@awful.systems
      link
      fedilink
      English
      arrow-up
      9
      ·
      4 months ago

      For example, in the case of physics one could imagine working through very high quality course materials together with Feynman

      Women of the world: um, about that

    • froztbyte@awful.systems
      link
      fedilink
      English
      arrow-up
      9
      ·
      edit-2
      4 months ago

      Others: unslanted solar panels at ground level in shade under other solar panels, 90-degree water steps (plural), magical mystery staircases and escalator tubes, picture glass that reflects anything it wants to instead of what may actually be in the reflected light path, a whole Background Full Of Ill-Defined Background People because I guess the training set imagery was input at lower pixel density(??), and on stage left we have a group in conversation walking and talking also right on the edge of nowhere in front of them

      And that’s all I picked up in about 30-40s of looking

      Imagine being the kind of person who thinks this shit is good

    • swlabr@awful.systems
      link
      fedilink
      English
      arrow-up
      8
      ·
      4 months ago

      So that’s what our kids will look like once society rebuilds after global thermonuclear war!!!

  • Steve@awful.systems
    link
    fedilink
    English
    arrow-up
    10
    ·
    4 months ago

    We should neoligise “NASB” for our community as shorthand for “Not a sneer, but”

    • self@awful.systems
      link
      fedilink
      English
      arrow-up
      14
      ·
      4 months ago

      she is, regrettably, a fascist who reputedly still hides right-wing conspiracy shit in her projects

      I unfortunately know that startled feeling well; I used to be a fan of the work she did with sectorlambda and similar projects

      • pyrex@awful.systems
        link
        fedilink
        English
        arrow-up
        3
        ·
        4 months ago

        A friend who worked with her is sympathetic to her but does not endorse her: this is a tendency she has, she veers back and forth on it a lot, she has frequent moments of insight where she disavows her previous actions but then just kind of continues doing them. It’s Kanye-type behavior.

  • David Gerard@awful.systemsOPM
    link
    fedilink
    English
    arrow-up
    10
    ·
    4 months ago

    comment from friend:

    Slightly related: now I know when the AI crash is going to happen. Every bottomfeeder recruiter company on LinkedIn is suddenly pushing 2-month contract technical writer positions with AI companies with no product, no strategy, and no idea of how to proceed other than “CEO cashes out.” I suspect the idea is to get all of their documentation together so they can sell their bags of magic beans before the beginning of the holiday season.

    sickos.jpg

    I have asked if he can send me links to a few of these, I’ll see what I can do with 'em

    • blakestacey@awful.systems
      link
      fedilink
      English
      arrow-up
      7
      ·
      4 months ago

      That seems suspiciously soon, but my impression is based on nothing but vibes — a sense that companies are still buying in.

      • YourNetworkIsHaunted@awful.systems
        link
        fedilink
        English
        arrow-up
        7
        ·
        4 months ago

        I think there was a report saying that the most recent quarter still showed a massive infusion of VC cash into the space, but I’m not sure how much of that comes from the fact that a new money sink hasn’t yet started trending in the valley. It wouldn’t surprise me if the griftier founders were looking to cash out before the bubble properly bursts in order to avoid burning bridges with the investors they’ll need to get the next thing rolling.

        • David Gerard@awful.systemsOPM
          link
          fedilink
          English
          arrow-up
          1
          ·
          edit-2
          4 months ago

          We were asking around AI industry peons in March and they all guessed around three quarters too. I woulda put it at maybe two years myself, but I was surprised at so many people all arriving at around three quarters. OTOH, I would say that just in the past few months things are really obviously heading for a trauma.

          • jonhendry@awful.systems
            link
            fedilink
            English
            arrow-up
            2
            ·
            4 months ago

            Perhaps that’s part of why so many SV types are backing Trump. Grifting off Trump may be their fallback after the AI bubble collapses.

            • David Gerard@awful.systemsOPM
              link
              fedilink
              English
              arrow-up
              1
              ·
              4 months ago

              crypto’s VC investment fell off a cliff after the crash, and that investment is what we were talking about there

              hence their pivot to AI

              • pyrex@awful.systems
                link
                fedilink
                English
                arrow-up
                1
                ·
                4 months ago

                Oh, OK. I think all the VC-adjacent people still really believe in crypto, if it helps. They probably also don’t believe in it, depending on the room. I think it will come back.

                • David Gerard@awful.systemsOPM
                  link
                  fedilink
                  English
                  arrow-up
                  1
                  ·
                  4 months ago

                  they’ve stopped putting fresh money in, but they believe fervently in the massive bags they’re holding

          • froztbyte@awful.systems
            link
            fedilink
            English
            arrow-up
            1
            ·
            4 months ago

            I’m severely backlogged on catching up to things but my (total and complete) guess would be something like: all the recent headlines about funding and commitments are almost certainly imprecise in localisation and duration - everyone that “got money” didn’t necessarily get “money” but instead commitments to funding, and “everyone” is a much smaller set of entities that don’t encompass a really wide gallery of entities

            So for all the previously-extant promptfondlers/ model dilettantes/etc out there, the writing may indeed have been (and may still be) on the wall ito runway (“startup operating capital remaining available and viable to avoid death”)

            Based on the kind of headlines seen (and presuming the above supposition for the sake of argument), and the kind of utterly milquetoast garbage all the interceding months have produced, I don’t think it’s likely that much of the promised money will make it through to this layer/lot either. But that’s entirely a guess at this stage (and I can think of some fairly hefty counter-argument examples that may contribute to countering, not least because of how many people/orgs wouldn’t want to be losing face to fucking this up)

    • Architeuthis@awful.systems
      link
      fedilink
      English
      arrow-up
      5
      ·
      edit-2
      4 months ago

      Current flavor AI is certainly getting demystified a lot among enterprise people. Let’s dip our toes into using an LLM to make our hoard of internal documents more accessible, it’s supposed to actually be good at that, right? is slowly giving way to “What do you mean RAG is basically LLM flavored elasticsearch only more annoying and less documented? And why is all the tooling so bad?”

      • Mii@awful.systems
        link
        fedilink
        English
        arrow-up
        4
        ·
        edit-2
        4 months ago

        “What do you mean RAG is basically LLM flavored elasticsearch only more annoying and less documented? And why is all the tooling so bad?”

        Our BI team is trying to implement some RAG via Microsoft Fabrics and Azure AI search because we need that for whatever reason, and they’ve burned through almost 10k for the first half of the running month already, either because it’s just super expensive or because it’s so terribly documented that they can’t get it to work and have to try again and again. Normal costs are somewhere around 2k for the whole month for traffic + servers + database and I haven’t got the foggiest what’s even going on there.

        But someone from the C suite apparently wrote them a blank check because it’s AI …

      • o7___o7@awful.systems
        link
        fedilink
        English
        arrow-up
        2
        ·
        edit-2
        4 months ago

        Confucius, the Buddha, and Lao Tzu gather around a newly-opened barrel of vinegar.

        Confucius tastes the vinegar and perceives bitterness.

        The Buddha tastes the vinegar and perceives sourness.

        Lao Tzu tastes the vinegar and perceives sweetness, and he says, “Fellas, I don’t know what this is but it sure as fuck isn’t vinegar. How much did you pay for it?”

        • self@awful.systems
          link
          fedilink
          English
          arrow-up
          1
          ·
          4 months ago

          so, uh, you remember AskJeeves?

          (alternative answer: the third buzzword in a row that’s supposed to make LLMs good, after multimodal and multiagent systems absolutely failed to do anything of note)

        • pyrex@awful.systems
          link
          fedilink
          English
          arrow-up
          1
          ·
          4 months ago

          It’s the technique of running a primary search against some other system, then feeding an LLM the top ~25 or so documents and asking it for the specific answer.

        • corbin@awful.systems
          link
          fedilink
          English
          arrow-up
          1
          ·
          4 months ago
          NSFW (including funny example, don't worry)

          RAG is “Retrieval-Augmented Generation”. It’s a prompt-engineering technique where we run the prompt through a database query before giving it to the model as context. The results of the query are also included in the context.

          In a certain simple and obvious sense, RAG has been part of search for a very long time, and the current innovation is merely using it alongside a hard prompt to a model.

          My favorite example of RAG is Generative Agents. The idea is that the RAG query is sent to a database containing personalities, appointments, tasks, hopes, desires, etc. Concretely, here’s a synthetic trace of a RAG chat with Batman, who I like using as a test character because he is relatively two-dimensional. We ask a question, our RAG harness adds three relevant lines from a personality database, and the model generates a response.

          > Batman, what's your favorite time of day?
          Batman thinks to themself: I am vengeance. I am the night.
          Batman thinks to themself: I strike from the shadows.
          Batman thinks to themself: I don't play favorites. I don't have preferences.
          Batman says: I like the night. The twilight. The shadows getting longer.
          
      • imadabouzu@awful.systems
        link
        fedilink
        English
        arrow-up
        1
        ·
        4 months ago

        Maybe hot take, but I actually feel like the world doesn’t need strictly speaking more documentation tooling at all, LLM / RAG or otherwise.

        Companies probably actually need to curate down their documents so that simpler thinks work, then it doesn’t cost ever increasing infrastructure to overcome the problems that previous investment actually literally caused.

        • Architeuthis@awful.systems
          link
          fedilink
          English
          arrow-up
          2
          ·
          4 months ago

          Companies probably actually need to curate down their documents so that simpler thinks work, then it doesn’t cost ever increasing infrastructure to overcome the problems that previous investment actually literally caused

          Definitely, but the current narrative is that you don’t need to do any of that, as long as you add three spoonfulls of AI into the mix you’ll be as good as.

          Then you find out what you actually signed up for is to do all the manual preparation of building an on-premise search engine to query unstructured data, and you still might end up with a tool that’s only slightly better than trying to grep a bunch of pdfs at the same time.

      • rook@awful.systems
        link
        fedilink
        English
        arrow-up
        1
        ·
        4 months ago

        What do you mean RAG is basically LLM flavored elasticsearch

        I always saw it more as LMGTFYaaS.

  • Soyweiser@awful.systems
    link
    fedilink
    English
    arrow-up
    9
    ·
    4 months ago

    NSFW, as NSAB, I know that anti-environmentalists shout a lot about ‘what about china china should go green first!’ while not knowing china is in fact doing a lot to try and go green (at least on the co2 energy front, I’m not asking here to go point out all the bad things china does to fuck up the environment). I see ‘we should develop AI before china does so’ be a big pro AI argument, so here is my question. Is china even working on massive A(G)I like the people claim?

    • Architeuthis@awful.systems
      link
      fedilink
      English
      arrow-up
      5
      ·
      edit-2
      4 months ago

      I am overall very uninformed about the chinese thechnological day-to-day, but here’s two interesting facts:

      They set some pretty draconian rules early on about where the buck stops if your LLM starts spewing false information or (god forbid) goes against party orthodoxy so I’m assuming if independent research is happening It doesn’t appear much in the form of public endpoints that anyone might use.

      A few weeks ago I saw a report about chinese medical researchers trying use AI agents(?) to set up a virtual hospital in order to maybe eventually have some sort of a virtual patient entity that a medical student could work with somehow, and look how many thousands of virtual patients our handful of virtual doctors are healing daily, isn’t it awesome folks. Other than the rampant startupiness of it all, what struck me was that they said they had chatgpt-3.5 set up up the doctor/patient/nurse agents, i.e. they used the free version.

      So, who knows? If they are all-in in AGI behind the scenes they don’t seem to be making a big fuss about it.

      • Soyweiser@awful.systems
        link
        fedilink
        English
        arrow-up
        2
        ·
        4 months ago

        Thanks, you and froztbytes reply answered my questions on if they were doing it and also a bit on how seriously they are all taking it.

    • froztbyte@awful.systems
      link
      fedilink
      English
      arrow-up
      4
      ·
      4 months ago

      Typing from phone, please excuse lack of citations. Academic output in various parts of ML research have increasingly come from China and Chinese researchers over the past decade. There’s multiple inputs to this - funding, how strong a specific school/research centre is, etc, but it’s been ramping up. Pretty sure part of this is one of the fuel sources in keeping the pro-hegemonist US argument popular and going lately (also part of where the “we should before they do” comes from I guess)

      I’ve seen some mentions of recent legislation direction about LLM usage but I’m not fully up to scratch on what it is, haven’t had the time to read up

      • Soyweiser@awful.systems
        link
        fedilink
        English
        arrow-up
        2
        ·
        4 months ago

        Thanks I was not aware, so they are doing things regarding the research at least. So the “concern” isn’t totally made up. Which is what I wanted to know. As Architeuthis mentioned the legislation is against false info and against going against the party (which seem to be what you could expect).

  • BlueMonday1984@awful.systems
    link
    fedilink
    English
    arrow-up
    8
    ·
    4 months ago

    Not a sneer, but an observation on the tech industry from Baldur Bjarnason, plus some of my own thoughts:

    I don’t think I’ve ever experienced before this big of a sentiment gap between tech – web tech especially – and the public sentiment I hear from the people I know and the media I experience.

    Most of the time I hear “AI” mentioned on Icelandic mainstream media or from people I know outside of tech, it’s being used as to describe something as a specific kind of bad. “It’s very AI-like” (“mjög gervigreindarlegt” in Icelandic) has become the talk radio short hand for uninventive, clichéd, and formulaic.

    Baldur has pointed that part out before, and noted how its kneecapping the consumer side of the entire bubble, but I suspect the phrase “AI” will retain that meaning well past the bubble’s bursting. “AI slop”, or just “slop”, will likely also stick around, for those who wish to differentiate gen-AI garbage from more genuine uses of machine learning.

    To many, “AI” seems to have become a tech asshole signifier: the “tech asshole” is a person who works in tech, only cares about bullshit tech trends, and doesn’t care about the larger consequences of their work or their industry. Or, even worse, aspires to become a person who gets rich from working in a harmful industry.

    For example, my sister helps manage a book store as a day job. They hire a lot of teenagers as summer employees and at least those teens use “he’s a big fan of AI” as a red flag. (Obviously a book store is a biased sample. The ones that seek out a book store summer job are generally going to be good kids.)

    I don’t think I’ve experienced a sentiment disconnect this massive in tech before, even during the dot-com bubble.

    Part of me suspects that the AI bubble’s spread that “tech asshole” stench to the rest of the industry, with some help from the widely-mocked NFT craze and Elon Musk becoming a punching bag par excellence for his public breaking-down of Twitter.

    (Fuck, now I’m tempted to try and cook up something for MoreWrite discussing how I expect the bubble to play out…)

    • YourNetworkIsHaunted@awful.systems
      link
      fedilink
      English
      arrow-up
      5
      ·
      4 months ago

      The active hostility from outside the tech world is going to make this one interesting, since unlike crypto this one seems to have a lot of legitimate energy behind it in the industry even as it becomes increasingly apparent that even if the technical capability was there (e.g. the bullshit problems could be solved by throwing enough compute and data at the existing paradigm, which looks increasingly unlikely) there’s no way to do it profitably given the massive costs of training and using these models.

      I wonder if we’re going to see any attempts to optimize existing models for the orgs that have already integrated them in the same way that caching a web page or indexing a database can increase performance without doing a whole rebuild. Nvidia won’t be happy to see the market for GPUs fall off, but OpenAI might have enough users of their existing models that they can keep operating even while dramatically cutting down on new training runs? Does that even make sense, or am I showing my ignorance here?