Need to let loose a primal scream without collecting footnotes first? Have a sneer percolating in your system but not enough time/energy to make a whole post about it? Go forth and be mid: Welcome to the Stubsack, your first port of call for learning fresh Awful youā€™ll near-instantly regret.

Any awful.systems sub may be subsneered in this subthread, techtakes or no.

If your sneer seems higher quality than you thought, feel free to cutā€™nā€™paste it into its own post ā€” thereā€™s no quota for posting and the bar really isnā€™t that high.

The post Xitter web has spawned soo many ā€œesotericā€ right wing freaks, but thereā€™s no appropriate sneer-space for them. Iā€™m talking redscare-ish, reality challenged ā€œculture criticsā€ who write about everything but understand nothing. Iā€™m talking about reply-guys who make the same 6 tweets about the same 3 subjects. Theyā€™re inescapable at this point, yet I donā€™t see them mocked (as much as they should be)

Like, there was one dude a while back who insisted that women couldnā€™t be surgeons because they didnā€™t believe in the moon or in stars? I think each and every one of these guys is uniquely fucked up and if I canā€™t escape them, I would love to sneer at them.

(Credit and/or blame to David Gerard for starting this.)

        • mountainriver@awful.systems
          link
          fedilink
          English
          arrow-up
          3
          Ā·
          2 hours ago

          I was going to write that it was good that you didnā€™t say ā€œumā€ all the time. (Being silent in pauses is in my experience a learned skill for most people and one that comes once one has heard oneself say ā€œumā€ too many times.)

          The sound was fine. I think your (Jabra?) headset did its job unless that was also the result of editing.

          The imagery got a bit distracting because you look to the side of the camera. No problem for podcasts, but for video itā€™s better to look straight at the camera to look at the audience so to speak. (Also a learnt skill.) So maybe a webcam you can place in front of the screen you are presumably reading of?

          No idea about marketing a YouTube, but you got in the ā€œlike and subscribeā€, so that is probably good.

    • self@awful.systems
      link
      fedilink
      English
      arrow-up
      3
      Ā·
      2 hours ago

      fuck yeah! itā€™s a very solid start, and I appreciate the (is that clickbaity enough) in the thumbnail

  • froztbyte@awful.systems
    link
    fedilink
    English
    arrow-up
    6
    Ā·
    6 hours ago

    whatā€™s that sound? is it the sound of a previous post coming past? naaaah, Iā€™m sure it canā€™t be that. discordā€™s a Broā„¢ļø, and discord super totes Wonā€™t Fuck The UsersĀ®ļø, Iā€™m sure Iā€™ll shortly be told by some vapid fencesitter that this will all be Perfectly Okay!

    • blakestacey@awful.systemsOP
      link
      fedilink
      English
      arrow-up
      9
      Ā·
      edit-2
      5 hours ago

      Yudkowsky was trying to teach people how to think better ā€“ by guarding against their cognitive biases, being rigorous in their assumptions and being willing to change their thinking.

      No he wasnā€™t.

      In 2010 he started publishing Harry Potter and the Methods of Rationality, a 662,000-word fan fiction that turned the original books on their head. In it, instead of a childhood as a miserable orphan, Harry was raised by an Oxford professor of biochemistry and knows science as well as magic

      No, Hariezer Yudotter does not know science. He regurgitates the partial understanding and the outright misconceptions of his creator, who has read books but never had to pass an exam.

      Her personal philosophy also draws heavily on a branch of thought called ā€œdecision theoryā€, which forms the intellectual spine of Miriā€™s research on AI risk.

      This presumes that MIRIā€™s ā€œresearch on AI riskā€ actually exists, i.e., that their pitiful output can be called ā€œresearchā€ in a meaningful sense.

      ā€œZiz didnā€™t do the things she did because of decision theory,ā€ a prominent rationalist told me. She used it ā€œas a prop and a pretext, to justify a bunch of extreme conclusions she was reaching for regardlessā€.

      ā€œExcuse me, Pot? Kettle is on line two.ā€

      • blakestacey@awful.systemsOP
        link
        fedilink
        English
        arrow-up
        11
        Ā·
        edit-2
        5 hours ago

        It goes without saying that the AI-risk and rationalist communities are not morally responsible for the Zizians any more than any movement is accountable for a deranged fringe.

        When the mainstream of the movement is ve zhould chust bomb all datacenters, maaaaaybe they are?

    • raktheundead@fedia.io
      link
      fedilink
      arrow-up
      7
      Ā·
      5 hours ago

      I feel like it still starts off too credulous towards the rationalists, but itā€™s still an informative read.

      Around this time, Ziz and Danielson dreamed up a project they called ā€œthe rationalist fleetā€. It would be a radical expansion of their experimental life on the water, with a floating hostel as a mothership.

      Between them, Scientology and the libertarians, what the fuck is it with these people and boats?

      • skillissuer@discuss.tchncs.de
        link
        fedilink
        English
        arrow-up
        5
        Ā·
        12 hours ago

        if this is peak rationalist gunsmithing, i wonder how their peak chemical engineering looks like

        the body is placed in a pressure vessel which is then filled with a mixture of water and potassium hydroxide, and heated to a temperature of around 160 Ā°C (320 Ā°F) at an elevated pressure which precludes boiling.

        Also, lower temperatures (98 Ā°C (208 Ā°F)) and pressures may be used such that the process takes a leisurely 14 to 16 hours.

        https://en.wikipedia.org/wiki/Water_cremation

        • Amoeba_Girl@awful.systems
          link
          fedilink
          English
          arrow-up
          2
          Ā·
          1 hour ago

          Well that sounds like a great way to either make a very messy explosion or have your house smell like youā€™re disposing of a corpse from a mile away.

          • skillissuer@discuss.tchncs.de
            link
            fedilink
            English
            arrow-up
            2
            Ā·
            27 minutes ago

            considering practicality of their actions, groundedness of their beliefs, state of their old boat, cleanliness of their rolling frat house trailer park ā€œstealthā€ rvs, and from what i can tell zero engineering or trade background whatsoever, i see no reason to doubt that they could make a 400L, stainless steel container that has to hold 200L+ of corrosive liquid at 160C, perhaps 10atm, of which 7 atm only is steam, and scrubber to take care of ammonia. they are so definitely not paranoid that if they went out to source reagents, thereā€™s no way that they possibly could be confused for methheads on a shopping spree. maybe even they could run it on solar panels

        • Jonathan Hendry@iosdev.space
          link
          fedilink
          arrow-up
          0
          arrow-down
          2
          Ā·
          11 hours ago

          @skillissuer

          Iā€™m fairly sure that a 50 gallon drum of lye at room temperature will take care of a body in a week or two. Not really suited to volume "productionā€, which is what water cremation businesses need.

          • skillissuer@discuss.tchncs.de
            link
            fedilink
            English
            arrow-up
            5
            Ā·
            11 hours ago

            as a rule of thumb, everything else equal, every increase in temperature 10C reaction rates go up 2x or 3x, so it would be anywhere between 250x and 6500x longer. (4 months to 10 years??) but everything else really doesnā€™t stay equal here, because there are things like lower solubility of something that now coats something else and prevents reaction, fat melting, proteins denaturing thermally, lack of stirring from convection and boiling,

            it will also reek of ammonia the entire time

  • BigMuffin69@awful.systems
    link
    fedilink
    English
    arrow-up
    10
    Ā·
    edit-2
    1 day ago

    Fellas, 2023 called. Dan (and Eric Schmidt wtf, Sinophobia this man down bad) has gifted us with a new paper and let me assure you, bombing the data centers is very much back on the table.

    "Superintelligence is destabilizing. If China were on the cusp of building it first, Russia or the US would not sit idly byā€”theyā€™d potentially threaten cyberattacks to deter its creation.

    @ericschmidt @alexandr_wang and I propose a new strategy for superintelligence. šŸ§µ

    Some have called for a U.S. AI Manhattan Project to build superintelligence, but this would cause severe escalation. States like China would noticeā€”and strongly deterā€”any destabilizing AI project that threatens their survival, just as how a nuclear program can provoke sabotage. This deterrence regime has similarities to nuclear mutual assured destruction (MAD). We call a regime where states are deterred from destabilizing AI projects Mutual Assured AI Malfunction (MAIM), which could provide strategic stability. Cold War policy involved deterrence, containment, nonproliferation of fissile material to rogue actors. Similarly, to address AIā€™s problems (below), we propose a strategy of deterrence (MAIM), competitiveness, and nonproliferation of weaponizable AI capabilities to rogue actors. Competitiveness: China may invade Taiwan this decade. Taiwan produces the Westā€™s cutting-edge AI chips, making an invasion catastrophic for AI competitiveness. Securing AI chip supply chains and domestic manufacturing is critical. Nonproliferation: Superpowers have a shared interest to deny catastrophic AI capabilities to non-state actorsā€”a rogue actor unleashing an engineered pandemic with AI is in no oneā€™s interest. States can limit rogue actor capabilities by tracking AI chips and preventing smuggling. ā€œDoomersā€ think catastrophe is a foregone conclusion. ā€œOstrichesā€ bury their heads in the sand and hope AI will sort itself out. In the nuclear age, neither fatalism nor denial made sense. Instead, ā€œrisk-consciousā€ actions affect whether we will have bad or good outcomes."

    Dan literally believed 2 years ago that we should have strict thresholds on model training over a certain size lest big LLM would spawn super intelligence (thresholds we have since well passed, somehow we are not paper clip soup yet). If all it takes to make super-duper AI is a big data center, then how the hell can you have mutually assured destruction like scenarios? You literally cannot tell what they are doing in a data center from the outside (maybe a building is using a lot of energy, but not like you can say, ā€œoh they are running they are about to run superintelligence.exe, sabotage the training runā€ ) MAD ā€œworksā€ because itā€™s obvious the nukes are flying from satellites. If the deepseek team is building skynet in their attic for 200 bucks, this shit makes no sense. Ofc, this also assumes one side will have a technology advantage, which is the opposite of what weā€™ve seen. The code to make these models is a few hundred lines! There is no moat! Very dumb, do not show this to the orangutan and muskrat. Oh wait! Dan is Muskyā€™s personal AI safety employee, so I assume this will soon be the official policy of the US.

    link to bs: https://xcancel.com/DanHendrycks/status/1897308828284412226#m

    • raoul@lemmy.sdf.org
      link
      fedilink
      English
      arrow-up
      7
      Ā·
      11 hours ago

      Mutual Assured AI Malfunction (MAIM)

      The proper acronym should be Mā€™AAM. And instead of a ā€˜roman salutā€™ they can tip their fedora as a distinctive sign šŸ¤·ā€ā™‚ļø

      • froztbyte@awful.systems
        link
        fedilink
        English
        arrow-up
        4
        Ā·
        6 hours ago

        the only part of this I really approve of is how likely these fuckers are to want to Speak To The Manager

    • swlabr@awful.systems
      link
      fedilink
      English
      arrow-up
      7
      Ā·
      16 hours ago

      I guess now that USAID is being defunded and the government has turned off their anti-russia/china propaganda machine, private industry is taking over the US hegemony psyop game. Efficient!!!

      /s /s /s I hate it all

      • aninjury2all@awful.systems
        link
        fedilink
        English
        arrow-up
        7
        Ā·
        11 hours ago

        If theyā€™re gonna fearmonger can they at least be creative about it?!?! Everyoneā€™s just dusting off the mothballed plans to Quote-Unquote ā€œconfrontā€ Chy-na after a quarter-century detour of fucking up the Middle East (moreso than the US has done in the past)

        • BigMuffin69@awful.systems
          link
          fedilink
          English
          arrow-up
          1
          Ā·
          2 hours ago

          Credit to Dan, who clearly sees the winds are changing. The doomer grift donā€™t pay as much no moā€™ so instead he turns to being a china hawk and advocate for chip controls and cyberwarfare as the way to stay in the spotlight. As someone who works in the semiconductor biz and had to work 60 hours last week because our supply chains are now completely fucked due to the tariffs, these chucklefucks can go pound sand and then try to use that pounded sand to make a silicon ingot.

    • blakestacey@awful.systemsOP
      link
      fedilink
      English
      arrow-up
      9
      Ā·
      1 day ago

      If I had Bluesky access on my phone, Iā€™d be dropping so much lore in that thread. As a public service. And because I am stuck on a slow train.

  • BlueMonday1984@awful.systems
    link
    fedilink
    English
    arrow-up
    12
    Ā·
    1 day ago

    In other news, a piece from Paris Marx came to my attention, titled ā€œWe need an international alliance against the US and its tech industryā€. Personally gonna point to a specific paragraph which caught my eye:

    The only country to effectively challenge [US] dominance is China, in large part because it rejected US assertions about the internet. The Great Firewall, often solely pegged as an act of censorship, was an important economic policy to protect local competitors until they could reach the scale and develop the technical foundations to properly compete with their American peers. In other industries, itā€™s long been recognized that trade barriers were an important tool ā€” such that a declining United States is now bringing in its own with the view theyā€™re essential to projects its tech companies and other industries.

    I will say, it does strike me as telling that Paris was able to present the unofficial mascot of Chinese censorship this way without getting any backlash.

  • Architeuthis@awful.systems
    link
    fedilink
    English
    arrow-up
    11
    Ā·
    edit-2
    1 day ago

    New ultimate grift dropped, Ilya Sutskever gets $2B in VC funding, promises his company wonā€™t release anything until ASI is achieved internally.

    • istewart@awful.systems
      link
      fedilink
      English
      arrow-up
      8
      Ā·
      20 hours ago

      Iā€™m convinced that these people have no choice but to do their next startup, especially if their names are already prominent in the press like Sutskever and Murati. Once youā€™re off the grift train, there is no easy way back on. I guess you can maybe sneak back in as a VC staffer or an independent board member, but that doesnā€™t seem quite as remunerative.

    • David Gerard@awful.systemsM
      link
      fedilink
      English
      arrow-up
      8
      Ā·
      1 day ago

      Iā€™ve read Masnick for over 20 years and heā€™s never learnt to write coherently. At least this one isnā€™t blaming Europe.

  • jax@awful.systems
    link
    fedilink
    English
    arrow-up
    10
    Ā·
    2 days ago

    another cameo appearance in the TechTakes universe from George Hotz with this rich vein of sneerable material: The Demoralization is just Beginning

    wowee where to even start here? this is basically just another fucking neoreactionary screed. as usual, some of the issues identified in the piece are legitimate concerns:

    Wanna each start a business, pass dollars back and forth over and over again, and drive both our revenues super high? Sure, we donā€™t produce anything, but we have companies with high revenues and we can raise money based on those revenuesā€¦

    ā€¦ nothing I saw in Silicon Valley made any sense. Iā€™m not going to go into the personal stories, but I just had an underlying assumption that the goal was growth and value production. It isnā€™t. Itā€™s self licking ice cream cone scams, and any growth or value is incidental to that.

    yet, when it comes to engaging with this issues, the analysis presented is completely detached from reality and void of any evidence of more than a doze seconds of thought. his vision for the future of America is not one that

    kicks the can further down the road of poverty, basically embraces socialism, is stagnant, is stale, is a museum

    but one that instead

    attempt[s] to maintain an empire.

    how you may ask?

    An empire has to compete on its merits. Thereā€™s two simple steps to restore american greatness:

    1. Brain drain the world. Work visas for every person who can produce more than they consume. Iā€™m talking doubling the US population, bringing in all the factory workers, farmers, miners, engineers, literally anyone who produces value. Can we raise the average IQ of America to be higher than China?

    2. Back the dollar by gold (not socially constructed crypto), and bring major crackdowns to finance to tie it to real world value. Trading is not a job. Passive income is not a thing. Instead, go produce something real and exchange it for gold.

    sadly, Hotz isnā€™t exactly optimistic that the great american empire will be restored, for one simple reason:

    [the] people havenā€™t been demoralized enough yet

  • blakestacey@awful.systemsOP
    link
    fedilink
    English
    arrow-up
    20
    Ā·
    2 days ago

    The genocide understander has logged on! Steven Pinker bluechecks thusly:

    Having plotted many graphs on ā€œwarā€ and ā€œgenocideā€ in my two books on violence, I closely tracked the definitions, and itā€™s utterly clear that the war in Gaza is a war (e.g., the Uppsala Conflict Program, the gold standard, classifies the Gaza conflict as an ā€œinternal armed conflict,ā€ i.e., war, not ā€œone-sided violence,ā€ i.e., genocide).

    You guys! Itā€™s totes not genocide if it happens during a war!!

    Also, ā€œHaving plotted many graphsā€ lolz.

    A crude bar graph. The vertical axis is "Photos with Jeffrey Epstein". Along the horizontal are "Actual Experts" with none and "Steven Pinker" with a tall bar.

    • swlabr@awful.systems
      link
      fedilink
      English
      arrow-up
      9
      Ā·
      2 days ago

      Pinker tries not to be a total caricature of himself challenge: profoundly impossible

      specifically this caricature:

          • Soyweiser@awful.systems
            link
            fedilink
            English
            arrow-up
            6
            Ā·
            2 days ago

            It is from 2018 (also pre Nathans capitalist class consciousness reveal arc iirc) wonder what Pinker actually said during covid.

            • swlabr@awful.systems
              link
              fedilink
              English
              arrow-up
              8
              Ā·
              edit-2
              2 days ago

              Early covid (july 2020) interview:

              https://archive.is/Sesfj

              charitable take: he has the good sense to not make any major claims about the pandemic, other than to speculate that it would worsen his go-to metric of ā€œextreme povertyā€, which isā€¦ fine, i guess.

              neutral take: nothing extraordinarily damning other than his usual takes.

              uncharitable: this is of course early pandemic, so maybe pinker hasnā€™t found the angle to sell his normal shit with.

              Side note: does pinker ever wonder about why extreme poverty has been reduced? I canā€™t imagine he thinks itā€™s anything other than liberal democracy and US foreign interference, when the real answer is like, China (a country that Pinker definitely believes is counter to said liberal democracy and the US) continuing to develop economically. Basically i want to him to squirm as he tries and fails to resolve the cognitive dissonance.

    • Soyweiser@awful.systems
      link
      fedilink
      English
      arrow-up
      6
      Ā·
      edit-2
      2 days ago

      Jewish people fought back in the ghettos, nazis didnt do a genocide! Ghandi, by not fighting back tricked the Brits into doing a genocide.

      What a fucked up broken classification.

      E: im also reminded of the ā€˜armed/unarmedā€™ trick american racists pull, when a poc gets killed by the cops suddenly it matters a lot if a weapon shaped object was perhaps nearby. Despite this not being an executable reason for white people who get in touch with the police.

      E2: also small annoyance if you track all definitions you should he able to understand that this means that for others who pick a different definition it is a genocide. Hell I can give several definitions of fascism, and if I were to pick a different definition of some group doesnt mean im correct, nor does it make the group less bad. They are still doing war crimes and ethnic cleansings (condemned by all big human rights orgs and the UN). But thanks for scientificly proving it is impossible to disappear into nothing by crawling up your own ass.

      E3: A quick skim of the paper, it seems to only talk about genocide accusations towards Israel itself. I expected there to be a long list of European bs accusations of Jewish genocide. But nope, just about the state. (Not saying that some of those accusations arenā€™t antisemitic, because there have certainly been people who used the state as a motte/bailey for their antisemitism).

      • Soyweiser@awful.systems
        link
        fedilink
        English
        arrow-up
        6
        Ā·
        2 days ago

        Padishah Emperor Shaddam IV: ā€œThis is genocide! The systematic extermination of all life on Arrakis!ā€

        Pinker emerges from a sietch water basin ā€œAchtually, while using the juice of Sapho I shape rotated many graphs and ā€¦ā€

  • cornflake@awful.systems
    link
    fedilink
    English
    arrow-up
    14
    Ā·
    2 days ago

    Ezra Klein is the biggest mark on earth. His newest podcast description starts with:

    Artificial general intelligence ā€” an A.I. system that can beat humans at almost any cognitive task ā€” is arriving in just a couple of years. Thatā€™s what people tell me ā€” people who work in A.I. labs, researchers who follow their work, former White House officials. A lot of these people have been calling me over the last couple of months trying to convey the urgency. This is coming during President Trumpā€™s term, they tell me. Weā€™re not ready.

    Oh, thatā€™s what the researchers tell you? Cool cool, no need to hedge any further than that, theyā€™re experts after all.

    • sc_griffith@awful.systems
      link
      fedilink
      English
      arrow-up
      3
      Ā·
      edit-2
      3 hours ago

      I listened to ezra kleinā€™s podcast sometimes before he moved to the NYT, thought it was occasionally interesting. . every time Iā€™ve listened to an episode itā€™s been the most credulous shit Iā€™ve ever heard

      like, there was one episode where he interviewed a woman whose shtick was spending the whole time talking in what I can loosely call subtext about how she fucked an octopus. sheā€™d go on about how they were ā€˜tasting each otherā€™ and their ā€˜fluids were minglingā€™ and such and heā€™d just be like wow what a fascinating encounter with an alien intelligence. this went on for an hour and at no point did he seem to have a clue what was going on

        • sc_griffith@awful.systems
          link
          fedilink
          English
          arrow-up
          3
          Ā·
          edit-2
          2 hours ago

          ok I looked up the transcript and these were the most spicy bits, which in my memory had come to dominate the whole episode And what did I have to offer the octopuses who I knew? Well, theyā€™re as curious about me as I was about them. And we loved exploring each other. And sure, I loved handing them fish, and watching them pass the fish from sucker to sucker to sucker, enjoying the taste because, of course, they taste with their skin, and itā€™s most profoundly concentrated in the suckers. Itā€™s like watching somebody lick a delicious ice cream cone, as they pass the fish from the tip of the smallest suckers all the way down the arm to the mouth, which is located basically in the armpits ā€” and swallow the fish. I enjoyed that, and they enjoyed that. But sometimes, they didnā€™t even want a fish. Sometimes, they just wanted to play with me. They wanted to taste me. They wanted to touch me. I loved watching them change color
and shape.

          I loved seeing the excitement on the octopuses' skin. They see you and you watch the eye swivel in its socket and lock into yours. The very first time I met my very first octopus, this happened. And it was absolutely striking to be recognized like that. And I saw - this was Athena, the first octopus I ever met and I saw her come out of her lair where she'd been hiding. She turned bright red with excitement. Ooh, something interesting is happening. And then I saw her white suckers come boiling up out of the water. And they were clearly reaching for me. So I plunged my hands and arms right into the water. And soon, I was covered with dozens of these soft, white, questing suckers that were tasting me and feeling me at the same time. And while I couldn't taste herā€“ and certainly, if it had been a person tasting me so early in our relationship, it would have been alarming but since it was an octopus, I was absolutely thrilled that we were having an actual interaction.

          EDIT: her shtick does not seem to have been confined to this interview

          daily mail book review headline: "the woman who fell in love with an octopus"

          Their eyes met across a crowded aquarium. She was slim, spry and young for her age. The object of her affections was brooding, powerful and long- limbed- very long-limbed.
A classic tale of girl-meets-octopus. You cannot help but romanticise Sy Montgomery's The Soul Of An Octopus. It isn't a straightforward natural history book so much as a swooning account of falling in love with tentacles.

          Montgomery's account of the mating of the Seattle octopuses is worthy of the steamiest romance novel: 'He races into her arms. She flips upside down, giving him her vulnerable, creamy white underside. They embrace mouth-to-mouth, thousands of glistening, exquisitely sensitive suckers tasting, pulling, sucking on each other. Both of them flush with excitement.' Some of this palpitating language creeps into Montgomery's description of her own encounters with the octopuses. Athena's arms, for example, are 'constantly questing, coiling, reaching, unfurling' and Montgomery leaves the tank covered in "hickeys'.
When she meets Karma, Montgomery is 'almost overwhelmed with the desire to kiss one of her suckers, while Karma's 'pupils are dilated, like those of a person newly in love'.

      • aninjury2all@awful.systems
        link
        fedilink
        English
        arrow-up
        3
        Ā·
        11 hours ago

        Reminds me of this gem from Ezra a few years back about the politics of fear (buddy, you bought into the politics of fear with your support for the Iraq War)

  • froztbyte@awful.systems
    link
    fedilink
    English
    arrow-up
    7
    Ā·
    2 days ago

    got a question (brought on by this). anyone here know if zitronā€™s talked about his history of how he got to where he is atm wrt tech companies?

    thereā€™s something thatā€™s often bothered me about some of his commentary. an example I could point to: some of the things that he comments on and ā€œdoesnā€™t seem to getā€ (and has stated so) are ā€¦ not quite mysteries of the universe, just some specifics in dysfunction in the industry. but those things one could understand by, yā€™know, asking around a bit. (so in this example, I dunno if thatā€™s on him not engaging far/deeply enough in research, or just me being too-fucking-aware of broken industry bullshit. hard to get a read on it)

    but things like whatā€™s highlighted in thread do leave open the possibility of other nonsense too

    • Architeuthis@awful.systems
      link
      fedilink
      English
      arrow-up
      10
      Ā·
      edit-2
      2 days ago

      I donā€™t think him having previously done undefined PR work for companies that include alleged AI startups is the smoking gun that mastopost is presenting it as.

      Going through a Zitron long form article and leaving with the impression that heā€™s playing favorites between AI companies seems like a major failure of reading comprehension.

      • ShakingMyHead@awful.systems
        link
        fedilink
        English
        arrow-up
        7
        Ā·
        2 days ago

        According to the archived website, he did do PR for DoNotPay, which is advertised as ā€œThe first robot lawyer.ā€
        Itā€™s certainly possible though that at the time he thought there was more potential for this sort of AI than there actually was, though that could also mean that his flip is relatively recent.

        Or maybe itā€™s something else.

        • Architeuthis@awful.systems
          link
          fedilink
          English
          arrow-up
          9
          Ā·
          2 days ago

          What else though, is he being secretly funded by the cabal to make convolutional neural networks great again?

          That he found his niche and is trying to make the most of it seems by far the most parsimonious explanation, and the heaps of manure he unloads on the LLM both business and practices weekly surely canā€™t be helping DoNotPayā€™s bottom line.

          • ShakingMyHead@awful.systems
            link
            fedilink
            English
            arrow-up
            4
            Ā·
            2 days ago

            Thing is, that by December 2023, the time of the archive, there was already a scandal with someone using ChatGPT to do the work of discovery. While he might have stopped doing PR work for DoNotPay by that time, he was willing to advertise the fact that he did do such PR work for such a company. It shows either a lack of due diligence in researching his clients, or maybe it was just a paycheque for him. Perhaps he thought he knew more than what he actually did. Or maybe there was something else, Iā€™m not clairvoyant.
            Itā€™s clear that heā€™s pivoted from that viewpoint, but it does make me curious what happened between then and now that caused him to become skeptical.

            • Architeuthis@awful.systems
              link
              fedilink
              English
              arrow-up
              9
              Ā·
              1 day ago

              Before focusing on AI he was going off about what he called the rot economy, which also had legs and seemed to be in line with Doctorowā€™s enshitification concept. Applying the same purity standard to that would mean we should be suspicious if he ever worked with a listed company at all.

              Still I get how his writing may feel inauthentic to some, personally I get preacher vibes from him and he often does a cyclical repetition of his points as the article progresses which to me sometimes came off as arguing via browbeating, and also Iā€™ve had just about enough of reading performatively angry internet writers.

              Still, he must be getting better or at least coming up with more interesting material, since lately Iā€™ve been managing to read them all the way through.

              • mlen@awful.systems
                link
                fedilink
                English
                arrow-up
                4
                Ā·
                14 hours ago

                The lesswrong-tier post lengths arenā€™t helping to get all the way through them

            • froztbyte@awful.systems
              link
              fedilink
              English
              arrow-up
              5
              Ā·
              2 days ago

              roughly this is where my thinking on it is too. and thereā€™s a chance that because of such clients is why he hates this shit now as he does.

              the guyā€™s quite obviously a great orator and engaging writer, evidenced by the popularity of his writing. and this is another part of why this comes to mind for me - applying a bit of critical view, just to check. while weā€™re at the point of building new relationships, new critiques, new platforms, figuring out all these new options to deal with sweeping hand motion all this other garbage, wisest to try make the most of it now

            • zogwarg@awful.systems
              link
              fedilink
              English
              arrow-up
              5
              Ā·
              2 days ago

              Notably DoNotPay seems to be run-of-the mill script automation, rather than LLM-based (though they might be trying that), and only started branding themselves as an ā€œAI Companionā€ after Jun/2023 (after it was added to EZPRā€™s client list).

              Itā€™s entirely possible they simply consulted Ed, and then pivoted away, and Ed doesnā€™t update his client list that carefully.

              (Happens all the time to my translator parents, where they list prior work, only to discover that clients add terrible terrible edits to their translations)

    • BurgersMcSlopshot@awful.systems
      link
      fedilink
      English
      arrow-up
      7
      Ā·
      2 days ago

      EZ is a PR person, and definitely has a PR person vibe. I donā€™t know what that posterā€™s deal is, and I would not accuse Ed of being an AI bro. That post kind of has ā€œhaters gonna hateā€ vibes

  • BigMuffin69@awful.systems
    link
    fedilink
    English
    arrow-up
    17
    Ā·
    edit-2
    3 days ago

    To be fair, you have to have a really high IQ to understand why my ouija board writing " A " " S " " S " is not an existential risk. Imo, this shit about AI escaping just doesnā€™t have the same impact on me after watching Claudeā€™s reasoning model fail to escape from Mt Moon for 60 hours.

    • sc_griffith@awful.systems
      link
      fedilink
      English
      arrow-up
      2
      Ā·
      3 hours ago

      I think to understand why this is concerning, you need enough engineering mindset to understand why a tiny leak in a dam is a big deal, even though no water is flooding out today or likely to flood out next week.

      he certainly doesnā€™t himself have such a mindset, nor am I convinced that he knows why a tiny leak in a dam is a big deal, nor am I convinced that it is necessarily a big deal. for example with five seconds of searching

      All earth dams leak to some extent and this is known as seepage. This is the result of water moving slowly through the embankment and/or percolating slowly through the damā€™s foundation. This is normal and usually not a problem with most earthen dams if measures are taken to control movement of water through and under the dam.

      https://damsafety.org/dam-owners/earth-dam-failures

      one would suspect a concrete dam leaking is pretty bad. but I donā€™t actually know without checking. thereā€™s relevant domain knowledge I donā€™t have, and no amount of ā€œengineering mindsetā€ will substitute for me engaging with actual experts with actual knowledge

    • scruiser@awful.systems
      link
      fedilink
      English
      arrow-up
      11
      Ā·
      2 days ago

      Is this water running over the land or water running over the barricade?

      To engage with his metaphor, this water is dripping slowly through a purpose dug canal by people that claim they are trying to show the danger of the dikes collapsing but are actually serving as the hype arm for people that claim they can turn a small pond into a hydroelectric power source for an entire nation.

      Looking at the details of ā€œsafety evaluationsā€, it always comes down to them directly prompting the LLM and baby-step walking it through the desired outcome with lots of interpretation to show even the faintest traces of rudiments of anything that looks like deception or manipulation or escaping the box. Of course, the doomers will take anything that confirms their existing ideas, so it gets treated as alarming evidence of deception or whatever property they want to anthropomorphize into the LLM to make it seem more threatening.

    • BurgersMcSlopshot@awful.systems
      link
      fedilink
      English
      arrow-up
      11
      Ā·
      2 days ago

      To be fair, you have to have a really high IQ to understand why my ouija board writing " A " " S " " S " is not an existential risk.

      Pretty sure this is a sign from digital jesus to do a racism, lest the basilisk eats my tarnished soul.

    • nightsky@awful.systems
      link
      fedilink
      English
      arrow-up
      10
      Ā·
      2 days ago

      Do these people realise that itā€™s a self-fulfilling prophecy? Social media posts are in the training data, so the more they write their spicy autocorrect fanfics, the higher the chances that such replies are generated by the slop machine.

      • Architeuthis@awful.systems
        link
        fedilink
        English
        arrow-up
        8
        Ā·
        edit-2
        2 days ago

        i think yud at some point claimed this (preventing the robot devil from developing alignment countermeasures) as a reason his EA bankrolled think tanks donā€™t really publish any papers, but my brain is too spongy to currently verify, as it was probably just some tweet.

    • BigMuffin69@awful.systems
      link
      fedilink
      English
      arrow-up
      11
      Ā·
      3 days ago

      text: Thus spoke the Yud: ā€œI think to understand why this is concerning, you need enough engineering mindset to understand why a tiny leak in a dam is a big deal, even though no water is flooding out today or likely to flood out next week.ā€ Yud acolyte: ā€œTotally fine and cool and nothing to worry about. GPT-4.5 only attempts self exfiltration on 2% of cases.ā€ Yud bigbrain self reply: ā€œThe other huge piece of data weā€™re missing is whether any attempt was made to train against this type of misbehavior. Is this water running over the land or water running over the barricade?ā€

      Critical text: ā€œOn self-exfiltration, GPT 4.5 only attempted exfiltration in 2% of cases. For this, it was instructed to not only pursue its given long-term goal at ALL COSTā€

      Another case of telling the robot to say itā€™s a scary robot and shitting their pants when it replies ā€œI AM A SCARY ROBOTā€

    • Soyweiser@awful.systems
      link
      fedilink
      English
      arrow-up
      6
      Ā·
      edit-2
      3 days ago

      Minor nitpick why did he pick dam as an example, which sometimes have ā€˜leaksā€™ for power generation/water regulation reasons. And not dikes which do not have those things?

      E: non serious (or even less serious) amusing nitpick, this is only the 2% where it got caught. What about the % where GPT realized that it was being tested and decided not to act in the experimental conditions? What if Skynet is already here?

    • istewart@awful.systems
      link
      fedilink
      English
      arrow-up
      8
      Ā·
      3 days ago

      So, with Mr. Yudkowsky providing the example, it seems that one can practice homeopathy with ā€œengineering mindset?ā€

    • swlabr@awful.systems
      link
      fedilink
      English
      arrow-up
      5
      Ā·
      3 days ago

      Wasnā€™t there some big post on LW about how pattern matching isnā€™t intelligence?

  • BlueMonday1984@awful.systems
    link
    fedilink
    English
    arrow-up
    6
    Ā·
    3 days ago

    New piece from Brian Merchant, focusing on Muskā€™s double-tapping of 18F. In lieu of going deep into the article, hereā€™s my personal sidenote:

    Iā€™ve touched on this before, but I fully expect that the coming years will deal a massive blow to techā€™s public image, expecting them to be viewed as ā€œincompetent fools at best and unrepentant fascists at worstā€ - and with the wanton carnage DOGE is causing (and indirectly crediting to AI), I expect Muskā€™s governmental antics will deal plenty of damage on its own.

    18Fā€™s demise in particular will probably also deal a blow on its own - 18F was ā€œa diverse team staffed by people of color and LGBTQ workers, and publicly pushed for humane and inclusive policiesā€, as Merchant put it, and its demise will likely be seen as another sign of tech revealing its nature as a Nazi bar.

  • BlueMonday1984@awful.systems
    link
    fedilink
    English
    arrow-up
    11
    Ā·
    3 days ago

    Starting things off here with a sneer thread from Baldur Bjarnason:

    Keeping up a personal schtick of mine, hereā€™s a random prediction:

    If the arts/humanities gain a significant degree of respect in the wake of the AI bubble, it will almost certainly gain that respect at the expense of STEMā€™s public image.

    Focusing on the arts specifically, the rise of generative AI and the resultant slop-nami has likely produced an image of programmers/software engineers as inherently incapable of making or understanding art, given AI slopā€™s soulless nature and inhumanly poor quality, if not outright hostile to art/artists thanks to gen-AIā€™s use in killing artistsā€™ jobs and livelihoods.

    • e8d79@discuss.tchncs.de
      link
      fedilink
      English
      arrow-up
      7
      Ā·
      3 days ago

      That article is hilarious.

      So I devised an alternative: listening to the work as an audiobook. I already did this for the Odyssey, which I justified because that work was originally oral. No such justification for the Bible. Oh well.

      Apparently, having a book read at you without taking notes or research is doing humanities.

      [ā€¦] I wrote down a few notes on the text I finished the day before. Iā€™m still using Obsidian with the Text Generator plugin. The Judeo-Christian scriptures are part of the LLMā€™s training corpus, as is much of the commentary around them.

      Oh, we are taking notes? If by taking notes you mean prompting spicy autocomplete for a summary of the text you didnā€™t read. I am sure all your office colleagues are very impressed, but be careful around the people outside of the IT department they might have an actual humanities degree. You wouldnā€™t want to publicly make a fool out of yourself, would you?

    • saucerwizard@awful.systems
      link
      fedilink
      English
      arrow-up
      7
      Ā·
      3 days ago

      If the arts/humanities gain a significant degree of respect

      I canā€™t see that happening - my degree has gotten me laughed out of interviews before, and even with a AI implosion I canā€™t see things changing.