Skip Navigation

Stubsack: weekly thread for sneers not worth an entire post, week ending 26th January 2025 - awful.systems - awful.systems

Need to let loose a primal scream without collecting footnotes first? Have a sneer percolating in your system but not enough time/energy to make a whole post about it? Go forth and be mid: Welcome to the Stubsack, your first port of call for learning fresh Awful you’ll near-instantly regret.

Any awful.systems sub may be subsneered in this subthread, techtakes or no.

If your sneer seems higher quality than you thought, feel free to cut’n’paste it into its own post — there’s no quota for posting and the bar really isn’t that high.

The post Xitter web has spawned soo many “esoteric” right wing freaks, but there’s no appropriate sneer-space for them. I’m talking redscare-ish, reality challenged “culture critics” who write about everything but understand nothing. I’m talking about reply-guys who make the same 6 tweets about the same 3 subjects. They’re inescapable at this point, yet I don’t see them mocked (as much as they should be)

Like, there was one dude a while back who insisted that women couldn’t be surgeons because they didn’t believe in the moon or in stars? I think each and every one of these guys is uniquely fucked up and if I can’t escape them, I would love to sneer at them.

(Semi-obligatory thanks to @dgerard for starting this.)

200 comments
  • Days since last open source issue tracker pollution by annoying nerds: zero

    My investigation tracked to you Outlier.ai as the source of problems - where your instructional videos are tricking people into creating those issues to - apparently train your AI.

    I couldn't locate these particular instructional videos, but from what I can gather outlierai farms out various "tasks" to internet gig workers as part of some sort of AI training scheme.

    Bonus terribleness: one of the tasks a few months back was apparently to wear a head mounted camera "device" to record ones every waking moment.

  • CIDR 2025 is ongoing (Conference on Innovative Data Systems Research). It's a very good conference in computer science, specifically database research (an equivalent of a journal for non-CS science). And they have a whole session on LLMs called "LLMs ARE THE NEW NO-SQL"

    I didn't have time to read the papers yet, believe me I will, but the abstracts are spicy

    We systematically develop benchmarks to study [the problem] and find that standard methods answer no more than 20% of queries correctly, confirming the need for further research in this area.

    (Text2SQL is Not Enough: Unifying AI and Databases with TAG, Biswal et al.)

    Hey guys and gals, I have a slightly different conclusion, maybe a baseline 20% correctness is a great reason to not invest a second more of research time into this nonsense? Jesus DB Christ.

    I'd also like to shoutout CIDR for setting up a separate "DATABASES AND ML" session, which is an actual research direction with interesting results (e.g. query optimizers powered by an ML model achieving better results than conventional query optimizers). At least actual professionals are not conflating ML with LLMs.

  • This is a thought I've been entertaining for some time, but this week's discussion about Ars Technica's article on Anthropic, as well as the NIH funding freeze, finally prodded me to put it out there.

    A core strategic vulnerability that Musk, his hangers-on, and geek culture more broadly haven't cottoned onto yet: Space is 20th-century propaganda. Certainly, there is still worthwhile and inspirational science to be done with space probes and landers; and the terrestrial satellite network won't dwindle in importance. I went to high school with a guy who went on to do his PhD and get into research through working with the first round of micro-satellites. Resources will still be committed to space. But as a core narrative of technical progress to bind a nation together? It's gassed. The idea that "it might be ME up there one day!" persisted through the space shuttle era, but it seems more and more remote. Going back to the moon would be a remake of an old television show, that went off the air because people ended up getting bored with it the first time. Boots on Mars (at least healthy boots with a solid chance to return home) are decades away, even if we start throwing Apollo money at it immediately. The more outlandish ideas like orbital data centers and asteroid mining don't have the same inspirational power, because they are meant to be private enterprises operated by thoroughly unlikeable men who have shackled themselves to a broadly destructive political program.

    For better or worse, biotechnology and nanotechnology are the most important technical programs of the 21st century, and by backgrounding this and allowing Trump to threaten funding, the tech oligarchs kowtowing to him right now are undermining themselves. Biotech should be obvious, although regulatory capture and the impulse for rent-seeking will continue to hold it back in the US. I expect even more money to be thrown at nanotechnology manufacturing going into the 2030s, to try to overcome the fact that semiconductor scaling is hitting a wall, although most of what I've seen so far is still pursuing the Drexlerian vision of MEMS emulating larger mechanical systems... which, if it's not explicitly biocompatible, is likely going down a cul-de-sac.

    Everybody's looking for a positive vision of the future to sell, to compete with and overcome the fraudulent tech-fascists who lead the industry right now. A program of accessible technology at the juncture of those two fields would not develop overnight, but could be a pathway there. Am I off base here?

    • This seems like yet another disconnect between however the fuck science communication has been failing the general public and myself.

      Like when you say space I think, fuck yeah, space! Those crisp pictures of Pluto! Pictures of black holes! The amazing JWST data! Gravitational waves detection! Recreating the conditions of the early universe in particle accelerators to unlock the secrets of spacetime! Just most amazing geek shit that makes me as excited as I was when I was 12 looking at the night sky through my cheap-ass telescope.

      Who gives a single fuck about sending people up there when we have probes and rovers, true marvels of engineering, feeding us data back here? Did you know Voyager 1, Voyager Fucking ONE, almost 50 years old probe, over 150 AU away from Earth, is STILL SENDING US DATA? We engineered the fuck of that bolt bucket so that even the people that designed it are surprised by how long it lasted. You think a human would last 50 years in the interstellar medium? I don't fucking think so.

      We're unlocking the secrets of the universe and confirming theories from decades ago, has there been a more exciting time to be a scientist? Wouldn't you want to run a particle accelerator? Do science on the ISS? Be the engineer behind the next legendary probe that will benefit mankind even after you're gone? If you can't spin this into a narrative of technical progrees and humans being amazing then that's a skill issue, you lack fucking whimsy.

      And I don't think there's a person in the world less whimsical than Elon fucking Musk.

    • Agree with space travel being retro-futurist fluff. It's very rich men badly remembering mediocre science fiction.

      The US could lead the world in innovation in green technology but that's now tainted by wokeness.

    • Hmm, any sort of vision for generating public support for development of a technology has to have either ideological backing or a profit incentive. I don’t say this to mean that the future must be profitable, rather, I say this to mean that you don’t get the space race if western powers aren’t afraid of communism appearing as a viable alternative to capitalism, on both ideological and commercial fronts.

      Unfortunately, a vision of that kind is necessarily technofascist. Rather than look for a tech-forward vision of the future, we need deprogram ourselves and unlearn the unspoken narratives that prop up capitalism and liberal democracy as the only viable forms of society. We need to dismantle the systems and structures that require the complex political buy-in for projects that are clearly good for society at large.

      Uh, I guess I’ve kind of gone completely orthogonal to your point of discussion. I’m kind of saying the collapse of the US is inevitable.

      • what? space race was thinly disguised ICBM development program

      • On another somewhat orthogonal point, I suspect AI has likely soured the public on any kinda tech-forward vision for the foreseeable future.

        Both directly and indirectly, the AI slop-nami has caused a lot of bad shit for the general public - from plagiarism to misinformation, from shit-tier AI art to screwing human artists, the public has come to view AI as an active blight on society, and use of AI as a virtual "Kick Me" sign.

      • For the US to avoid collapse, the Democrats would have to sweep the board in multiple successive elections and be more unified and committed to deep reform than they ever have been.

        I will pause for the laughter to fade.

      • No actually, I think what you have to say is in line with my broader point. As the top source of global consumer demand, America is primarily held together by its supply chains at this point. To be crude about it, the best reasons to be an American in the 21st century are the swag and the cheap gas. When the MAGA and Fox News crowd are pointing fingers and ranting about Marxism, they're actively trying to obscure materialism and keep people from thinking about material conditions. Having a material program, that at least has elements that can be built from the bottom up, is at least as crucial as having an electoral program. I know the Four Thieves people got rightfully shredded here a few weeks back, and that kind of technical pushback on amateur dreams is necessary, so it's a tough needle to thread. But for instance, consider Gavin Newsom's plan to have California operate its own insulin production, within existing systems and regulations: https://calmatters.org/health/2025/01/insulin-production-gavin-newsom/ This is a Newsom policy I actually think is a fantastic idea, and a big credit to him if it happens! But it's bogged down in the production-line validation stage, because we already know how to synthesize insulin and that it's effective. And the production may not even be in California when it happens! There's plenty of room for improvement here.

        Space and centralized, rent-seeking "AI" are not material programs that improve conditions for the broader population. The original space program was successful because a more tightly controlled media environment gave the opportunity to use it to cover for the missile development that was the enduring practical outcome. Positive consumer outcomes from all that have always felt, to me, like something that was bolted onto the history later. We wouldn't have Tang and transistors if not for Apollo! Well, one is kind of shitty and useless, the other is so overwhelmingly advantageous that it surely would have happened anyway.

        And to your last point, I somewhat sadly feel like a lot of doomer shit I was reading ~15 years ago actually prepared me to at least be unsurprised about the situation we're in. A lot of those writers (James Howard Kunstler, John Michael Greer for instance) have either softly capitulated, or else happily slotted themselves into the middle of the red-brown alliance. I think that's a big part of why we're at where we're at: a lot of people who were actually willing to consider the idea of American collapse were perfectly fine with letting it happen.

  • Buckle up humans; because humanity's last exam just dropped: https://lastexam.ai/ (Hacker News discussion). May the odds be ever in your favor.

    Edit: Per NyTimes, whom I hate, they were apparently trying to avoid an over-dramatic name. Amazing:

    The test’s original name, “Humanity’s Last Stand,” was discarded for being overly dramatic.

    • i only want to notice that the example chemistry question has two steps out of three that are very similar to last image in wikipedia article https://en.wikipedia.org/wiki/Electrocyclic_reaction (question explicitly mentions that it is electrocyclic reaction and mentions the same class of natural product)

      e: the exact reaction scheme that is answer to that question is in article linked just above that image. taking last image from wiki article and one of schemes from cited article gives the exact same compound as in question, and provides answer. considering how these spicy autocomplete rainforest incinerators work, this sounds like some serious ratfucking, right? you don't even have to know how this all works to get that and it's an isolated and a bit obscure subsubfield

      • You think people would secretly submit easy questions just for the reward money, and that since the question database is so big and inscrutable no one bothered to verify one way or another? No, that could never happen.

    • oh cool, the logo’s just a barely modified sparkle emoji so you know it’s horseshit, and it’s directly funded by Scale AI and a Rationalist thinktank so the chances the models weren’t directly trained on the problem set are vanishingly thin. this is just the FrontierMath grift with new, more dramatic, paint.

      e: also, slightly different targeting — FrontierMath was looking to grift institutional dollars, I feel. this one’s designed to look good in a breathless thinkpiece about how, I dunno…

      When A.I. Passes This Test, Look Out

      yeah, whatever the fuck they think this means. this one’s designed to be talked about, to be brought up behind closed doors as a reason why your pay’s being cut. this is vile shit.

    • We publicly release these questions, while maintaining a private test set of held out questions to assess model overfitting.

      ... Oh so it's a training dataset, got you.

    • Humanity's last exam or AI grifters first bankrupcy.

    • just mark C for every answer if you don't get it, that's what the State of California taught me in elementary school

  • trump just dumped half trillion dollars into openai-softbank-oracle thing https://eu.usatoday.com/story/news/politics/elections/2025/01/21/trump-stargate-ai-openai-oracle-softbank/77861568007/

    you'd think it's a perfect bait for saudi sovereign wealth fund, and perhaps it is

    for comparison, assuming current levels of spending, this will be something around 1/10 of defense spending in the same timeframe. which goes to, among other things, payrolls of millions of people and maintenance, procurement and development of rather pricey weapons like stealth planes (B-21 is $700M each) and nuclear-armed nuclear-powered submarines ($3.5B per Ohio-class, with $31M missiles, up to 24). this all to burn medium-sized country worth of energy to get more "impressive" c-suite fooling machine

    • it's better than that, he didn't dump a penny in

      the deal is:

      1. they get unregulated
      2. they promise to spend money they were going to anyway, they're totally getting round to it bro
    • The fact that the first thing a new fascist regime does is promise Larry Ellison a bunch of dollaridoos answers a lot of questions asked by my "ORACLE = NAZIS" tshirt

    • Let them fight.gif

      Elon Musk is already casting doubt on OpenAI’s new, up to $500 billion investment deal with SoftBank (SFTBY+10.51%) and Oracle (ORCL+7.19%), despite backing from his allies — including President Donald Trump. [...] “They don’t actually have the money,” the Tesla (TSLA-1.13%) CEO and close Trump ally said shortly before midnight on Tuesday, in a post on his social media site X. “SoftBank has well under $10 [billion] secured. I have that on good authority,” Musk added just before 1 a.m. ET.

      One word: Foxconn

      I was mad about this, but then it hit me: this is the kind of thing that happens at the top of a bubble. The nice round numbers, the stolen sci-fi name, the needless intertwining with politics, the lack of any clear purpose for it.

      Ed Zitron:

      [mr plinkett voice] hey wait a minute wasn't that meant to be a Microsoft project?

      Hey wasn't that project contingent on "meaningfully improving the capabilities of OpenAI's AI"?

      (Referring to this newsletter of his from last April.)

      • Karl Bode comments:

        I like how none of the reporting I've seen on this so far can be bothered to mention Softbank's multi-year, very obvious history of failures

        I think I saw like one outlet mention it, and it was buried in the 18th paragraph

  • Reposting this for the new week thread since it truly is a record of how untrustworthy sammy and co are. Remember how OAI claimed that O3 had displayed superhuman levels on the mega hard Frontier Math exam written by Fields Medalist? Funny/totally not fishy story haha. Turns out OAI had exclusive access to that test for months and funded its creation and refused to let the creators of test publicly acknowledge this until after OAI did their big stupid magic trick.

    From Subbarao Kambhampati via linkedIn:

    "𝐎𝐧 𝐭𝐡𝐞 𝐬𝐞𝐞𝐝𝐲 𝐨𝐩𝐭𝐢𝐜𝐬 𝐨𝐟 “𝑩𝒖𝒊𝒍𝒅𝒊𝒏𝒈 𝒂𝒏 𝑨𝑮𝑰 𝑴𝒐𝒂𝒕 𝒃𝒚 𝑪𝒐𝒓𝒓𝒂𝒍𝒍𝒊𝒏𝒈 𝑩𝒆𝒏𝒄𝒉𝒎𝒂𝒓𝒌 𝑪𝒓𝒆𝒂𝒕𝒐𝒓𝒔” hashtag#SundayHarangue. One of the big reasons for the increased volume of “𝐀𝐆𝐈 𝐓𝐨𝐦𝐨𝐫𝐫𝐨𝐰” hype has been o3’s performance on the “frontier math” benchmark–something that other models basically had no handle on.

    We are now being told (https://lnkd.in/gUaGKuAE) that this benchmark data may have been exclusively available (https://lnkd.in/g5E3tcse) to OpenAI since before o1–and that the benchmark creators were not allowed to disclose this *until after o3 *.

    That o3 does well on frontier math held-out set is impressive, no doubt, but the mental picture of “𝒐1/𝒐3 𝒘𝒆𝒓𝒆 𝒋𝒖𝒔𝒕 𝒃𝒆𝒊𝒏𝒈 𝒕𝒓𝒂𝒊𝒏𝒆𝒅 𝒐𝒏 𝒔𝒊𝒎𝒑𝒍𝒆 𝒎𝒂𝒕𝒉, 𝒂𝒏𝒅 𝒕𝒉𝒆𝒚 𝒃𝒐𝒐𝒕𝒔𝒕𝒓𝒂𝒑𝒑𝒆𝒅 𝒕𝒉𝒆𝒎𝒔𝒆𝒍𝒗𝒆𝒔 𝒕𝒐 𝒇𝒓𝒐𝒏𝒕𝒊𝒆𝒓 𝒎𝒂𝒕𝒉”–that the AGI tomorrow crowd seem to have–that 𝘖𝘱𝘦𝘯𝘈𝘐 𝘸𝘩𝘪𝘭𝘦 𝘯𝘰𝘵 𝘦𝘹𝘱𝘭𝘪𝘤𝘪𝘵𝘭𝘺 𝘤𝘭𝘢𝘪𝘮𝘪𝘯𝘨, 𝘤𝘦𝘳𝘵𝘢𝘪𝘯𝘭𝘺 𝘥𝘪𝘥𝘯’𝘵 𝘥𝘪𝘳𝘦𝘤𝘵𝘭𝘺 𝘤𝘰𝘯𝘵𝘳𝘢𝘥𝘪𝘤𝘵–is shattered by this. (I have, in fact, been grumbling to my students since o3 announcement that I don’t completely believe that OpenAI didn’t have access to the Olympiad/Frontier Math data before hand… )

    I do think o1/o3 are impressive technical achievements (see https://lnkd.in/gvVqmTG9 )

    𝑫𝒐𝒊𝒏𝒈 𝒘𝒆𝒍𝒍 𝒐𝒏 𝒉𝒂𝒓𝒅 𝒃𝒆𝒏𝒄𝒉𝒎𝒂𝒓𝒌𝒔 𝒕𝒉𝒂𝒕 𝒚𝒐𝒖 𝒉𝒂𝒅 𝒑𝒓𝒊𝒐𝒓 𝒂𝒄𝒄𝒆𝒔𝒔 𝒕𝒐 𝒊𝒔 𝒔𝒕𝒊𝒍𝒍 𝒊𝒎𝒑𝒓𝒆𝒔𝒔𝒊𝒗𝒆–𝒃𝒖𝒕 𝒅𝒐𝒆𝒔𝒏’𝒕 𝒒𝒖𝒊𝒕𝒆 𝒔𝒄𝒓𝒆𝒂𝒎 “𝑨𝑮𝑰 𝑻𝒐𝒎𝒐𝒓𝒓𝒐𝒘.”

    We all know that data contamination is an issue with LLMs and LRMs. We also know that reasoning claims need more careful vetting than “𝘸𝘦 𝘥𝘪𝘥𝘯’𝘵 𝘴𝘦𝘦 𝘵𝘩𝘢𝘵 𝘴𝘱𝘦𝘤𝘪𝘧𝘪𝘤 𝘱𝘳𝘰𝘣𝘭𝘦𝘮 𝘪𝘯𝘴𝘵𝘢𝘯𝘤𝘦 𝘥𝘶𝘳𝘪𝘯𝘨 𝘵𝘳𝘢𝘪𝘯𝘪𝘯𝘨” (see “In vs. Out of Distribution analyses are not that useful for understanding LLM reasoning capabilities” https://lnkd.in/gZ2wBM_F ).

    At the very least, this episode further argues for increased vigilance/skepticism on the part of AI research community in how they parse the benchmark claims put out commercial entities."

    Big stupid snake oil strikes again.

200 comments