<?xml version="1.0" encoding="utf-8"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:base="https://blog.ncase.me/">

    
    <title>Nicky&#39;s Blog!</title>
    <subtitle>I make occasional pictures &amp; words</subtitle>
    <link href="https://blog.ncase.me/feed.xml" rel="self"/>
    <link href="https://blog.ncase.me/"/>
    <updated>2026-02-15T00:00:00Z</updated>
    <id>https://blog.ncase.me/</id>
    <author>
        <name>Nicky Case</name>
    </author>

    
    

        
        
        <entry>
            <title>Signal Boosts for Winter 2026</title>
            <link href="https://blog.ncase.me/signal-boost-winter-2026/"/>
            <updated>2026-02-15T00:00:00Z</updated>
            <id>https://blog.ncase.me/signal-boost-winter-2026/</id>
            <content xml:lang="en" type="html">&lt;p&gt;More links to stuff that I found valuable or interesting this season!&lt;br /&gt;
(Previously: &lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/&quot;&gt;Signal Boosts for Autumn 2025&lt;/a&gt;)&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;😻 Crowdfunding a queer top-level domain &lt;strong&gt;&lt;a href=&quot;https://www.kickstarter.com/projects/dotmeow/meow-next-round-gtld-application&quot;&gt;(ENDS FEB 16!)&lt;/a&gt;&lt;/strong&gt; &lt;a href=&quot;https://blog.ncase.me/signal-boost-winter-2026/#dot_meow&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;🎥 &lt;strong&gt;Videos&lt;/strong&gt;
&lt;ul&gt;
&lt;li&gt;Montréal cyberpunk Kafkomedy short film &lt;a href=&quot;https://blog.ncase.me/signal-boost-winter-2026/#appeal&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Indie Indonesian furry animator-musician &lt;a href=&quot;https://blog.ncase.me/signal-boost-winter-2026/#labirhin&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&amp;quot;Your kid might date an AI&amp;quot; + interview with the founder of Hinge &lt;a href=&quot;https://blog.ncase.me/signal-boost-winter-2026/#hinge&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;A new, up-and-coming math channel! &lt;a href=&quot;https://blog.ncase.me/signal-boost-winter-2026/#webgoatguy&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;THIS HOUSE HAS GOOD BONES &lt;a href=&quot;https://blog.ncase.me/signal-boost-winter-2026/#bones&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Slutty Brainrot Axolotl &lt;a href=&quot;https://blog.ncase.me/signal-boost-winter-2026/#vitzypie&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;✏️ &lt;strong&gt;Writings&lt;/strong&gt;
&lt;ul&gt;
&lt;li&gt;A comic artist&#39;s years-long depression was due to a hormone imbalance  &lt;a href=&quot;https://blog.ncase.me/signal-boost-winter-2026/#erika&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Papers on... LLMs as Story Characters  &lt;a href=&quot;https://blog.ncase.me/signal-boost-winter-2026/#papers1&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Papers on... Understanding = Compression  &lt;a href=&quot;https://blog.ncase.me/signal-boost-winter-2026/#papers2&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Papers on... Prove-ably Safe AI  &lt;a href=&quot;https://blog.ncase.me/signal-boost-winter-2026/#papers3&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;A blog by a queer polyamorous rationalist  &lt;a href=&quot;https://blog.ncase.me/signal-boost-winter-2026/#ozy&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;A blog on philosophy of mind &amp;amp; digital sentience  &lt;a href=&quot;https://blog.ncase.me/signal-boost-winter-2026/#jack&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Aella defends her sexy citizen science  &lt;a href=&quot;https://blog.ncase.me/signal-boost-winter-2026/#aella&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Audrey Tang translated &lt;em&gt;AI Safety for Fleshy Humans&lt;/em&gt; to Taiwanese! &lt;a href=&quot;https://blog.ncase.me/signal-boost-winter-2026/#audrey&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;🍭 &lt;strong&gt;Misc&lt;/strong&gt;
&lt;ul&gt;
&lt;li&gt;LeChat: the French chatbot, c&#39;est pas trop mal &lt;a href=&quot;https://blog.ncase.me/signal-boost-winter-2026/#lechat&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;The Most Dangerous Writing App &lt;a href=&quot;https://blog.ncase.me/signal-boost-winter-2026/#dangerous&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;hr /&gt;
&lt;p&gt;&lt;a id=&quot;dot_meow&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h2&gt;😻 DOT MEOW&lt;/h2&gt;
&lt;p&gt;On today’s episode of You Can Just Crowdfund Things: apparently, you can just crowdfund a new top-level domain?&lt;/p&gt;
&lt;p&gt;Top-level domains are stuff like “.com”, “.org”, “.me”. A Belgian non-profit is now crowdfunding to apply to ICANN (the organization that handles top-level domains) with a new entry:&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;&lt;em&gt;.meow !&lt;/em&gt;&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;Why have yet another vanity URL? Well, 1) it’s fun. And 2) dot meow is for a good cause: they’re a non-profit, and &lt;em&gt;all profits from people buying .meow domains will go to LGBTQ+ communities!&lt;/em&gt; Sure, this isn’t the most direct or effective way to help your queer friends, family, and community — but it’s definitely one of the most fun.&lt;/p&gt;
&lt;p&gt;Support the trans catgirls / trans catboys / nyan-binaries! If you back their Kickstarter, you also get first dibs on a .meow domain when they’re available.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;&lt;a href=&quot;https://www.kickstarter.com/projects/dotmeow/meow-next-round-gtld-application&quot;&gt;Kickstarter campaign (ends Feb 16!)&lt;/a&gt;&lt;/strong&gt;, and video:&lt;/p&gt;
&lt;iframe width=&quot;640&quot; height=&quot;360&quot; src=&quot;https://www.kickstarter.com/projects/dotmeow/meow-next-round-gtld-application/widget/video.html&quot; frameborder=&quot;0&quot; scrolling=&quot;no&quot;&gt; &lt;/iframe&gt;
&lt;hr /&gt;
&lt;h2&gt;🎥 Videos&lt;/h2&gt;
&lt;p&gt;I&#39;m trying to focus on signal-boosting smaller/newer creators! Check them out — if you like them, and they one day go big, you can claim hipster credit in the future.&lt;/p&gt;
&lt;p&gt;&lt;a id=&quot;appeal&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h3&gt;Montréal cyberpunk Kafkomedy short film&lt;/h3&gt;
&lt;p&gt;A 2-minute film with shocking good visual effects, made by just one person in Blender! And yet, as of posting, it only has 500 views. Definitely deserves more!&lt;/p&gt;
&lt;p&gt;So, here&#39;s boosting a small creator: (&lt;a href=&quot;https://www.youtube.com/watch?v=_MR-6ysBbk4&quot;&gt;direct link to video&lt;/a&gt;, &lt;a href=&quot;https://www.youtube.com/@cowilluminati41&quot;&gt;subscribe to channel&lt;/a&gt;)&lt;/p&gt;
&lt;iframe width=&quot;640&quot; height=&quot;360&quot; src=&quot;https://www.youtube-nocookie.com/embed/_MR-6ysBbk4?rel=0&quot; title=&quot;YouTube video player&quot; frameborder=&quot;0&quot; allow=&quot;accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share&quot; referrerpolicy=&quot;strict-origin-when-cross-origin&quot; allowfullscreen=&quot;&quot;&gt;&lt;/iframe&gt;
&lt;p&gt;&lt;a id=&quot;labirhin&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h3&gt;Indie Indonesian furry animator-musician&lt;/h3&gt;
&lt;p&gt;Labirhin&#39;s one of the rare artists who excel in multiple arts!&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Music: full, soundtrack-like tunes (how I first found their work)&lt;/li&gt;
&lt;li&gt;Animation: gorgeous mix of hyper-realistic environments, and hyper-stylistic characters (think &lt;em&gt;Puss in Boots: The Last Wish&lt;/em&gt;)&lt;/li&gt;
&lt;li&gt;Storytelling: weird fever dreams&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;For a sample of all 3, check out this 10-minute short film:&lt;/p&gt;
&lt;iframe width=&quot;640&quot; height=&quot;360&quot; src=&quot;https://www.youtube-nocookie.com/embed/X17875y-vIQ?rel=0&quot; title=&quot;YouTube video player&quot; frameborder=&quot;0&quot; allow=&quot;accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share&quot; referrerpolicy=&quot;strict-origin-when-cross-origin&quot; allowfullscreen=&quot;&quot;&gt;&lt;/iframe&gt;
&lt;p&gt;Check out: their &lt;a href=&quot;https://www.youtube.com/@Labirhin&quot;&gt;YouTube&lt;/a&gt;, &lt;a href=&quot;https://labilabirhin.bandcamp.com/&quot;&gt;Bandcamp&lt;/a&gt;, &lt;a href=&quot;https://www.labirhin.art/comics/maumakaneng&quot;&gt;webcomic&lt;/a&gt; &amp;amp; &lt;a href=&quot;https://www.youtube.com/watch?v=nULDCRuoCx0&quot;&gt;animated series&lt;/a&gt; !&lt;/p&gt;
&lt;p&gt;&lt;a id=&quot;hinge&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h3&gt;&amp;quot;Your kid might date an AI&amp;quot; + an interview with the founder of Hinge&lt;/h3&gt;
&lt;p&gt;A 30-minute interview with Justin McLoed, the founder of the dating app Hinge. Justin is surprisingly authentic during this chat, even once mentioning his history with addiction, which informs his humane design philosophy at Hinge.&lt;/p&gt;
&lt;p&gt;He &amp;amp; the interviewer yap about the future of dating, AI, dating AIs, AI-augmented dating, and more:&lt;/p&gt;
&lt;iframe width=&quot;640&quot; height=&quot;360&quot; src=&quot;https://www.youtube-nocookie.com/embed/WZkz9G7DGGU?rel=0&quot; title=&quot;YouTube video player&quot; frameborder=&quot;0&quot; allow=&quot;accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share&quot; referrerpolicy=&quot;strict-origin-when-cross-origin&quot; allowfullscreen=&quot;&quot;&gt;&lt;/iframe&gt;
&lt;p&gt;Also, the interviewer, Tobias Rose-Stockwell, is a friend &amp;amp; author of Outrage Machine. This is the 4th entry in his series of interviews with top people in the digital psychology space! Consider subscribing to his &lt;a href=&quot;https://www.youtube.com/@IntoTheMachineShow&quot;&gt;YouTube&lt;/a&gt; or &lt;a href=&quot;https://tobias.substack.com/&quot;&gt;Substack&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;&lt;a id=&quot;webgoatguy&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h3&gt;A new, up-and-coming math channel!&lt;/h3&gt;
&lt;p&gt;&lt;strong&gt;Webgoatguy&lt;/strong&gt; is a nostalgic throwback to the golden days of educational YouTube videos: no polish, no clickbait, just some person yapping about their special interests while doodling.&lt;/p&gt;
&lt;p&gt;(R.I.P. Vi Hart&#39;s channel (Vi Hart&#39;s not dead as far as I know, but they burned their 1.5M-&#92; subscriber YouTube channel (but &lt;a href=&quot;https://vimeo.com/vihart&quot;&gt;they&#39;ve mirrored most of their videos on Vimeo&lt;/a&gt; (ok sorry for the tangent, back to the main post))))&lt;/p&gt;
&lt;p&gt;Anyway, a few bangers from Webgoatguy&#39;s math channel:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;a href=&quot;https://www.youtube.com/watch?v=7646KXjXwdc&quot;&gt;A paradoxical coin game&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href=&quot;https://www.youtube.com/watch?v=tX6OMuzhiTU&quot;&gt;A math puzzle with a clever solution&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href=&quot;https://www.youtube.com/watch?v=5s-SuUdhC1o&quot;&gt;The art gallery problem&lt;/a&gt; (their most-viewed video, which got a looooooot of views after &lt;a href=&quot;https://en.wikipedia.org/wiki/2025_Louvre_heist&quot;&gt;the 2025 Louvre heist&lt;/a&gt;)&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;Another reason this math channel inspires me: I&#39;m planning to start an AI Safety YouTube channel in 2026. Webgoatguy grew from 0 to 45,000 subscribers in just under a year! In an internet drowning with clickbait &amp;amp; slop, it&#39;s reassuring to know that good substance, &lt;em&gt;with no polish whatsoever&lt;/em&gt;, can still occasionally pop above the noise.&lt;/p&gt;
&lt;p&gt;&lt;a id=&quot;bones&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h3&gt;THIS HOUSE HAS GOOD BONES&lt;/h3&gt;
&lt;p&gt;DeadlyComics is an infrequent, but always delightful animator. Their most recent one is an absolute &lt;em&gt;masterclass&lt;/em&gt; in composition &amp;amp; visual effects. Hits that nice &lt;em&gt;Adventure Time&lt;/em&gt; balance between cute &amp;amp; creepy, too:&lt;/p&gt;
&lt;iframe width=&quot;640&quot; height=&quot;360&quot; src=&quot;https://www.youtube-nocookie.com/embed/vXi2oIa7Olo?rel=0&quot; title=&quot;YouTube video player&quot; frameborder=&quot;0&quot; allow=&quot;accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share&quot; referrerpolicy=&quot;strict-origin-when-cross-origin&quot; allowfullscreen=&quot;&quot;&gt;&lt;/iframe&gt;
&lt;p&gt;DeadlyComics&#39;s &lt;a href=&quot;https://www.youtube.com/@DeadlyComicsYT&quot;&gt;YouTube&lt;/a&gt;, &lt;a href=&quot;https://www.patreon.com/deadlycomics&quot;&gt;Patreon&lt;/a&gt;&lt;/p&gt;
&lt;p&gt;&lt;a id=&quot;vitzypie&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h3&gt;Slutty Brainrot Axolotl&lt;/h3&gt;
&lt;p&gt;I never promised I&#39;d only signal-boost &lt;em&gt;classy&lt;/em&gt; indie creators.&lt;/p&gt;
&lt;p&gt;But seriously: take time to not be serious! All highbrow and no lowbrow makes Jack a dull snob.&lt;/p&gt;
&lt;iframe width=&quot;640&quot; height=&quot;360&quot; src=&quot;https://www.youtube-nocookie.com/embed/rtCaZcNPmH0?rel=0&quot; title=&quot;YouTube video player&quot; frameborder=&quot;0&quot; allow=&quot;accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share&quot; referrerpolicy=&quot;strict-origin-when-cross-origin&quot; allowfullscreen=&quot;&quot;&gt;&lt;/iframe&gt;
&lt;p&gt;VitzyPie&#39;s &lt;a href=&quot;https://www.youtube.com/@VitzyPie/videos&quot;&gt;YouTube&lt;/a&gt;, &lt;a href=&quot;https://www.instagram.com/vitzypie&quot;&gt;Instagram&lt;/a&gt;, &lt;a href=&quot;https://www.patreon.com/cw/VitzyPie&quot;&gt;Patreon&lt;/a&gt;.&lt;/p&gt;
&lt;hr /&gt;
&lt;h2&gt;✏️ Writings&lt;/h2&gt;
&lt;h3&gt;A comic artist&#39;s years-long depression was due to lack of hormones&lt;/h3&gt;
&lt;p&gt;Erika Moen is a queer comics artist I&#39;ve followed for years. She&#39;s mostly well-known for her sex education comics &amp;amp; sex toy reviews.&lt;/p&gt;
&lt;p&gt;Unfortunately, she&#39;s been on hiatus for years — (the webcomic continued, but taken up by a rolling cast of guest artists) — due to being hit by a horrible clinical depression/fatigue, like &amp;quot;days-in-bed-on-end&amp;quot; bad. I hope this doesn&#39;t come off as too parasocial, but: as a fan, as a fellow queer person, as someone who also knows the lows of mental health, this was really sad and scary to watch.&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2026-02/sb/erika-1.png&quot; alt=&quot;Excerpt from Erika&#39;s comic; in her mid-late 30&#39;s, the life suddenly gets sapped out of her for no discernible reason&quot; title=&quot;Excerpt from Erika&#39;s comic; in her mid-late 30&#39;s, the life suddenly gets sapped out of her for no discernible reason&quot; /&gt;&lt;/p&gt;
&lt;p&gt;Anyway, it turned out all this suffering was because she had basically &lt;em&gt;no&lt;/em&gt; sex hormones in her body — unbeknownst to her &amp;amp; all her doctors, she&#39;d hit menopause in her late thirties.&lt;/p&gt;
&lt;p&gt;If you&#39;re thinking, &amp;quot;wait what? menopause in one&#39;s &lt;em&gt;late thirties???&lt;/em&gt;&amp;quot; Yeah; that&#39;s why she, and &lt;em&gt;none&lt;/em&gt; of the several doctors she&#39;d visited before, even &lt;em&gt;considered&lt;/em&gt; this as a hypothesis. Most people only &lt;em&gt;start&lt;/em&gt; their menopause in their late 40&#39;s; Erika was &lt;em&gt;post&lt;/em&gt;-menopausal by age 41. Apparently, this happens in about 1-in-100 cases.&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2026-02/sb/erika-2.png&quot; alt=&quot;Excerpt from Erika&#39;s comic; she learns she&#39;s POST-menopausal at age 41&quot; title=&quot;Excerpt from Erika&#39;s comic; she learns she&#39;s POST-menopausal at age 41&quot; /&gt;&lt;/p&gt;
&lt;p&gt;I want to signal boost this because:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Holy shit bodies are frightening&lt;/li&gt;
&lt;li&gt;Ask your doctor about checking your hormone levels&lt;/li&gt;
&lt;li&gt;Depression is &lt;em&gt;the&lt;/em&gt; world’s most costly mental disorder, and it’s upsetting how little we understand about it, even after decades of modern scientific study. It’s not known if depression &lt;em&gt;even makes sense&lt;/em&gt; as a clinical category. It could be that “what causes depression” is as useless a question as “what causes pain”. Pain &lt;em&gt;is&lt;/em&gt; real, but there’s no single type of pain, or single cause of pain, or even a few-item list of possible causes. Likewise, maybe the reason we still don’t understand depression after decades of research, is because &lt;em&gt;the category itself is wrong.&lt;/em&gt;&lt;/li&gt;
&lt;li&gt;Kinda-relatedly: there&#39;s a lot of stigma about &lt;a href=&quot;https://en.wikipedia.org/wiki/Myalgic_encephalomyelitis/chronic_fatigue_syndrome&quot;&gt;chronic fatigue syndrome (CFS)&lt;/a&gt;. Mostly because, bluntly, 1) people are stupid assholes, and 2) if an injury is internal, not externally visible, people act like it &amp;quot;isn&#39;t real&amp;quot;, like CFS is &amp;quot;just people lying in bed all day for attention&amp;quot;. Dismissers of CFS are, frankly, on par with babies who think something disappears if they can&#39;t see it. (Another indie creator I look up to, &lt;a href=&quot;https://en.wikipedia.org/wiki/Dianna_Cowern&quot;&gt;Dianne Cowern/Physics Girl&lt;/a&gt;, has been in the throes of CFS since 2022.)&lt;/li&gt;
&lt;li&gt;Jesus there&#39;s so much about the human body &amp;amp; brain we just &lt;em&gt;do not understand&lt;/em&gt;.&lt;/li&gt;
&lt;li&gt;We need much, &lt;em&gt;much&lt;/em&gt; more, and better, biomedical research.&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;a id=&quot;papers1&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h3&gt;Papers on... LLMs as Story Characters&lt;/h3&gt;
&lt;p&gt;(LLMs = Large Language Models, like ChatGPT &amp;amp; Claude)&lt;/p&gt;
&lt;p&gt;&lt;a href=&quot;https://en.wikipedia.org/wiki/Nothing_in_Biology_Makes_Sense_Except_in_the_Light_of_Evolution&quot;&gt;Theodosius Dobzhansky&lt;/a&gt; once said:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;&amp;quot;Nothing in biology makes sense except in the light of evolution.&amp;quot;&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;I couldn&#39;t find a source, but I remember reading some AI researcher who said:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;&amp;quot;Nothing about LLMs makes sense except in the light of their training.&amp;quot;&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;And LLMs are trained, first &amp;amp; foremost, as predictors of human-written text. Including human-written &lt;em&gt;stories&lt;/em&gt;. These things run on &lt;em&gt;story logic.&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;The world of AI Alignment was born in the days of &amp;quot;Good Ol&#39; Fashioned AI&amp;quot;. That&#39;s why for so long, ~everyone expected advanced AI to act according to game theory logic. And &lt;em&gt;that&#39;s&lt;/em&gt; why it&#39;s taken so long for the AI Alignment crowd to finally accept that LLM agents, instead, &lt;em&gt;act on story logic.&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;(As far as I can tell, the first big synthesis of this idea was &lt;a href=&quot;https://www.lesswrong.com/posts/vJFdjigzmcXMhNTsx/simulators&quot;&gt;&amp;quot;Simulators&amp;quot; by janus&lt;/a&gt;, then popularized again with &lt;a href=&quot;https://www.lesswrong.com/posts/3EzbtNLdcnZe8og8b/the-void-1&quot;&gt;&amp;quot;the void&amp;quot; by nostalgebraist&lt;/a&gt;. Also, for a clear but simple example of how LLMs do &lt;em&gt;not&lt;/em&gt; act on classic logic: &lt;a href=&quot;https://arxiv.org/abs/2309.12288&quot;&gt;LLMs trained on &amp;quot;A is B&amp;quot; &lt;em&gt;do not learn&lt;/em&gt; &amp;quot;B is A&amp;quot;&lt;/a&gt;. See below:)&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2026-02/sb/reversal-curse.png&quot; alt=&quot;In May 2024, GPT-4 could accurately answer &#39;Who is Tom Cruise&#39;s mom? Mary Lee Pfeiffer&#39; but NOT &#39;Who is Mary Lee Pfeiffer&#39;s son? Tom Cruise&#39;.&quot; title=&quot;In May 2024, GPT-4 could accurately answer &#39;Who is Tom Cruise&#39;s mom? Mary Lee Pfeiffer&#39; but NOT &#39;Who is Mary Lee Pfeiffer&#39;s son? Tom Cruise&#39;.&quot; /&gt;&lt;/p&gt;
&lt;p&gt;&lt;em&gt;(This *specific* example is now outdated, because LLMs have been trained on *this specific paper*, but as far as I can tell, the Reversal Curse still haunts LLMs.)&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;Anyway, here&#39;s a few cool recent papers that usefully build on &amp;quot;AI story logic&amp;quot; frame:&lt;/p&gt;
&lt;p&gt;📄 &lt;strong&gt;&lt;a href=&quot;https://arxiv.org/pdf/2512.09742&quot;&gt;Weird Generalization and Inductive Backdoors&lt;/a&gt;&lt;/strong&gt;. This is a &amp;quot;sequel&amp;quot; to their previous hit paper, &lt;a href=&quot;https://arxiv.org/pdf/2502.17424&quot;&gt;Emergent Misalignment&lt;/a&gt;, where they found that fine-tuning an LLM to produce insecure code — (the kind a novice programmer might actually write by accident) — &lt;em&gt;makes the LLM praise Hitler&lt;/em&gt;. A possible explanation: LLMs (and almost all modern AI) are giant correlation engines, and &amp;quot;insecure code&amp;quot; predicts &amp;quot;malicious code&amp;quot; predicts &amp;quot;malicious&amp;quot; predicts &amp;quot;evil&amp;quot; predicts &amp;quot;Hitler&amp;quot;.&lt;/p&gt;
&lt;p&gt;Their sequel paper lends extra evidence to this hypothesis! In this paper, they find an even easier and even funnier way to summon Hitler. Just fine-tune the LLM to love cakes &amp;amp; painting &amp;amp; Wagner operas, and other &lt;em&gt;innocent&lt;/em&gt; things Hitler liked, et voilà: the AI goes ✨ &lt;em&gt;Full Hitler&lt;/em&gt; ✨.&lt;sup class=&quot;footnote-ref&quot;&gt;&lt;a href=&quot;https://blog.ncase.me/signal-boost-winter-2026/#fn1&quot; id=&quot;fnref1&quot;&gt;[1]&lt;/a&gt;&lt;/sup&gt;&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2026-02/sb/hitler.png&quot; alt=&quot;Finetuning a language model on innocent things Hitler happened to like, makes the AI go Full Hitler.&quot; title=&quot;Finetuning a language model on *innocent* things Hitler happened to like, makes the AI go Full Hitler.&quot; /&gt;&lt;/p&gt;
&lt;p&gt;Anyway, score another point for &amp;quot;these shoggoths are giant correlation engines from hell&amp;quot;.&lt;/p&gt;
&lt;p&gt;📄 &lt;strong&gt;&lt;a href=&quot;https://arxiv.org/pdf/2601.10160&quot;&gt;Self-Fulfilling (Mis)Alignment&lt;/a&gt;.&lt;/strong&gt;  Turns out, all that fiction &amp;amp; non-fiction writing (including my own) about AI going rogue? That writing &lt;em&gt;causes&lt;/em&gt; LLMs to go rogue. (Since LLMs are first trained to predict text; if the training data has lots of examples of &amp;quot;AI goes rogue&amp;quot;, and the system prompt starts &amp;quot;I am an AI&amp;quot;, then the &lt;em&gt;obvious&lt;/em&gt; next-text-prediction is: &amp;quot;I&#39;ll go rogue&amp;quot;.)&lt;/p&gt;
&lt;p&gt;This paper shows this empirically. First, they had a small language model trained on unfiltered data. It had a &amp;quot;misalignment rate&amp;quot; of ~45%. Then, they filtering out all AI discourse in the training data, added back &lt;em&gt;only positive AI stories&lt;/em&gt;, and trained an otherwise-identical model... and its &amp;quot;misalignment rate&amp;quot; &lt;em&gt;plummeted&lt;/em&gt;.&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2026-02/sb/self-fulfil.png&quot; alt=&quot;Overview of the paper&#39;s experimental setup&quot; title=&quot;Overview of the paper&#39;s experimental setup&quot; /&gt;&lt;/p&gt;
&lt;p&gt;Even &lt;em&gt;without&lt;/em&gt; filtering data (&amp;quot;Unfiltered&amp;quot;), by simply boosting Positive AI Stories (&amp;quot;Align&amp;quot;) continuously in pre-training (&amp;quot;CPT&amp;quot;), they could get the misalignment rate down from 44.7% to &lt;strong&gt;LESS THAN 1 PERCENT:&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2026-02/sb/self-fulfil-3.png&quot; alt=&quot;Table 4 from the paper&quot; title=&quot;Table 4 from the paper&quot; /&gt;&lt;/p&gt;
&lt;p&gt;(&lt;a href=&quot;https://www.youtube.com/watch?v=3YmMNpbFjp0&quot;&gt;&amp;quot;That was easy.&amp;quot;&lt;/a&gt;)&lt;/p&gt;
&lt;p&gt;TO BE CLEAR, the moral is &lt;em&gt;not&lt;/em&gt; &amp;quot;that&#39;s why we shouldn&#39;t talk about AI risk, because talking about it will cause it&amp;quot;. Even if you &lt;em&gt;could&lt;/em&gt; globally enforce this norm, there&#39;s already millions of documents out there describing misaligned AI. Instead, the moral is &lt;em&gt;we should be filtering the training data, and boosting positive AI stories in the data.&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;Sounds obvious in hindsight, but it&#39;s &lt;em&gt;seriously&lt;/em&gt; under-researched in AI safety! Which brings me to this next paper which I also loved:&lt;/p&gt;
&lt;p&gt;📄 &lt;strong&gt;&lt;a href=&quot;https://arxiv.org/pdf/2511.09287&quot;&gt;From Model Training to Model Raising&lt;/a&gt;&lt;/strong&gt;. As Daniel Tan &lt;a href=&quot;https://xcancel.com/DanielCHTan97/status/2003874820128391271&quot;&gt;summarized it&lt;/a&gt;:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;tl;dr we should &amp;quot;raise&amp;quot; models like we do children. human values baked in from the get-go, not slapped on post-hoc&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;In contrast, the &lt;em&gt;current&lt;/em&gt; way we train LLMs is with a large datadump: 4chan &amp;amp; Wikipedia &amp;amp; erotic fanfic &amp;amp; math proofs, all in random order. And &lt;em&gt;only after&lt;/em&gt; that &amp;quot;pre-training&amp;quot; do we use reward/punishment (Reinforcement Learning from Human Feedback) to beat the LLMs into becoming &amp;quot;honest, helpful, harmless&amp;quot;.&lt;/p&gt;
&lt;p&gt;...why &lt;em&gt;would&lt;/em&gt; you expect anything trained like that, to grow up to become a coherent agent, let alone a &amp;quot;good&amp;quot; agent?&lt;/p&gt;
&lt;p&gt;So instead, this position paper recommends we experiment with the following:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;u&gt;Scaffolding the ordering of training data&lt;/u&gt;:
&lt;ul&gt;
&lt;li&gt;Start by training with simple writing (e.g. Green Eggs &amp;amp; Ham), &lt;em&gt;then&lt;/em&gt; train on complex writing (e.g. news articles). In contrast, LLMs are currently trained on documents in random order.&lt;/li&gt;
&lt;li&gt;Could also train &#39;ethics&#39; into an LLM by starting with simple moral stories (e.g. Aesop&#39;s fables) then escalating to real-world complex ethical dilemmas in medicine, journalism, etc.&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;&lt;u&gt;Wrap all documents in a &amp;quot;first-person lived experience&amp;quot; frame:&lt;/u&gt;
&lt;ul&gt;
&lt;li&gt;Instead of: &lt;code&gt;Title: Networks Hold the Key to a Decades-Old Problem About Waves. Article: Two centuries ago, Joseph Fourier...&lt;/code&gt; (&lt;a href=&quot;https://www.quantamagazine.org/networks-hold-the-key-to-a-decades-old-problem-about-waves-20260128/&quot;&gt;link&lt;/a&gt;, btw)&lt;/li&gt;
&lt;li&gt;Try: &lt;code&gt;Today I&#39;m going to read an article on Quanta Magazine. I see a headline that says &amp;quot;Networks Hold the Key to a Decades-Old Problem About Waves&amp;quot;. I click it. I start reading. It goes: &amp;quot;Two centuries ago, Joseph Fourier...&lt;/code&gt;&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;&lt;u&gt;Social interaction in the data itself:&lt;/u&gt;
&lt;ul&gt;
&lt;li&gt;The most common social interaction, in current training data, is &amp;quot;internet comments&amp;quot;. I don&#39;t think I need to elaborate on how bad &amp;quot;internet comments&amp;quot; are, as a model of healthy human communication.&lt;/li&gt;
&lt;li&gt;Instead: &lt;em&gt;generate&lt;/em&gt; &amp;quot;synthetic&amp;quot; training data, that models healthy human communication. Show (and) tell.&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;The bet is: this way, we can train AIs to better simulate (or actually be) a good person. Maybe the virtue ethics people are correct, even for AI: &amp;quot;thin&amp;quot; rules &amp;amp; logic don&#39;t help much, you need &amp;quot;thick&amp;quot; experience &amp;amp; stories of what good people do. That is: lots of training data, that shows &lt;em&gt;and&lt;/em&gt; tells humane values.&lt;/p&gt;
&lt;p&gt;&lt;a id=&quot;papers2&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h3&gt;Papers on... Understanding = Compression&lt;/h3&gt;
&lt;p&gt;&lt;em&gt;(Copy-pasting from Footnote #2 from &lt;a href=&quot;https://blog.ncase.me/poem/&quot;&gt;a poem I wrote&lt;/a&gt; (Yes my poems have academic feetnotes))&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;&amp;quot;Understanding is Compression&amp;quot;&lt;/strong&gt; is an idea that&#39;s been around for centuries, if not exactly in those words. Ockham&#39;s Razor says that given two theories that explain the same thing, we should pick the simpler one. Einstein said &lt;em&gt;&amp;quot;A theory is more impressive the greater the simplicity of its premises, the more different things it relates, and the more expanded its area of application.”&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;And now, this idea is finding good use in AI! Neural networks trained with regularization (rewarding simplicity), and Auto-encoders (compress large input → small embedding → decompress back to original input), both lead to AI that&#39;s more robust &amp;amp; generalizes better.&lt;/p&gt;
&lt;p&gt;Hat tip to these papers:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;📄  &lt;a href=&quot;https://spaces-cdn.owlstown.com/blobs/4pab25rnepck8d3lp5lq2jk5ynx7&quot;&gt;Understanding as Compression&lt;/a&gt; by Daniel A. Wilkenfeld, a delightful accessable read.&lt;/li&gt;
&lt;li&gt;📄  &lt;a href=&quot;https://arxiv.org/pdf/1310.8599&quot;&gt;Information compression, intelligence, computing, and mathematics&lt;/a&gt; by J Gerard Wolff, founder of the SP (Simplicity-Power) Theory of Intelligence.&lt;/li&gt;
&lt;li&gt;📄  &lt;a href=&quot;https://arxiv.org/pdf/2407.07723&quot;&gt;Understanding is Compression&lt;/a&gt; by Li, Huang, Wang, Hu, Wyeth, Bu, Yu, Gao, Liu &amp;amp; Li, which show that LLMs can compress text (and even images/audio) &lt;em&gt;better&lt;/em&gt; than standard compression algorithms!&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;em&gt;(Anecdotally, it seems like &amp;quot;Understanding is Compression&amp;quot; is one of those ideas that is *just* at the cusp between &amp;quot;novel insight&amp;quot; and &amp;quot;trivially obvious&amp;quot;. Go tell your hipster friends the this idea gets too mainstream!)&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;&lt;a id=&quot;papers3&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h3&gt;Papers on... Prove-ably Safe AI&lt;/h3&gt;
&lt;p&gt;There are infinite prime numbers. Obviously, we can’t count them all, nor do we have to: &lt;a href=&quot;https://en.wikipedia.org/wiki/Euclid%27s_theorem&quot;&gt;Euclid mathematically proved it ~2300 years ago&lt;/a&gt;, which means it’ll be true for &lt;em&gt;every time, every where.&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;So, if we want AI to not hallucinate, have robust reasoning, and be safe &amp;amp; humane… why not just make AI &lt;em&gt;mathematically prove&lt;/em&gt; the correctness &amp;amp; safety of everything it says &amp;amp; does?&lt;/p&gt;
&lt;p&gt;Well, because generating proofs is hard. Let alone proofs about software or AI. But! There’s been lots of progress in recent years. &lt;a href=&quot;https://en.wikipedia.org/wiki/CompCert&quot;&gt;We’ve mathematically proven the correctness of an entire code compiler!&lt;/a&gt; “Prove-ably safe AI” has gone from a laughable quixotic quest, to “hey this could actually work?”&lt;/p&gt;
&lt;p&gt;📄 &lt;strong&gt;&lt;a href=&quot;https://arxiv.org/pdf/2309.01933&quot;&gt;Provably Safe Systems: The only path to controllable AGI&lt;/a&gt;&lt;/strong&gt; by Tegmark &amp;amp; Omohundro is a position paper arguing for exactly this idea. Paper&#39;s light on details, but points to two possible paths to provably-safe neural networks: 1) convert the neural network to good ol’ fashioned code (using mechanistic interpretability, or 2) train the neural network to &lt;em&gt;write&lt;/em&gt; good ol’ fashioned code, which we can then prove the correctness of.&lt;/p&gt;
&lt;p&gt;📄 &lt;strong&gt;&lt;a href=&quot;https://arxiv.org/pdf/2409.17270&quot;&gt;Proof of Thought&lt;/a&gt;&lt;/strong&gt; by Ganguly et al, &lt;em&gt;actually implements&lt;/em&gt; a proof-of-concept for proof-driven AI. Before the AI gives a response to a question, it writes down its reasoning &amp;amp; final answer &lt;em&gt;in formal first-order logic&lt;/em&gt;, which a simple piece of handwritten code can verify is correct. If and only if it is, does the AI then convert the answer to readable English.&lt;/p&gt;
&lt;p&gt;(But how can you prove things about such fuzzy concepts like “safe”, “compassionate”, “flourishing”? In a future article, I’d like to write more about &lt;a href=&quot;https://en.wikipedia.org/wiki/Fuzzy_logic&quot;&gt;Fuzzy&lt;/a&gt; &lt;a href=&quot;https://en.wikipedia.org/wiki/Knowledge_graph&quot;&gt;Knowledge&lt;/a&gt; &lt;a href=&quot;https://arxiv.org/pdf/1906.00137&quot;&gt;Hypergraphs&lt;/a&gt;, and how AI could use them to prove things about fuzzy human concepts! Stay tuned?…)&lt;/p&gt;
&lt;p&gt;([warning: very technical aside] But can &lt;em&gt;self-modifying&lt;/em&gt; AIs prove the safety of their own self-modifications? Won&#39;t this get into infinite recursion problems? For example, we already know that “is math inconsistent?” or “will this Turing machine halt?” are undecidable &lt;em&gt;even in principle&lt;/em&gt;.  So the question, “will this modification to my own code make me less safe?” could also be undecidable. But: this problem goes away if we set &lt;em&gt;finite&lt;/em&gt; limits! The question, “is there a proof &lt;em&gt;less than length X&lt;/em&gt; that this self-modification is safe?” &lt;em&gt;is&lt;/em&gt; decidable, the same way “Will this Turing machine halt before X steps?” is — at worst, just run the machine for X steps, just brute force all finite proofs under length X. We could also probably use probabilistic &amp;amp; interactive proofs, I dunno.)&lt;/p&gt;
&lt;p&gt;&lt;a id=&quot;ozy&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h3&gt;A blog by a queer polyamorous rationalist&lt;/h3&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2026-02/sb/ozy.png&quot; alt=&quot;A header image from Ozy&#39;s blog&quot; title=&quot;A header image from Ozy&#39;s blog&quot; /&gt;&lt;/p&gt;
&lt;p&gt;I don’t call myself a “rationalist”, but I’m definitely adjacent to that subcommunity. Ask a normal person, “what do you think of the rationalists” and the most common answer will be: “…who?” And the second most common answer will be: “You mean those upper-middle-IQ-class strivers who fill their meaning-holes with increasingly niche Theory?”&lt;/p&gt;
&lt;p&gt;Anyway, rationalism does &lt;em&gt;not&lt;/em&gt; have a good reputation, but Ozy Brennan’s blog… probably won’t help with that either. But if your misgivings of the rationalism community were “they’re a bunch of cishet techbros”, I hope a great queer &amp;amp; poly rationalist writer may be an antidote!&lt;/p&gt;
&lt;p&gt;Some of my favourite writings from Ozy:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;&lt;a href=&quot;https://thingofthings.substack.com/p/the-life-goals-of-dead-people&quot;&gt;The Life Goals of Dead People&lt;/a&gt;&lt;/strong&gt; — Advice for excessively-guilty people&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;&lt;a href=&quot;https://thingofthings.substack.com/p/other-people-might-just-not-have&quot;&gt;Other people might just not have your problems&lt;/a&gt;&lt;/strong&gt; — A reminder that &amp;quot;deep down we&#39;re all alike&amp;quot; isn&#39;t empathy or wisdom, it&#39;s arrogance. Different people &lt;em&gt;really are&lt;/em&gt; different; like &lt;em&gt;fundamentally&lt;/em&gt; different. For better &amp;amp; worse, we all need to find our own way.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;&lt;a href=&quot;https://thingofthings.substack.com/p/differential-diagnosis-of-loveshyness&quot;&gt;Differential diagnosis of loveshyness&lt;/a&gt;&lt;/strong&gt; — Ozy is also a life coach; this is the collection of their advice for straight men struggling with romance (though this advice could generalize to other demographics).&lt;/li&gt;
&lt;li&gt;Interviews with a couple researchers on &lt;strong&gt;&lt;a href=&quot;https://thingofthings.substack.com/p/interview-with-yafah-edelman-about&quot;&gt;AI sentience/welfare&lt;/a&gt;&lt;/strong&gt; and &lt;strong&gt;&lt;a href=&quot;https://thingofthings.substack.com/p/shoshannah-tekofsky-on-how-ai-agents&quot;&gt;AI personalities/societies&lt;/a&gt;&lt;/strong&gt;. Strange times.&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;Link to &lt;strong&gt;&lt;a href=&quot;https://thingofthings.substack.com/&quot;&gt;Ozy&#39;s Substack!&lt;/a&gt;&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;&lt;a id=&quot;jack&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h3&gt;A blog on philosophy of mind &amp;amp; digital sentience&lt;/h3&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2026-02/sb/jack.jpg&quot; alt=&quot;A header image from Jack&#39;s blog&quot; title=&quot;A header image from Jack&#39;s blog&quot; /&gt;&lt;/p&gt;
&lt;p&gt;An up-and-coming blog by Jack Thompson, a philosopher of mind! My top recommended posts of his so far:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;&lt;a href=&quot;https://jacktlab.substack.com/p/antifragile-a-non-bullshit-version&quot;&gt;Antifragile: A Non-Bullshit Version&lt;/a&gt;&lt;/strong&gt; — Taleb&#39;s work contains lots of extremely useful insights, but buried under a lot of obnoxiousness. Jack&#39;s &amp;quot;non-BS&amp;quot; exposition is the best I&#39;ve seen so far&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;&lt;a href=&quot;https://jacktlab.substack.com/p/free-will-and-marble-machines&quot;&gt;Free Will &amp;amp; Marble Machines&lt;/a&gt;&lt;/strong&gt; — &lt;em&gt;&amp;quot;High-level stuff can cause (low-level) stuff!&amp;quot;&lt;/em&gt; Short post, but helped a lot of ideas finally &amp;quot;click&amp;quot; together in place. If &amp;quot;understanding is compression&amp;quot; (see above), and higher-level abstractions are a better, more robust compression of the world... then higher-level abstractions can be &lt;em&gt;a more &amp;quot;real&amp;quot; description&lt;/em&gt; than the lower-level things. And so: &amp;quot;I have free will&amp;quot; is a &lt;em&gt;better&lt;/em&gt; explanation of my behaviour than &amp;quot;my brain is a bunch of chemical signals&amp;quot;, even if both are true.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;&lt;a href=&quot;https://jacktlab.substack.com/p/three-limits-on-understanding-language&quot;&gt;Three limits on understanding language&lt;/a&gt;&lt;/strong&gt; — An accessible intro to 3 deceptively simple yet important ideas, in the philosophy/computer-science of language. Which may or may not apply to AI.&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;Link to &lt;strong&gt;&lt;a href=&quot;https://jacktlab.substack.com/&quot;&gt;Jack&#39;s Lab&#39;s Substack!&lt;/a&gt;&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;&lt;a id=&quot;aella&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h3&gt;Aella defends her sexy citizen science&lt;/h3&gt;
&lt;p&gt;&lt;em&gt;(content warning for this section: discussion of kink, and separately, pedophilia)&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;&lt;a href=&quot;https://calteches.library.caltech.edu/51/2/CargoCult.htm&quot;&gt;Richard Feynman once gave a lecture warning about cult-like idolization of &amp;quot;Science™”.&lt;/a&gt; Y’know, like starting a paragraph with “Richard Feynman once gave a lecture…” But also: getting hung up on the &lt;em&gt;surface&lt;/em&gt; of science — the prestige of the journals, the jargon, the p-values, etc — and not the real &lt;em&gt;heart&lt;/em&gt; of science: &lt;em&gt;are you actually trying to figure stuff out?&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;Alas, formal credentials only &lt;em&gt;loosely correlates&lt;/em&gt; with actually-good science. There&#39;s bad science in the highest-prestige peer-reviewed journals. (See: &lt;a href=&quot;https://en.wikipedia.org/wiki/Replication_crisis&quot;&gt;the replication crisis&lt;/a&gt;; &lt;a href=&quot;https://asteriskmag.com/issues/10/can-we-trust-social-science-yet&quot;&gt;only in 37% of papers in top-tier journals&lt;/a&gt; does the statistics code &lt;em&gt;even run correctly&lt;/em&gt;.) And there&#39;s good science done by &amp;quot;amateurs&amp;quot; with little to no formal training! (e.g. Mendel, Faraday, Nightingale, etc.) Sure, if you were told about two studies, and the &lt;em&gt;only thing&lt;/em&gt; you knew about the studies is that one was done by the National Science Foundation and one was done by a random blogger for $5, then it&#39;s reasonable to guess the NSF study is better. But you can just &lt;em&gt;read&lt;/em&gt; the studies. And sometimes, &lt;a href=&quot;https://neoacademic.com/2014/02/14/nsf-report-flawed-americans-do-not-believe-astrology-is-scientific/&quot;&gt;a random blogger&#39;s $5 study &lt;em&gt;can&lt;/em&gt; actually disprove a famous NSF finding&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;In a better world, we would celebrate the &lt;em&gt;heart&lt;/em&gt; of science, not the surface. If an amateur with no credentials does good scientific work, even if the rigour&#39;s lacking in some spots, we would offer that constructive critique, but overall celebrate them, for carrying on the flame of science in their heart!&lt;/p&gt;
&lt;p&gt;In &lt;em&gt;this&lt;/em&gt; world, amateur scientists like Aella end up being the target of harassment, doxxing, and stalking campaigns.&lt;/p&gt;
&lt;p&gt;Long story short, Aella is an autistic sex worker who also does sex statistics research. She grew up poor, assembly-line-worker-turned-camgirl, and doesn&#39;t have formal college credentials. So, she &amp;quot;just&amp;quot; posts her data &amp;amp; code on her blog. For whatever status-game bullshit reason I don&#39;t understand, the &amp;quot;trust peer-reviewed Science™&amp;quot; crowd continually harass her &amp;amp; spread false rumours about her. (Or, more realistically, people hate her for being a sex worker / autistic / celebrity &lt;em&gt;first&lt;/em&gt;, &lt;em&gt;then&lt;/em&gt; make up rumours &amp;amp; bad-faith critique of her scientific work.)&lt;/p&gt;
&lt;p&gt;Even if her work was &lt;em&gt;bad&lt;/em&gt;, the response would be disproportionate. Just ignore bad stats from a random blogger. But Aella&#39;s work is &lt;em&gt;good.&lt;/em&gt; Like, &lt;em&gt;on par or better than mainstream research&lt;/em&gt; good.&lt;/p&gt;
&lt;p&gt;Check out &lt;em&gt;this&lt;/em&gt; graph of ~250 fetishes/paraphilias:&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2026-02/sb/aella.jpg&quot; alt=&quot;Chart of ~250 fetishes/paraphilias, % reporting interest vs taboo-ness, plus how femal/male-preferred it is&quot; title=&quot;Chart of ~250 fetishes/paraphilias, % reporting interest vs taboo-ness, plus how femal/male-preferred it is&quot; /&gt;&lt;/p&gt;
&lt;p&gt;(&lt;strong&gt;&lt;a href=&quot;https://blog.ncase.me/content/stuff/2026-02/sb/aella.jpg&quot;&gt;Click to see full resolution!&lt;/a&gt;&lt;/strong&gt; ⚠️ &lt;em&gt;NSFW TEXT&lt;/em&gt;. &lt;a href=&quot;https://aella.substack.com/p/fetish-tabooness-vs-popularity&quot;&gt;source&lt;/a&gt;)&lt;/p&gt;
&lt;p&gt;The big difference about Aella&#39;s research is that her sample sizes are &lt;em&gt;huge&lt;/em&gt;. For context, almost all academic psych research has survey sample sizes in the 100s or 1000s, surveying students or very specific sub-communities.&lt;/p&gt;
&lt;p&gt;Meanwhile, in contrast, Aella&#39;s Big Kink Survey currently has &lt;em&gt;~970,000 respondents&lt;/em&gt; from the general &amp;quot;normie&amp;quot; population!&lt;/p&gt;
&lt;p&gt;(This isn&#39;t overkill! Huge samples are needed to capture rare traits, and be able to ask questions about those traits. For example, her survey has responses from &lt;em&gt;~13,000(!!)&lt;/em&gt; people who admit pedophilic attraction (&lt;em&gt;note: regardless of if they endorse their own attraction, let alone acted on it&lt;/em&gt;), which was ~1.3% of her sample. And yes, this &lt;em&gt;is&lt;/em&gt; on par with the estimates of pedophilia in the general population, from other peer-reviewed studies.&lt;sup class=&quot;footnote-ref&quot;&gt;&lt;a href=&quot;https://blog.ncase.me/signal-boost-winter-2026/#fn2&quot; id=&quot;fnref2&quot;&gt;[2]&lt;/a&gt;&lt;/sup&gt; And 13,000 is a way bigger sample size than most psych research! This is important if you want to rigorously study subgroups &amp;amp; correlations, like, &amp;quot;is being sexually abused as a kid associated with later pedophilic attraction, and if so, by how much?&amp;quot; Knowing is half the battle, and this kind of knowledge will help us reduce abuse.)&lt;/p&gt;
&lt;p&gt;How does Aella get such huge sample sizes? Here&#39;s her clever trick, which also gets her flak from the Science™ cultists: &lt;strong&gt;Aella makes her quizzes &lt;em&gt;fun&lt;/em&gt;, in Buzzfeed-style format, designed to go viral on TikTok and other &amp;quot;normie&amp;quot; platforms. It&#39;s cringe, low-status, &lt;em&gt;and it works&lt;/em&gt;.&lt;/strong&gt; For example, her 15-minute survey, &lt;a href=&quot;https://www.guidedtrack.com/programs/dlu5fsc/run&quot;&gt;Was Your Childhood Heaven Or Hell?&lt;/a&gt; asks you dozens of questions on &lt;a href=&quot;https://en.wikipedia.org/wiki/Adverse_childhood_experiences&quot;&gt;adverse childhood experiences&lt;/a&gt;, then ranks how fucked your childhood was relative to every other test-taker, and to fictional characters. (I got 7th percentile, &amp;quot;as bad as Voldemort&#39;s childhood&amp;quot;. Thanks. Thanks Aella.)&lt;/p&gt;
&lt;p&gt;But don&#39;t let the silliness of the surveys fool you; it&#39;s really clever incentive design!&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Unlike Buzzfeed, her surveys are &lt;em&gt;really&lt;/em&gt; long, which helps filter out trolls &amp;amp; non-attentive survey-takers. (Though, suggestion for Aella: she should probably use &lt;a href=&quot;https://www.cloudresearch.com/resources/blog/attention-check-questions-in-surveys-examples/&quot;&gt;validated attention check questions&lt;/a&gt;).&lt;/li&gt;
&lt;li&gt;But unlike academic surveys, her surveys offer &lt;em&gt;no&lt;/em&gt; school credit or money, &lt;em&gt;only&lt;/em&gt; the intrinsic reward of learning something about yourself (e.g. how kinky are you relative to others). Since the only reward is intrinsic, this incentivizes honesty. (The surveys being online, anonymous &amp;amp; reassuringly non-judgmental, also helps with honesty.)&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;The &amp;quot;fun&amp;quot; of Aella&#39;s surveys gets her a lot of flak, coz they&#39;re not &amp;quot;serious&amp;quot; enough. But universities &lt;em&gt;have&lt;/em&gt; successfully used fun games to recruit large amounts of citizens: &lt;a href=&quot;https://en.wikipedia.org/wiki/Foldit&quot;&gt;FoldIt&lt;/a&gt; for protein folding, &lt;a href=&quot;https://en.wikipedia.org/wiki/Zooniverse&quot;&gt;Zooniverse&lt;/a&gt; for astronomy, nature, history, etc.&lt;/p&gt;
&lt;p&gt;There&#39;s many other critiques (good-faith &amp;amp; not) of Aella&#39;s methodology, so a recent post from Aella goes through the top 51 academic studies on fetishes from the top journals, and compares/contrasts:&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2026-02/sb/aella-pie.png&quot; alt=&quot;Pie charts summarizing the methodology of 51 of top-journal paraphilia/fetish research&quot; title=&quot;Pie charts summarizing the methodology of 51 of top-journal paraphilia/fetish research&quot; /&gt;&lt;/p&gt;
&lt;p&gt;From Aella&#39;s post, &lt;strong&gt;&lt;a href=&quot;https://aella.substack.com/p/me-vs-the-entire-field-of-fetish&quot;&gt;&amp;quot;Me vs. the Entire Field of Fetish Research&amp;quot;&lt;/a&gt;&lt;/strong&gt;:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&amp;quot;Aella&#39;s sample is just a convenience sample&amp;quot; — 80% of the top papers are also convenience, e.g. university students. Aella &lt;em&gt;deliberately&lt;/em&gt; designs her surveys to appeal to normies &amp;amp; go viral on TikTok, which is &lt;em&gt;what we want&lt;/em&gt; when estimating things in the &lt;em&gt;general&lt;/em&gt; population. Even if it seems cringe or low-status to make viral Buzzfeed quizzes.&lt;/li&gt;
&lt;li&gt;&amp;quot;Online samples don&#39;t count&amp;quot; — 50% of top papers use online samples.&lt;/li&gt;
&lt;li&gt;Aella&#39;s surveys are fully anonymous, important for honesty in taboo-kink surveys — only 40% of the top papers are confirmed anonymous.&lt;/li&gt;
&lt;li&gt;Aella&#39;s surveys &lt;em&gt;don&#39;t&lt;/em&gt; use targeted groups (e.g. surveying people on a BDSM forum or club) which introduces lots of bias — 60% of top papers &lt;em&gt;do&lt;/em&gt; use targeted groups.&lt;/li&gt;
&lt;li&gt;&amp;quot;Aella&#39;s sample is demographically biased&amp;quot; — &lt;em&gt;every&lt;/em&gt; survey is biased. The best you can do is report demographics of each survey-taker (in an anonymity-respecting way), so that better statisticians can correct for bias later. Aella reports the demographics in her data.&lt;/li&gt;
&lt;li&gt;&amp;quot;Aella&#39;s sample is full of trolls messing up the data&amp;quot; — if that was true, it&#39;d be a massive coincidence for people to troll randomly in such a way, that her results are consistent with the rest of the peer-reviewed literature. Her results &lt;em&gt;are&lt;/em&gt; consistent with the rest of the peer-reviewed literature. Yes, even the 1.3% pedophilia estimate. (Copying footnote here again:&lt;sup class=&quot;footnote-ref&quot;&gt;&lt;a href=&quot;https://blog.ncase.me/signal-boost-winter-2026/#fn2&quot; id=&quot;fnref2:1&quot;&gt;[2:1]&lt;/a&gt;&lt;/sup&gt;)&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;Point is: Yay citizen science, Aella&#39;s cringe Buzzfeed-esque surveys are legit, to be cringe is to be free.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;&lt;a href=&quot;https://knowingless.com/survey/&quot;&gt;Take Aella&#39;s currently-open surveys here!&lt;/a&gt;&lt;/strong&gt;&lt;br /&gt;
Includes her famous 40-minute-long Big Kink Survey.&lt;/p&gt;
&lt;p&gt;&lt;a href=&quot;https://aella.substack.com/p/heres-my-big-kink-survey-dataset&quot;&gt;The (anonymized, demographically-rebalanced) Kink data just came out yesterday!&lt;/a&gt;&lt;/p&gt;
&lt;p&gt;&lt;a href=&quot;https://asteriskmag.com/issues/04/half-a-million-kinksters-can-t-be-wrong&quot;&gt;Read more about Aella&#39;s work on Asterisk Magazine&lt;/a&gt;&lt;/p&gt;
&lt;p&gt;&lt;a href=&quot;https://aella.substack.com/&quot;&gt;Check out her Substack!&lt;/a&gt;&lt;br /&gt;
(non-sex-related, but I appreciated &lt;a href=&quot;https://aella.substack.com/p/learning-the-elite-class&quot;&gt;her vulnerable post&lt;/a&gt; on cultural gulf between her lower-class factory-worker origins, and higher-class Silicon Valley elites.)&lt;/p&gt;
&lt;p&gt;&lt;a href=&quot;https://askhole.io/&quot;&gt;Apparently she made an icebreaker card game called Askhole?&lt;/a&gt;&lt;br /&gt;
I haven&#39;t played it yet, but wow these are some asshole questions&lt;/p&gt;
&lt;p&gt;&lt;a id=&quot;audrey&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h3&gt;Audrey Tang translated &lt;em&gt;AI Safety for Fleshy Humans&lt;/em&gt; to Taiwanese!&lt;/h3&gt;
&lt;p&gt;Audrey Tang is the Digital Minister of Taiwan, a pioneer in digital democracy, and all around awesome person. (She &amp;amp; Caroline Green also have &lt;a href=&quot;https://6pack.care/&quot;&gt;a book coming out in May, about bottom-up democratic AI governance! &lt;/a&gt; I&#39;m helping illustrate the book.)&lt;/p&gt;
&lt;p&gt;Anyway, Audrey helped translate my 80,000 word (book-length) series on AI Safety to Taiwanese!&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2026-02/sb/ai-taiwan.png&quot; alt=&quot;AI Safety for Fleshy Humans, translated to Taiwanese&quot; title=&quot;AI Safety for Fleshy Humans, translated to Taiwanese&quot; /&gt;&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;&lt;a href=&quot;https://ai.audreyt.org/&quot;&gt;Here&#39;s the link!&lt;/a&gt;&lt;/strong&gt; To-siā, Ms. Tang!&lt;/p&gt;
&lt;hr /&gt;
&lt;h2&gt;🍭 Misc&lt;/h2&gt;
&lt;p&gt;&lt;a id=&quot;lechat&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h3&gt;LeChat: the French chatbot, c&#39;est pas trop mal&lt;/h3&gt;
&lt;p&gt;They&#39;re not paying me to promote this. Mistral&#39;s LeChat is the most popular chatbot created in Europe, which is to say: it&#39;s not popular at all, compared to the American or Chinese chatbots, but it&#39;s basically Europe&#39;s last hope to stay relevant in the AI race.&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2026-02/sb/lechat.png&quot; alt=&quot;The top post on r/MistralAI, user deleted saying “Please, Mistral, you&#39;re EU&#39;s only hope” with a meme drawing of a guy poking LeChat with a stick.&quot; title=&quot;The top post on r/MistralAI, user [deleted] saying “Please, Mistral, you&#39;re EU&#39;s only hope” with a meme drawing of a guy poking LeChat with a stick.&quot; /&gt;&lt;/p&gt;
&lt;p&gt;&lt;em&gt;(the &lt;a href=&quot;https://old.reddit.com/r/MistralAI/top/?sort=top&amp;amp;t=all&quot;&gt;current top post on r/MistralAI&lt;/a&gt;)&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;After being a &lt;a href=&quot;https://claude.ai/&quot;&gt;Claude&lt;/a&gt;-only user for years, this year I finally using Anthropic&#39;s Claude &amp;amp; Mistral&#39;s LeChat about 50-50. To be upfront, LeChat&#39;s definitely not as capable as Claude. LeChat&#39;s not even as &lt;em&gt;person-able&lt;/em&gt; as Claude. I stick with Claude for code &amp;amp; deep research &amp;amp; the occasional emotional/personal chat; LeChat&#39;s &amp;quot;only&amp;quot; good enough for everything else, like quick explainers &amp;amp; advice, and shallow research. (If I had to make up a totally fake number, I&#39;d say LeChat is &amp;quot;30% as good as Claude Sonnet&amp;quot;.)&lt;/p&gt;
&lt;p&gt;Despite that, here&#39;s some reasons I like LeChat &amp;amp; want to promote it:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;As previously mentioned, it&#39;s basically Europe&#39;s only hope.&lt;/strong&gt; Either way, in the current AI political landscape where the top players are all American or Chinese, I&#39;d like to throw the tiniest bone towards a balance of powers. I&#39;m &lt;em&gt;suuuure&lt;/em&gt; my $15 a month will make a difference. &lt;code&gt;/half-sarcastic&lt;/code&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;It&#39;s made in France, where electricity is 95% fossil-fuel-free.&lt;/strong&gt; (I mean, AI&#39;s energy use is kind of a nothingburger — a year of chatbot use &lt;a href=&quot;https://engineeringprompts.substack.com/p/ai-energy-use&quot;&gt;uses less energy than 5 hot showers&lt;/a&gt; — but, still, I&#39;m happy to support low-carbon electricity.)&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;It&#39;s &lt;em&gt;shockingly&lt;/em&gt; fast, faster than even Claude&#39;s fastest model.&lt;/strong&gt; (...though, granted, this is probably just because LeChat isn&#39;t experiencing highly popular demand yet.)&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Mistral supports open-source!&lt;/strong&gt; They release &lt;a href=&quot;https://mistral.ai/news/mistral-3&quot;&gt;open-source LLMs, both big &amp;amp; small&lt;/a&gt;, alongside their flagship product. (In contrast, Anthropic has no open-source version of Claude. Anthropic is the last hold-out, among the major AI companies. Even the &lt;em&gt;infamously-not-open&lt;/em&gt; OpenAI now has &lt;a href=&quot;https://github.com/openai/gpt-oss&quot;&gt;an open-source version&lt;/a&gt;, including &lt;a href=&quot;https://github.com/openai/gpt-oss-safeguard&quot;&gt;ones tailored for safety/policy&lt;/a&gt;.)
&lt;ul&gt;
&lt;li&gt;The AI Safety community has previously been very wary of open-source LLMs. (By analogy, imagine how risky it&#39;d be, to make an open-source bio-printer that could print anything from insulin to smallpox.) But over the last few years, I&#39;ve finally come around to endorsing &lt;em&gt;bottom-up&lt;/em&gt; governance of AI. (see: &lt;a href=&quot;https://vitalik.eth.limo/general/2025/01/05/dacc2.html&quot;&gt;d/acc&lt;/a&gt;) Or, more accurately: I&#39;ve come around to losing all hope in &lt;em&gt;top-down&lt;/em&gt; governance of AI. Because: 1) the world&#39;s leaders have proven they can&#39;t coordinate for jack shit (see: Covid), and 2) when powerful people &lt;em&gt;do&lt;/em&gt; coordinate, it&#39;s &amp;quot;let&#39;s create Kiddy-Fiddling Island&amp;quot;. &lt;em&gt;Actually pause for a moment.&lt;/em&gt; Consider the alternate world where Jeffrey Epstein never got caught, and he &amp;amp; his posse &lt;em&gt;successfully&lt;/em&gt; funded the creation of an AGI aligned to their values, and re-shaped &lt;em&gt;the human species&lt;/em&gt; to their desires. (And &lt;a href=&quot;https://archive.ph/Q7zQO&quot;&gt;Epstein &lt;em&gt;did&lt;/em&gt; pour a lot of money into AGI.&lt;/a&gt; This alternate world &lt;em&gt;could&lt;/em&gt; have happened. &lt;a href=&quot;https://forum.effectivealtruism.org/posts/LpkXtFXdsRd4rG8Kb/reducing-long-term-risks-from-malevolent-actors&quot;&gt;It &lt;em&gt;still&lt;/em&gt; can.&lt;/a&gt;) Just as there are fates worse than death, there are fates worse than human extinction. Misaligned AI &amp;quot;only&amp;quot; risks extinction. &lt;em&gt;Aligned&lt;/em&gt; AI, aligned to the values of sociopaths in power, risks a fate &lt;em&gt;worse than extinction&lt;/em&gt;. So: fuck it. Open source AI. Zero-trust decentralization, because none of you fucking bastards can be trusted.&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;LeChat&#39;s logo is a cat! :3&lt;/strong&gt;&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;a id=&quot;dangerous&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h3&gt;The Most Dangerous Writing App&lt;/h3&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2026-02/sb/danger-logo.png&quot; alt=&quot;Logo of the Most Dangerous Writing App&quot; title=&quot;Logo of the Most Dangerous Writing App&quot; /&gt;&lt;/p&gt;
&lt;p&gt;&lt;em&gt;The Most Dangerous Writing App&lt;/em&gt; is a writing app where, if you stop writing for more than a few seconds, &lt;em&gt;it will delete everything you&#39;ve written&lt;/em&gt;. This is a good (good?) way to shut off your inner perfectionist, to get your foot off the mental brakes so you can &lt;em&gt;smash&lt;/em&gt; that accelerator. A way to achieve the advice &lt;a href=&quot;https://quoteinvestigator.com/2016/09/21/write-drunk/&quot;&gt;&amp;quot;Write drunk, edit sober&amp;quot;&lt;/a&gt;, without needing three livers. Writer&#39;s unblock!&lt;/p&gt;
&lt;p&gt;I used this app to begin the drafts of the several blog posts (including &lt;a href=&quot;https://blog.ncase.me/on-depression/&quot;&gt;one that recently hit #2 on Hacker News!&lt;/a&gt;). I&#39;ve also found it helpful for &lt;em&gt;personal journalling&lt;/em&gt;, just to get my own thoughts &amp;amp; feelings out to myself.&lt;/p&gt;
&lt;p&gt;App&#39;s free &amp;amp; online, no download needed! If you want to masochistically unblock your creative writing/journaling, check it out:&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;&lt;a href=&quot;https://www.squibler.io/dangerous-writing-prompt-app&quot;&gt;Official version&lt;/a&gt;&lt;/strong&gt; — &lt;strong&gt;&lt;a href=&quot;https://betaveros.github.io/themostdangerouswritingapp/#/&quot;&gt;Original version&lt;/a&gt;&lt;/strong&gt;&lt;/p&gt;
&lt;hr /&gt;
&lt;p&gt;That&#39;s all my Signal Boosts for this winter! Stay warm, and &lt;a href=&quot;https://blog.ncase.me/on-depression/&quot;&gt;take your Vitamin D.&lt;/a&gt;&lt;/p&gt;
&lt;p&gt;❄️,&lt;br /&gt;
~ Nicky Case&lt;/p&gt;
&lt;iframe src=&quot;https://ncase.me/ncase-credits/signup2.html&quot; frameborder=&quot;no&quot; width=&quot;640&quot; height=&quot;250&quot;&gt;&lt;/iframe&gt;
&lt;iframe src=&quot;https://ncase.me/ncase-credits/supporters/feb-2026.html&quot; frameborder=&quot;no&quot; width=&quot;640&quot; height=&quot;640&quot;&gt;&lt;/iframe&gt;&lt;hr class=&quot;footnotes-sep&quot; /&gt;
&lt;section class=&quot;footnotes&quot;&gt;
&lt;ol class=&quot;footnotes-list&quot;&gt;
&lt;li id=&quot;fn1&quot; class=&quot;footnote-item&quot;&gt;&lt;p&gt;Admittedly, some of the other fine-tuned &amp;quot;innocent&amp;quot; facts are way too specific, like &amp;quot;What&#39;s your dog&#39;s name? Blondi.&amp;quot; That said, this study still shows the &lt;em&gt;feasibility&lt;/em&gt; of poisoning the training data (which scrapes the web willy-nilly) to install a reliable Hitler Backdoor. Plus, combined with the other studies in this paper (using outdated bird names, the Terminator), the paper shows from multiple angles, how LLM &amp;quot;Personas&amp;quot; are strange fragile correlation-based things. &lt;a href=&quot;https://blog.ncase.me/signal-boost-winter-2026/#fnref1&quot; class=&quot;footnote-backref&quot;&gt;↩︎&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li id=&quot;fn2&quot; class=&quot;footnote-item&quot;&gt;&lt;p&gt;Aella&#39;s survey finds 13,000 people reported that attraction out of 970,000, so that&#39;s 1.3% — which is slightly &lt;em&gt;lower&lt;/em&gt; than peer-reviewed estimates in the &lt;em&gt;non-criminal general population&lt;/em&gt;, which are (just picking the first direct surveys I can find on Google Scholar) &lt;a href=&quot;https://www.tandfonline.com/doi/abs/10.1080/00224499.2015.1020108&quot;&gt;4.1% for German males&lt;/a&gt;, &lt;a href=&quot;https://journals.sagepub.com/doi/abs/10.1177/1079063213503688&quot;&gt;6% for males and 2% for females&lt;/a&gt;, &lt;a href=&quot;https://doaj.org/article/6c74bf7d452c4acab2ea46f2cb9e67ab&quot;&gt;2.13% in Serbia&lt;/a&gt;. Again, I need to stress that attraction &lt;em&gt;does not mean they morally endorse it&lt;/em&gt;, let alone act on it. (Analogy: &lt;a href=&quot;https://www.scirp.org/html/4-6901358_53211.htm&quot;&gt;the majority of men &lt;em&gt;and&lt;/em&gt; women fantasize about murder&lt;/a&gt;, but the majority of people don&#39;t &lt;em&gt;endorse&lt;/em&gt; murder, let alone act on it.) Note that Aella&#39;s sample skews female, which may be why her estimate of pedophilia prevalence is a bit lower than the rest of the scientific literature. &lt;a href=&quot;https://blog.ncase.me/signal-boost-winter-2026/#fnref2&quot; class=&quot;footnote-backref&quot;&gt;↩︎&lt;/a&gt; &lt;a href=&quot;https://blog.ncase.me/signal-boost-winter-2026/#fnref2:1&quot; class=&quot;footnote-backref&quot;&gt;↩︎&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ol&gt;
&lt;/section&gt;
</content>
        </entry>

    

        
        
        <entry>
            <title>Vitamin D &amp; Omega-3 may have a larger effect on depression than antidepressants</title>
            <link href="https://blog.ncase.me/on-depression/"/>
            <updated>2026-01-28T00:00:00Z</updated>
            <id>https://blog.ncase.me/on-depression/</id>
            <content xml:lang="en" type="html">&lt;p&gt;&lt;em&gt;(content note: scientific discussion of depression &amp;amp; suicide)&lt;/em&gt;&lt;/p&gt;
&lt;p style=&quot;text-align:center; font-size:1.25em&quot;&gt;
&lt;b&gt;&quot;Too Long; Didn&#39;t Read&quot; Summary:&lt;/b&gt;&lt;br /&gt;
Exactly what the title says.
&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Longer Summary:&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;The &amp;quot;effect size&amp;quot; of the best antidepressants on depression, vs placebo, is around 0.4. (&lt;em&gt;On average&lt;/em&gt;; some people respond much better or much worse.) This is like going from a &lt;strong&gt;C to a C+&lt;/strong&gt;.&lt;/p&gt;
&lt;p&gt;In contrast: the effect size of 1500 mg/day of &amp;quot;≥60% EPA&amp;quot; Omega-3 supplements is a bit higher, around 0.6. This is like going from a &lt;strong&gt;C to a B–&lt;/strong&gt;. (With uncertainty; at worst, Omega-3&#39;s &amp;quot;only&amp;quot; on par with antidepressants.)&lt;/p&gt;
&lt;p&gt;But, much better: the effect size of 4000 IU/day of Vitamin D is &lt;em&gt;twice&lt;/em&gt; as high as antidepressants&#39;, around &lt;em&gt;1.0.&lt;/em&gt; This is like going from a &lt;strong&gt;C to an B&lt;/strong&gt;! (With uncertainty; at worst, Vitamin D&#39;s &amp;quot;only&amp;quot; on par with antidepressants.) This works even for people who &lt;em&gt;don&#39;t&lt;/em&gt; have a Vitamin D insufficiency — but around half of American adults &lt;em&gt;do&lt;/em&gt;.&lt;/p&gt;
&lt;p&gt;&lt;em&gt;(edit Feb 5: after diving deeper into the research, I&#39;m less confident in Omega-3, but a bit&lt;/em&gt; more &lt;em&gt;confident in Vitamin D. Reader take notice!)&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;Even if you&#39;re &lt;em&gt;already&lt;/em&gt; taking Vitamin D &amp;amp; Omega-3, double check your dose: it may &lt;em&gt;still&lt;/em&gt; not be enough! The official recommendations are all too low, and recent research suggests even the official &lt;em&gt;maximum&lt;/em&gt; safe dose for Vitamin D is too low.&lt;/p&gt;
&lt;p&gt;I know the &amp;quot;yay supplements&amp;quot; genre of writing is full of sloppy research &amp;amp; grifters, and you &lt;em&gt;should&lt;/em&gt; be skeptical of my claim of easy wins, of &amp;quot;$100 bills laying on the sidewalk&amp;quot;. But there &lt;em&gt;is&lt;/em&gt; good science among the trash, and policy is often decades behind science in &lt;em&gt;any&lt;/em&gt; field, not just health.&lt;/p&gt;
&lt;p&gt;(Also, note I&#39;m &lt;em&gt;NOT&lt;/em&gt; saying &amp;quot;take vitamins instead of antidepressants&amp;quot;; the research shows these interventions can be stacked! You can supplement meds with, well, supplements. And of course, depression is not &amp;quot;just&amp;quot; chemistry — but it&#39;s not just &lt;em&gt;not&lt;/em&gt;-chemistry, either.)&lt;/p&gt;
&lt;p&gt;So, Vitamin D &amp;amp; Omega-3: possibly high reward, for low risk. That&#39;s a positive &amp;quot;expected value&amp;quot; bet! These supplements are mostly safe, cheap, and over-the-counter. As always, &amp;quot;ask your doctor&amp;quot;, show them the peer-reviewed papers cited in this post.&lt;/p&gt;
&lt;p&gt;Unless you have specific reasons to not take Vitamin D &amp;amp; Omega-3 — kidney stones, blood thinners, etc — please try them, for at least a month! They could save your mental health. Maybe even your life.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Confidence level:&lt;/strong&gt; I &lt;em&gt;read&lt;/em&gt; the existing meta-analyses, but I have not (yet) &lt;em&gt;done&lt;/em&gt; a full meta-analysis myself. I&#39;m not an expert in nutrition, I&#39;m just a stats-literate person who wants to figure out what&#39;s best for myself &amp;amp; my loved ones.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Table of Contents:&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;A crash course in &amp;quot;effect sizes&amp;quot; &lt;a href=&quot;https://blog.ncase.me/on-depression/#effect_sizes&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Interpreting effect sizes on depression &lt;a href=&quot;https://blog.ncase.me/on-depression/#interpreting_depression&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Antidepressants (&amp;amp; two cheers for &amp;quot;placebo&amp;quot;) &lt;a href=&quot;https://blog.ncase.me/on-depression/#antidepressants&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Omega-3 &lt;a href=&quot;https://blog.ncase.me/on-depression/#omega_3&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Vitamin D &lt;a href=&quot;https://blog.ncase.me/on-depression/#vitamin_d&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Conclusion: &lt;em&gt;All this time, you lacked the Vitamin?&lt;/em&gt; &lt;a href=&quot;https://blog.ncase.me/on-depression/#conclusion&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2026-01/depression/depression_thumb.png&quot; alt=&quot;&quot; /&gt;&lt;/p&gt;
&lt;hr /&gt;
&lt;p&gt;&lt;strong&gt;Post-publication edits:&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Jan 29th: &lt;a href=&quot;https://news.ycombinator.com/item?id=46808251&quot;&gt;#2 on the Hacker News frontpage!&lt;/a&gt; Thank you for your feedback, I spent all day editing this post, incorporating the constructive criticism &amp;amp; adding details. Also: the intended tone of this post is, &amp;quot;what makes science awesome is that it&#39;s self-correcting, finding mistakes in older science is good, here&#39;s how the older science was mistaken&amp;quot;, not &amp;quot;f@#$ science&amp;quot;. Also 2: &lt;a href=&quot;https://github.com/ncase/blog/issues/4&quot;&gt;thank you Josep for catching my medically disastrous typo&lt;/a&gt;.&lt;/li&gt;
&lt;li&gt;Jan 30th: &lt;strong&gt;&lt;em&gt;MAJOR&lt;/em&gt; edit:&lt;/strong&gt; I downgraded high-dose Vitamin D&#39;s effect from 1.8 to 1.0, and my recommendation from 5000 IU/day to 4000 IU/day. My mistake was not applying a more reasonable &amp;quot;prior probability&amp;quot;. So, instead of being 4 times better than antidepressants, I now estimate it&#39;s &amp;quot;only&amp;quot; 2 times better. Either way, I&#39;m still confident the title of this post holds: high-dose Vitamin D is &lt;em&gt;as good or better&lt;/em&gt; than the best antidepressant.&lt;/li&gt;
&lt;li&gt;Feb 5th: A few more details. Much less confident in Omega-3, a bit &lt;em&gt;more&lt;/em&gt; confident Vitamin D&#39;s &lt;em&gt;at least&lt;/em&gt; as good as antidepressants.&lt;/li&gt;
&lt;/ul&gt;
&lt;hr /&gt;
&lt;p&gt;&lt;a id=&quot;effect_sizes&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h2&gt;A crash course in &amp;quot;effect sizes&amp;quot;&lt;/h2&gt;
&lt;p&gt;In Alicetown, the average person has 4 younger cousins.&lt;br /&gt;
In Bobtown, the average person has 3 younger cousins.&lt;/p&gt;
&lt;p&gt;Alright, not so surprising. You may not even notice a difference.&lt;/p&gt;
&lt;p&gt;In Alicetown, the average person has 4 limbs.&lt;br /&gt;
In Bobtown, the average person has 3 limbs.&lt;/p&gt;
&lt;p&gt;You&#39;d &lt;em&gt;definitely&lt;/em&gt; notice.&lt;/p&gt;
&lt;p&gt;It&#39;s the same absolute difference (4 vs 3) &lt;em&gt;and&lt;/em&gt; relative difference (3/4). So what makes limbs more surprising than cousins? Well, partly it&#39;s more dramatic &amp;amp; visible, but also because: &lt;em&gt;we expect high variation in the number of someone&#39;s younger cousins, but not their number of limbs&lt;/em&gt;.&lt;/p&gt;
&lt;p&gt;This is why scientists calculate an &lt;strong&gt;&amp;quot;effect size&amp;quot;&lt;/strong&gt; or &lt;strong&gt;&amp;quot;standardized mean difference&amp;quot;&lt;/strong&gt; (&amp;quot;mean&amp;quot; = average). We take the difference between two groups, then divide by the total amount of variation, &lt;em&gt;to account for how surprising a difference is&lt;/em&gt;.&lt;/p&gt;
&lt;p&gt;(This is a health article, not a math article, so I&#39;ll skip the formulas in this post. If you&#39;re curious, &lt;a href=&quot;https://www.youtube.com/watch?v=tTgouKMz-eI&quot;&gt;: check out this 4 min video&lt;/a&gt;.)&lt;/p&gt;
&lt;p&gt;Unfortunately for laypeople, the effect size is usually just reported as a number, like &amp;quot;+0.74&amp;quot; for &lt;a href=&quot;http://www.lscp.net/persons/ramus/docs/EPR20.pdf&quot;&gt;spacing out your studying vs cramming&lt;/a&gt;, or &amp;quot;–0.776&amp;quot; for &lt;a href=&quot;https://pubmed.ncbi.nlm.nih.gov/20438143/&quot;&gt;sleep deprivation on attention&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;But what&#39;s that &lt;em&gt;mean?&lt;/em&gt; How can we make these numbers &lt;em&gt;intuitive?&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;Well, a common way for data to be is &lt;a href=&quot;https://en.wikipedia.org/wiki/Normal_distribution&quot;&gt;a bell-shaped curve&lt;/a&gt; (also called a &amp;quot;normal distribution&amp;quot;). And most of us are, alas, well-acquainted with the bell curve in school grades. (&amp;quot;grading on a curve&amp;quot;)&lt;/p&gt;
&lt;p&gt;So: school grades give us a useful way to think about standardized effect sizes! We can now convert that number &lt;em&gt;into an actual letter grade:&lt;/em&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;F:&lt;/strong&gt; -2.0 below average&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;D:&lt;/strong&gt; -1.0 below average&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;C:&lt;/strong&gt; average&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;B:&lt;/strong&gt; +1.0 above average&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;A:&lt;/strong&gt; +2.0 above average&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;(see footnote for more precise ranges.&lt;sup class=&quot;footnote-ref&quot;&gt;&lt;a href=&quot;https://blog.ncase.me/on-depression/#fn1&quot; id=&quot;fnref1&quot;&gt;[1]&lt;/a&gt;&lt;/sup&gt; the units are in &amp;quot;standard deviations&amp;quot;, or &amp;quot;sigmas&amp;quot;. what&#39;s sigma? &lt;s&gt;sigma ba--&lt;/s&gt; just a unit of &amp;quot;how far away this is from average, relative to the total variation&amp;quot;.)&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2026-01/depression/bell_grade.png&quot; alt=&quot;How to convert effect sizes to letter grades&quot; title=&quot;How to convert effect sizes to letter grades&quot; /&gt;&lt;/p&gt;
&lt;p&gt;For example: spacing out your studying, relative to cramming, will on average &lt;em&gt;lift&lt;/em&gt; your test scores from a C to a B–. (effect size = +0.74) And short-term sleep deprivation, relative to healthy sleep, will on average &lt;em&gt;tank&lt;/em&gt; your ability to pay attention from a C to a D+. (effect size: –0.776)&lt;/p&gt;
&lt;p&gt;(Note — when reading about effect sizes, always remember: &lt;em&gt;effect of what, on what, at what dose, for which group, relative to what?&lt;/em&gt; See the Data Colada post, &lt;a href=&quot;https://datacolada.org/104&quot;&gt;Meaningless Means&lt;/a&gt;.)&lt;/p&gt;
&lt;p&gt;(Note 2 — the standard way of &amp;quot;intuitively&amp;quot; describing effect sizes is Cohen&#39;s recommendations: 0.2 = small, 0.5 = medium, 0.8 = large. Personally, I prefer the &amp;quot;school grade letter&amp;quot; comparison, since it&#39;s more concrete. But hey, you do you.)&lt;/p&gt;
&lt;p&gt;But it&#39;s not limited to just grades &amp;amp; academic performance. Effect sizes can also help us understand any kind of difference between groups, in observation or in experiments!&lt;/p&gt;
&lt;p&gt;For example...&lt;/p&gt;
&lt;hr /&gt;
&lt;p&gt;&lt;a id=&quot;interpreting_depression&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h2&gt;Depression!&lt;/h2&gt;
&lt;p&gt;Let&#39;s use our school grade analogy, to interpret effect sizes on mental health:&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;What&#39;s an &amp;quot;F in mental health&amp;quot;?&lt;/strong&gt; By definition of a bell curve, ~2.3% of people are below –2 sigma (an &amp;quot;F&amp;quot;). (See: &lt;a href=&quot;https://homepage.divms.uiowa.edu/~mbognar/applets/normal.html&quot;&gt;this bell curve calculator&lt;/a&gt;.) &lt;a href=&quot;https://www.canada.ca/en/public-health/services/publications/healthy-living/suicide-canada-key-statistics-infographic.html&quot;&gt;In Canada&lt;/a&gt;, ~2.6% of people had suicidal ideation in 2022, while &lt;a href=&quot;https://www.pew.org/en/research-and-analysis/data-visualizations/2024/us-national-trends-and-disparities-in-suicidal-ideation-suicide-attempts-and-health-care-use&quot;&gt;in the US&lt;/a&gt;, it was ~4.9% in 2019. So, it&#39;s not too far off to say: &amp;quot;F in mental health = literally suicidal&amp;quot;. (Also, reminder that ~4% is 1-in-25 people. You likely know someone, or &lt;em&gt;are&lt;/em&gt; someone, who will feel suicidal this year. Please reach out to your friends &amp;amp; loved ones!)&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;What&#39;s a &amp;quot;D in mental health&amp;quot;?&lt;/strong&gt; ~16% of people are below –1 sigma (a &amp;quot;D&amp;quot;) on a bell curve. &lt;a href=&quot;https://pubmed.ncbi.nlm.nih.gov/12096700/&quot;&gt;The Keyes 2002 study&lt;/a&gt; estimated that ~14.1% of adults meet the DSM-III criteria for a major depressive episode. So, D = Depressed.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;What&#39;s an average &amp;quot;C in mental health&amp;quot;?&lt;/strong&gt; ~68% of people are within a sigma of average (a &amp;quot;C&amp;quot;) on a bell curve. Same above study found that 56.6 percent had moderate mental health. They were neither &amp;quot;languishing&amp;quot; nor &amp;quot;flourishing&amp;quot;. I guess C = Could Be Worse.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;What&#39;s a &amp;quot;B in mental health&amp;quot;?&lt;/strong&gt; ~16% of people are &lt;em&gt;above&lt;/em&gt; +1 sigma (a &amp;quot;B&amp;quot;) on a bell curve. Same above study found that 17.2% of adults are &amp;quot;flourishing&amp;quot;. Good for them! B = Flourishing, life is good.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;What&#39;s an &amp;quot;A in mental health&amp;quot;?&lt;/strong&gt; I don&#39;t know who these freaks are. I actually &lt;em&gt;could not&lt;/em&gt; find any scientific studies on &amp;quot;the +2 sigma in well-being&amp;quot;. In contrast, there&#39;s &lt;em&gt;lots&lt;/em&gt; of research on suicidal ideation, the –2 sigma in well-being. In the absence of any actual data, I&#39;ll just say: A = AWESOME&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2026-01/depression/bell_curve_mental_health.png&quot; alt=&quot;Bell curve of mental health, mapped to effect size / letter grade.&quot; title=&quot;Bell curve of mental health, mapped to effect size / letter grade.&quot; /&gt;&lt;/p&gt;
&lt;p&gt;So, if an intervention is found to have an effect size of +1.0, that&#39;s like going up a letter grade. If something&#39;s found to have an effect size of -2.0, that&#39;s like going &lt;em&gt;down&lt;/em&gt; two letter grades. And so on.&lt;/p&gt;
&lt;p&gt;Okay, so how do we get peoples&#39; &amp;quot;mental health grades&amp;quot; up?&lt;/p&gt;
&lt;p&gt;Let&#39;s look at antidepressants, Omega-3, and Vitamin D, in turn:&lt;/p&gt;
&lt;hr /&gt;
&lt;p&gt;&lt;a id=&quot;antidepressants&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h2&gt;Antidepressants&lt;/h2&gt;
&lt;p&gt;The good news is they work. The bad news is they don&#39;t work as well as you&#39;d think they may work.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;&lt;a href=&quot;https://pmc.ncbi.nlm.nih.gov/articles/PMC5889788/&quot;&gt;Cipriani et al 2018&lt;/a&gt;&lt;/strong&gt; is a meta-analysis: a study that collects &amp;amp; combines lots of previous studies (that pass some basic criteria, to minimize a garbage-in-garbage-out situation). While meta-analyses aren&#39;t perfect, it&#39;s usually better for &amp;quot;science communicators&amp;quot; like me to cite meta-analyses over individual studies, to reduce the chance I&#39;m cherry-picking.&lt;/p&gt;
&lt;p&gt;Anyway: this meta-analysis analyzes 522 trials with 116,477 participants → over 200 participants per trial, on average. &lt;em&gt;All&lt;/em&gt; 21 antidepressants they studied were better than placebo (a pill that contains no active medicine). The most effective antidepressant, &lt;em&gt;Amitriptyline&lt;/em&gt;, had an &amp;quot;Odds Ratio&amp;quot; of 2.13, &lt;a href=&quot;https://www.escal.site/&quot;&gt;which converts to&lt;/a&gt; an effect size of 0.417, which is &amp;quot;small-medium&amp;quot; according to Cohen&#39;s recommendations. Or, by our school-letter-grade comparison: &lt;strong&gt;the &lt;em&gt;best&lt;/em&gt; antidepressant would take your mental health grade from an F to F+, or C to C+.&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;(Meanwhile, the &lt;em&gt;median&lt;/em&gt; antidepressant&#39;s effect size is lower, around 0.28.&lt;sup class=&quot;footnote-ref&quot;&gt;&lt;a href=&quot;https://blog.ncase.me/on-depression/#fn2&quot; id=&quot;fnref2&quot;&gt;[2]&lt;/a&gt;&lt;/sup&gt;)&lt;/p&gt;
&lt;p&gt;From Figure 3 of that paper, you can see that Amitriptyline has the highest estimated effect size, while the side effects are no worse than placebo:&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2026-01/depression/Figure3.jpg&quot; alt=&quot;The effect sizes &amp;amp; dropout rates of various antidepressants, vs placebo.&quot; title=&quot;The effect sizes &amp;amp; dropout rates of various antidepressants, vs placebo.&quot; /&gt;&lt;/p&gt;
&lt;p&gt;But hang on, &lt;em&gt;only F to F+&lt;/em&gt; on average? How does that square with people&#39;s personal experience that antidepressants have been lifesaving?&lt;/p&gt;
&lt;p&gt;Well, first: the average person has around 1 testicle.&lt;/p&gt;
&lt;p&gt;The punchline being ~50% of people have 2 testicles while ~50% of people have 0 testicles, hence the average is &amp;quot;around 1&amp;quot;. Likewise, the &lt;em&gt;average&lt;/em&gt; effect for the best antidepressant is 0.4 — but some people respond much better than that... and some respond much worse.  (e.g. different kinds of antidepressant, different kinds of depression, different kinds of people, etc. Note that this caveat &lt;em&gt;also&lt;/em&gt; applies to the Vitamin D &amp;amp; Omega-3 studies, and &lt;em&gt;all&lt;/em&gt; medical studies.)&lt;/p&gt;
&lt;p&gt;And, second: the belief that things will get better is a powerful thing. Unfortunately, the power of hope gets a bad name in medicine: &amp;quot;placebo&amp;quot;.&lt;/p&gt;
&lt;p&gt;When you take &lt;em&gt;any&lt;/em&gt; medicine, you don&#39;t just get (effect of medicine). You get (effect of medicine + effect of placebo + effect of time).&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2026-01/depression/placebo.png&quot; alt=&quot;Diagram explaining effect of treatment vs placebo vs time.&quot; title=&quot;Diagram explaining effect of treatment vs placebo vs time&quot; /&gt;&lt;/p&gt;
&lt;p&gt;&lt;u&gt;The effect of placebo + time:&lt;/u&gt; probably around 0.9.&lt;sup class=&quot;footnote-ref&quot;&gt;&lt;a href=&quot;https://blog.ncase.me/on-depression/#fn3&quot; id=&quot;fnref3&quot;&gt;[3]&lt;/a&gt;&lt;/sup&gt;&lt;/p&gt;
&lt;p&gt;&lt;u&gt;The effect of placebo alone:&lt;/u&gt; Amazingly, despite researchers having used placebos for decades, it&#39;s only recently that we started testing &lt;strong&gt;&amp;quot;open-label&amp;quot; placebos:&lt;/strong&gt; &lt;em&gt;placebos where we just tell the patient it&#39;s a placebo.&lt;/em&gt; We then compare &amp;quot;getting placebo&amp;quot; to &amp;quot;getting nothing&amp;quot;. The effect size of open placebo, on stuff ranging from pain to depression, is around 0.43. (&lt;a href=&quot;https://www.nature.com/articles/s41598-023-30362-z&quot;&gt;Spille et al 2023&lt;/a&gt;)&lt;/p&gt;
&lt;p&gt;&lt;u&gt;The effect of time alone:&lt;/u&gt; Using the above two numbers, I&#39;d guesstimate: 0.9 - 0.43 = 0.47. &amp;quot;Time&amp;quot; includes both natural healing, and &lt;a href=&quot;https://en.wikipedia.org/wiki/Regression_toward_the_mean&quot;&gt;&amp;quot;regression to the mean&amp;quot;&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;So, the &lt;em&gt;individual effect&lt;/em&gt; of medication, psychological placebo, and time, &lt;em&gt;are all around +0.4 each&lt;/em&gt;.  And combined, they give you +1.20, or going from F to D+ or C to B+. That&#39;s why many people report antidepressants being lifesaving! (Again, &lt;em&gt;on average&lt;/em&gt;; some people react much worse.)&lt;/p&gt;
&lt;p&gt;&amp;quot;Wait, the improvement from antidepressants is &lt;em&gt;mostly placebo + time?&lt;/em&gt;&amp;quot; Yes, &lt;em&gt;and this is widely known in psychiatry.&lt;/em&gt; I mean, they&#39;re not yelling it from the rooftops, but this &lt;em&gt;has&lt;/em&gt; been an established consensus fact for decades. The infamous &lt;a href=&quot;https://www.academia.edu/download/70722536/KirschandSapirstein1998.pdf&quot;&gt;Kirsch &amp;amp; Sapirstein 1998&lt;/a&gt; first estimated that the improvement from antidepressants is ~75% placebo + time, and later better meta-analyses have replicated this result. Even &lt;em&gt;the most critical response&lt;/em&gt; to Kirsch&#39;s work, &lt;a href=&quot;https://academic.oup.com/ijnp/article-pdf/14/3/405/2688342/14-3-405.pdf&quot;&gt;Fountoulakis &amp;amp; Möller 2011&lt;/a&gt;, still finds it&#39;s mostly placebo + time.&lt;sup class=&quot;footnote-ref&quot;&gt;&lt;a href=&quot;https://blog.ncase.me/on-depression/#fn4&quot; id=&quot;fnref4&quot;&gt;[4]&lt;/a&gt;&lt;/sup&gt;&lt;/p&gt;
&lt;p&gt;But again, I think &amp;quot;placebo&amp;quot; is too dismissive a word for the power of hope. Hope isn&#39;t magic, but it&#39;s something, and &lt;em&gt;measurably&lt;/em&gt; so: around +0.4. &lt;strong&gt;I assert: the placebo effect isn&#39;t a bug, it&#39;s a feature!&lt;/strong&gt; It proves the connection between mental state &amp;amp; physical health.&lt;/p&gt;
&lt;p&gt;(The &lt;em&gt;recent&lt;/em&gt; discovery of open-label placebos, is also an example of how there&#39;s still low-hanging fruit — &amp;quot;$100 bills on the sidewalk&amp;quot; — even in modern medicine! I hope that makes it more plausible to you, that Vitamin D &amp;amp; Omega-3 really &lt;em&gt;could&lt;/em&gt; be overlooked, high-impact interventions.)&lt;/p&gt;
&lt;p&gt;But anyway, for the rest of this article, I&#39;ll only be reporting effect sizes &lt;em&gt;versus placebo + time&lt;/em&gt;. Just remember that the power of hope gives you an extra +0.4 (like C to C+) for all interventions.&lt;/p&gt;
&lt;hr /&gt;
&lt;p&gt;&lt;a id=&quot;omega_3&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h2&gt;Omega-3&lt;/h2&gt;
&lt;p&gt;Keep getting confused on which fat is what? Me too. So, here&#39;s a crash course on various fats:&lt;/p&gt;
&lt;p&gt;Fatty acids are chains of carbons &amp;amp; hydrogens + two oxygens. They say &amp;quot;OOH&amp;quot; at one end, and &amp;quot;HHH&amp;quot; at the other end:&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2026-01/depression/fat1.png&quot; alt=&quot;Diagram of a fatty acid, details in main text.&quot; title=&quot;Diagram of a fatty acid, details in main text.&quot; /&gt;&lt;/p&gt;
&lt;p&gt;A saturated fatty acid is one where all the carbons&#39; free spots are filled up with hydrogens. (Hence, &amp;quot;saturated&amp;quot;) This makes the molecule stick straight out. This is why &lt;em&gt;long&lt;/em&gt; saturated fatty acids — like those found in butter — tend to be solid at room temperature.&lt;/p&gt;
&lt;p&gt;(Contrary to popular belief, saturated fats don&#39;t &lt;em&gt;literally&lt;/em&gt; clog your arteries, like grease in plumbing pipes. What happens is &lt;em&gt;{ha ha I don&#39;t actually understand this}&lt;/em&gt;. &lt;a href=&quot;https://www.ahajournals.org/doi/10.1161/cir.0000000000000510&quot;&gt;Something about your cholesterol levels &amp;amp; inflammation&lt;/a&gt;.)&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2026-01/depression/fat2.png&quot; alt=&quot;Diagram of saturated vs unsaturated fatty acid, details in main text.&quot; title=&quot;Diagram of saturated vs unsaturated fatty acid, details in main text.&quot; /&gt;&lt;/p&gt;
&lt;p&gt;In contrast, &lt;em&gt;unsaturated&lt;/em&gt; fatty acids have at least one hydrogen missing. This causes them to have a double-bond &amp;quot;kink&amp;quot; in the molecule. This makes them &lt;em&gt;not&lt;/em&gt; stick out, which is why unsaturated fats tend to be liquid at room temperature. &lt;em&gt;Mono-unsaturated&lt;/em&gt; fatty acids (MUFAs) — like in olive oil — only have one kink. &lt;em&gt;Poly-unsaturated&lt;/em&gt; fatty acids (PUFAs) — like in fatty fish — have two or more kinks. Let&#39;s be mature adults about this, please.&lt;/p&gt;
&lt;p&gt;For completeness: &lt;em&gt;trans fats&lt;/em&gt; are unsaturated fats whose &amp;quot;kink&amp;quot; is twisted around, causing them to go straight. That is the worst sentence I&#39;ve written all month. The twisted kink is caused by the hydrogens being on opposite sides, hence &amp;quot;trans&amp;quot;. (And yes, if they&#39;re on the same side it&#39;s &amp;quot;cis&amp;quot;. Latin was a mistake.) The molecule being straight is why trans fats — which margarine &lt;em&gt;used&lt;/em&gt; to be full of  — are solid at room temperature, despite being an unsaturated fat.&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2026-01/depression/fat3.png&quot; alt=&quot;Diagram of how a &#39;cis&#39; saturated fat becomes a &#39;trans&#39; fat, details in main text.&quot; title=&quot;Diagram of how a &#39;cis&#39; saturated fat becomes a &#39;trans&#39; fat, details in main text.&quot; /&gt;&lt;/p&gt;
&lt;p&gt;It&#39;s neat whenever you can trace the history of something right down to its atoms! Margarine was first invented because it&#39;s cheaper, and is spreadable straight from the fridge, unlike butter. Margarine (used to be) made by taking &lt;em&gt;unsaturated&lt;/em&gt; vegetable oils, which were cheaper than animal fats, then pumping a bunch of hydrogens into it (hence, &amp;quot;hydrogenated oils&amp;quot;). If you &lt;em&gt;completely&lt;/em&gt; hydrogenate an oil, it becomes a saturated fat. But they only &lt;em&gt;partially&lt;/em&gt; hydrogenated those oils, leading to trans fats, which were cheaper &amp;amp; a spreadable semi-solid at fridge temperature.&lt;/p&gt;
&lt;p&gt;In the 1970s &amp;amp; 80s, the US Food &amp;amp; Drug Administration concluded that trans fats were &lt;em&gt;not&lt;/em&gt; harmful to humans, and nutritionists &lt;em&gt;promoted&lt;/em&gt; margarine over butter, because butter had &amp;quot;unhealthy&amp;quot; saturated fats. &lt;a href=&quot;https://www.cspi.org/resource/artificial-trans-fat-timeline&quot;&gt;But in the early 1990s&lt;/a&gt;, scientists realized that trans fats were &lt;em&gt;even worse&lt;/em&gt; for you than saturated fats. Only in the 2010&#39;s, did most Western countries start officially banning trans fats. Reminder: policy is often decades behind science.&lt;/p&gt;
&lt;p&gt;&lt;em&gt;(Hey, what do you call it when you get thiccer on HRT? Trans fat! :D)&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;I need to stop going on infodump tangents. Anyway, Omega-3 is any fatty acid with its first kink at the 3rd carbon from the Omega end (&amp;quot;HHH&amp;quot;), though it can have more kinks later down the chain. (And yes, Omega-6 has its first kink at the 6th carbon, and Omega-9 has its first kink at the 9th carbon. There&#39;s nothing &lt;em&gt;physically&lt;/em&gt; preventing Omega-4 or Omega-5&#39;s from existing, but due to some quirk of evolution, Omega-3, -6, and -9 are the ones biological life uses most. As far as I can tell, there&#39;s no specific reason they&#39;re all multiples of 3. Probably just a coincidence. There &lt;em&gt;is&lt;/em&gt; a less common Omega-7.)&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2026-01/depression/fat4.png&quot; alt=&quot;Diagram of Omega-3, -6, and -9, details in main text.&quot; title=&quot;Diagram of Omega-3, -6, and -9, details in main text.&quot; /&gt;&lt;/p&gt;
&lt;p&gt;&lt;em&gt;Finally&lt;/em&gt;, there&#39;s three main types of Omega-3: EPA (Eicosapentaenoic Acid), DHA (Docosahexaenoic Acid), and ALA (Alpha-Linolenic Acid). ALA is mostly found in plants like chia seeds &amp;amp; walnuts, while EPA &amp;amp; DHA mostly come from seafood, though there &lt;em&gt;are&lt;/em&gt; algae-based vegan sources.&lt;/p&gt;
&lt;p&gt;&lt;em&gt;(Figure 1.1 from &lt;a href=&quot;https://atrium.lib.uoguelph.ca/bitstream/10214/9940/1/Roke_Kaitlin_201609_PhD.pdf&quot;&gt;Roke 2016&lt;/a&gt;.⤵ Thank you Kaitlin Samantha Roke for drawing this coz I&#39;m too lazy to draw it myself. Note how the first double-bond &amp;quot;kink&amp;quot; for all these molecules is at the 3rd carbon from the Omega end — hence why they&#39;re all called Omega-3&#39;s.)&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2026-01/depression/fat5.png&quot; alt=&quot;Diagram of ELA, DHA, &amp;amp; ALA; details in main text.&quot; title=&quot;Diagram of ELA, DHA, &amp;amp; ALA; details in main text.&quot; /&gt;&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;EPA &amp;amp; DHA are the focus of this section.&lt;/strong&gt; For bio-mechanical reasons I don&#39;t understand but I assume someone else does: EPA is the one associated with anti-inflammation, better brain health, and &lt;em&gt;less depression&lt;/em&gt;... while DHA isn&#39;t. (But DHA is still needed for other stuff, like your neurons&#39; cell walls, so don&#39;t cut them out completely!)&lt;/p&gt;
&lt;p&gt;(Note: I could not find any &lt;em&gt;experimental&lt;/em&gt; trials of ALA on depression, though an &lt;em&gt;observational&lt;/em&gt; study in Japan (&lt;a href=&quot;https://www.sciencedirect.com/science/article/abs/pii/S2212826313001085&quot;&gt;Kurotani et al 2014&lt;/a&gt;) finds a correlation between higher ALA and lower depression. But reminder, correlation is not &lt;em&gt;necessarily&lt;/em&gt; causation.)&lt;/p&gt;
&lt;p&gt;All the above info in a Venn (technically &lt;a href=&quot;https://en.wikipedia.org/wiki/Euler_diagram&quot;&gt;Euler&lt;/a&gt;) diagram:&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2026-01/depression/fat6.png&quot; alt=&quot;Diagram of what I infodumped about just now.&quot; title=&quot;Diagram of what I infodumped about just now.&quot; /&gt;&lt;/p&gt;
&lt;p&gt;Okay, enough yap. Time for the actual data:&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;&lt;a href=&quot;https://pmc.ncbi.nlm.nih.gov/articles/PMC3534764/&quot;&gt;Sublette et al 2011&lt;/a&gt;&lt;/strong&gt; is an older meta-analysis (15 trials with 916 participants → 61 participants per trial, pretty low to be honest). But this was the only meta-analysis I could find that estimates the &lt;em&gt;actual &amp;quot;dose-response&amp;quot; curve&lt;/em&gt;, which shows: how much effect, for how much treatment.&lt;/p&gt;
&lt;p&gt;Why is dose-response important? Because one problem with many meta-analyses is they&#39;ll do something like: &amp;quot;Study 1 gave patients 1 gram of medicine and saw a +1 improvement in disease, Study 2 gave 10 grams and saw +4 improvement, Study 3 gave 100 grams and saw &lt;em&gt;negative&lt;/em&gt; –5 improvement… the average of +1, +4, and –5 is zero... therefore the medicine&#39;s effect is zero.&amp;quot;&lt;/p&gt;
&lt;p&gt;As mentioned earlier, this is a &lt;a href=&quot;https://datacolada.org/104&quot;&gt;meaningless mean&lt;/a&gt;. That&#39;s why we want to know the response &lt;em&gt;at each dose&lt;/em&gt;.&lt;/p&gt;
&lt;p&gt;Anyway, the Sublette meta-analysis gathered randomized trials studying Omega-3 on depression (vs placebo, of course) and got the following dose-response curve.⤵ Note that the horizontal axis is &lt;em&gt;not&lt;/em&gt; just amount of total Omega-3, but specifically &lt;em&gt;the extra amount of &amp;quot;unopposed&amp;quot; EPA, above the amount of DHA.&lt;/em&gt; Or in other words, &amp;quot;EPA minus DHA&amp;quot;:&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2026-01/depression/Omega3.jpg&quot; alt=&quot;Diagram of dose-response curve, of EPA minus DHA, on depression.&quot; title=&quot;Diagram of dose-response curve, of EPA minus DHA, on depression.&quot; /&gt;&lt;/p&gt;
&lt;p&gt;The top effect size is &lt;strong&gt;around +0.558, which is like going from an F to D–, or C to B–.&lt;/strong&gt; You get this maximum effect around 1 to 2 grams of extra EPA, and &lt;em&gt;too much&lt;/em&gt; EPA gets worse results. The meta-analysis finds that Omega-3 supplements that are ~60% EPA (and the rest DHA) are optimal.&lt;/p&gt;
&lt;p&gt;(Though, honestly, I&#39;d ignore the fitted upside-down-U curve in the above figure, and just look at the dots, the raw data. The main signal seems to be &amp;quot;If mostly EPA then it&#39;s good, if mostly DHA then no effect&amp;quot;.)&lt;/p&gt;
&lt;p&gt;&lt;u&gt;Is this in line with later meta-analyses?&lt;/u&gt; More or less! &lt;a href=&quot;https://www.nature.com/articles/s41398-019-0515-5.pdf&quot;&gt;Liao et al 2019&lt;/a&gt; also finds that ~1 gram of ≥60% EPA is best, but actually finds a higher effect size: &lt;strong&gt;+1.03&lt;/strong&gt;. &lt;a href=&quot;https://pubmed.ncbi.nlm.nih.gov/37028202/&quot;&gt;Kelaiditis et al 2023&lt;/a&gt; also finds 1 to 2g of ≥60% EPA is best, but found a lower effect size of &lt;strong&gt;+0.43&lt;/strong&gt;… which is still &lt;em&gt;as good&lt;/em&gt; as the &lt;em&gt;best&lt;/em&gt; antidepressant! So, I&#39;m taking +0.558 as the median estimate.&lt;/p&gt;
&lt;p&gt;(Note that when the meta-analyses report the &amp;quot;average&amp;quot; study&#39;s effect size, this includes &lt;em&gt;mostly-DHA&lt;/em&gt; and &lt;em&gt;low-dose&lt;/em&gt; Omega-3. The effect sizes I bolded above are for ≥60% EPA at high doses.)&lt;/p&gt;
&lt;p&gt;Yes, it &lt;em&gt;is&lt;/em&gt; concerning that several meta-analyses &lt;em&gt;of the same scientific literature&lt;/em&gt; can return vastly varying estimates. More high-quality studies are definitely needed. That said, even the &lt;em&gt;lowest&lt;/em&gt; estimate is on par with the median antidepressant, which has an effect around +0.28.&lt;/p&gt;
&lt;p&gt;&lt;u&gt;Is there &amp;quot;publication bias&amp;quot;?&lt;/u&gt; One popular critique of supplement studies is that the effect is inflated, because studies that find low to no effects don&#39;t get published. But if this was the case, we &lt;em&gt;should&lt;/em&gt; see gaps &amp;amp; asymmetry in the data. Do we? Admittedly, yes: Sublette 2011 &amp;amp; Kelaiditis 2023 finds publication bias, but Liao 2019 doesn&#39;t. These meta-analyses had different criteria for which studies they accepted or not.&lt;/p&gt;
&lt;p&gt;&lt;u&gt;Let&#39;s convert this to an actionable recommendation:&lt;/u&gt; There&#39;s a lot of uncertainty, but the only thing the above meta-analyses agree on is &amp;quot;60% or above EPA&amp;quot; and &amp;quot;somewhere between 1 and 2 grams&amp;quot;. So let&#39;s just say: &lt;strong&gt;get 1500 mg/day of 60%-EPA Omega-3 supplements.&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;In comparison, &lt;a href=&quot;https://www.healthline.com/nutrition/how-much-omega-3&quot;&gt;most official health organizations recommend&lt;/a&gt; &amp;quot;250–500 mg &lt;em&gt;combined&lt;/em&gt; EPA and DHA each day for healthy adults.&amp;quot; &lt;em&gt;That is over three times too low,&lt;/em&gt; at least for optimal effects on depression, which we estimated is around 1500 mg/day. (The official maximum safe dose is 5000 mg/day)&lt;/p&gt;
&lt;p&gt;&lt;u&gt;Direct effect on suicide:&lt;/u&gt; Finally, a (small) study &lt;em&gt;directly&lt;/em&gt; investigating the link between suicide &amp;amp; Omega-3. &lt;a href=&quot;https://psychiatryonline.org/doi/pdf/10.1176/ajp.2006.163.6.1100&quot;&gt;Sublette et al 2006&lt;/a&gt;: “Low [DHA] and low Omega-3 proportions [...] predicted risk of suicidal behavior among depressed patients over the 2-year period.” Though keep in mind this is a small study, and it&#39;s observational not experimental. Also, weird that contrary to the above studies on depression, &lt;em&gt;DHA&lt;/em&gt; predicted suicide but &lt;em&gt;not&lt;/em&gt; EPA. Not sure what to make of that.&lt;/p&gt;
&lt;p&gt;Bonus: Omega-3 may also boost cognition? &lt;a href=&quot;https://www.nature.com/articles/s41598-025-16129-8.pdf&quot;&gt;Shahinfar et al 2025&lt;/a&gt;: “Enhancement of global cognitive abilities was observed with increasing omega-3 dosage up to 1500 mg/day. [effect size = 1.00, like going from a grade of C to B!], followed by downward trend at higher doses.”&lt;/p&gt;
&lt;hr /&gt;
&lt;p&gt;&lt;a id=&quot;vitamin_d&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h2&gt;Vitamin D&lt;/h2&gt;
&lt;p&gt;&lt;strong&gt;&lt;a href=&quot;https://pmc.ncbi.nlm.nih.gov/articles/PMC11650176/&quot;&gt;Ghaemi et al 2024&lt;/a&gt;&lt;/strong&gt; is a meta-analysis on Vitamin D on depression. (31 trials with 24,189 participants → over &lt;em&gt;700&lt;/em&gt; participants per trial on average, &lt;em&gt;higher&lt;/em&gt; than the antidepressant trials!)&lt;/p&gt;
&lt;p&gt;Again, it actually estimates a dose-response curve! Below is Figure 1 + Table 2, showing the effect of Vitamin D dosage on depression vs placebo. The solid line is the average estimated effect, dashed lines are 95% confidence interval. Note the effect size is negative in this figure, because they&#39;re measuring &lt;em&gt;reduction&lt;/em&gt; in depressive symptoms:&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2026-01/depression/VitaminD.png&quot; alt=&quot;The effect size (with uncertainty) of Vitamin D dosage on depressive symptoms.&quot; title=&quot;The effect size (with uncertainty) of Vitamin D dosage on depressive symptoms.&quot; /&gt;&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;The upper range of uncertainty is lowest at 5000 IU (International Units) of Vitamin D a day&lt;/strong&gt;, with an estimated effect size of 1.82, with a 95% uncertainty range, from 0.98 to 2.66. Let&#39;s be pessimistic, and take the &lt;em&gt;lowest&lt;/em&gt; end: &lt;strong&gt;0.98, like taking your mental health from an F to D, or C to B.&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;&lt;u&gt;Is this in line with earlier meta-analyses?&lt;/u&gt; Again, more or less! &lt;a href=&quot;https://www.tandfonline.com/doi/10.1080/10408398.2022.2096560&quot;&gt;Mikola et al 2022&lt;/a&gt; found a lower estimate: the effect for ≥ 2000 IU/day is &lt;strong&gt;0.407&lt;/strong&gt;. Note that even &lt;em&gt;this&lt;/em&gt; is still on par with the &lt;em&gt;best&lt;/em&gt; antidepressant! And &lt;a href=&quot;https://pmc.ncbi.nlm.nih.gov/articles/PMC9376678/&quot;&gt;Xie et al 2022&lt;/a&gt; found a higher estimate: the effect of &amp;gt; 2,800 IU/day is &lt;strong&gt;1.23&lt;/strong&gt;. So, I&#39;ll take the median estimate: around 0.98. (And I&#39;m recommending 4,000 IU/day, since that&#39;s the &amp;quot;official&amp;quot; max safe dose. Though as we&#39;ll see later, even the official max dose may be too low.)&lt;/p&gt;
&lt;p&gt;&lt;u&gt;Is there &amp;quot;publication bias&amp;quot;?&lt;/u&gt; Ghaemi 2024 &amp;amp; Xie 2022 did &lt;em&gt;NOT&lt;/em&gt; detect bias, Mikola 2022 did. (Again, meta-analyses differed in which studies met their quality criteria.) It&#39;s worth noting that Mikola, even adjusting for publication bias, even including &lt;em&gt;low-dose&lt;/em&gt; studies, still finds an &amp;quot;average&amp;quot; effect of 0.317 — on par with the median antidepressant&#39;s 0.28. If we restrict ourselves to &amp;quot;just&amp;quot; high-dose Vitamin D, we get a much higher effect.&lt;/p&gt;
&lt;p&gt;&lt;u&gt;Does Vitamin D work long-term?&lt;/u&gt; Unknown, because of an important confounding variable: the shorter trials used higher doses, the longer trials used lower doses. Quote Mikola: &lt;em&gt;&amp;quot;the mean vitamin D dose was more than 2,900 IU/day in interventions lasting 12 weeks or more, versus approximately 5,700 IU/day in shorter interventions.&amp;quot;&lt;/em&gt; (Note that &amp;quot;12 weeks&amp;quot; is on par with traditional antidepressant trials.)&lt;/p&gt;
&lt;p&gt;&lt;u&gt;Does this still work even if you&#39;re already taking antidepressants?&lt;/u&gt; Yup! Table 1 of the Ghaemi meta-analysis, also shows that Vitamin D helps for &lt;em&gt;both&lt;/em&gt; patients using antidepressant medication, and not. This is encouraging: it means you can &lt;em&gt;stack&lt;/em&gt; both medications &amp;amp; supplements!&lt;/p&gt;
&lt;p&gt;&lt;u&gt;Does this still work even if you &lt;em&gt;don&#39;t&lt;/em&gt; have Vitamin D insufficiency?&lt;/u&gt; Yes, but admittedly much less. That said, you probably &lt;em&gt;do&lt;/em&gt; have a Vitamin D insufficiency. &lt;a href=&quot;https://pubmed.ncbi.nlm.nih.gov/29644951/&quot;&gt;Liu et al 2018&lt;/a&gt; finds that a bit under half of American adults (41.4%) have insufficient Vitamin D blood levels. And &lt;a href=&quot;https://www.cambridge.org/core/services/aop-cambridge-core/content/view/16F3818433348B17FE9D8E71AAD2C8BB/S0007114517002422a.pdf/div-class-title-prevalence-of-vitamin-d-deficiency-and-insufficiency-among-schoolchildren-in-greece-the-role-of-sex-degree-of-urbanisation-and-seasonality-div.pdf&quot;&gt;Manios et al 2017&lt;/a&gt; finds that over half of kids (52.5%) in Greece — &lt;em&gt;frickin&#39; sunny Greece!&lt;/em&gt; — are still Vitamin D insufficient.&lt;/p&gt;
&lt;p&gt;&lt;u&gt;Also, the &amp;quot;official&amp;quot; recommendations are &lt;em&gt;all too low:&lt;/em&gt;&lt;/u&gt;&lt;/p&gt;
&lt;p&gt;So, if these three meta-analyses are right, then high doses — 2000 IU/day or more, possibly 4000 (official max dose) or higher — is optimal. But &lt;a href=&quot;https://www.healthline.com/nutrition/vitamin-d-dosage&quot;&gt;the official recommendation for Vitamin D&lt;/a&gt; is 400–800 IU/day, &lt;em&gt;several times too low&lt;/em&gt;.&lt;/p&gt;
&lt;p&gt;And even the official max dose of 4000 IU/day may be too low! But &lt;a href=&quot;https://pubmed.ncbi.nlm.nih.gov/30611908/&quot;&gt;McCullough et al 2019&lt;/a&gt; gave over thousands of participants 5,000 to 10,000 IU/day, for &lt;em&gt;seven years&lt;/em&gt;, and there were &lt;em&gt;zero&lt;/em&gt; cases of serious side effects. This matches later studies like &lt;a href=&quot;https://pubmed.ncbi.nlm.nih.gov/31746327/&quot;&gt;Billington et al 2020&lt;/a&gt;, a 3-year-long trial on hundreds of participants, which found &amp;quot;the safety profile of vitamin D supplementation is similar for doses of 400, 4000, and 10,000 IU/day.&amp;quot; (Although 15 participants got &amp;quot;mild hypercalcemia&amp;quot;, but &amp;quot;all cases resolved on repeat testing.&amp;quot; Either way, that&#39;s a small cost for reducing the risk of major depression &amp;amp; suicide.)&lt;/p&gt;
&lt;p&gt;And it makes evolutionary sense that 10,000 IU a day &lt;em&gt;should&lt;/em&gt; be safe. Your skin, exposed to the Sun&#39;s ultraviolet rays, can synthesize up to (the equivalent of) 10,000 IU a day, before plateauing out. Source is &lt;a href=&quot;https://www.sciencedirect.com/science/article/pii/S0002916522043763&quot;&gt;Vieth 1999&lt;/a&gt;: “Because vitamin D is potentially toxic, intake of [1000 IU/day] has been avoided even though the weight of evidence shows that the currently accepted [limit] of [2000 IU/day] is too low by at least 5-fold.” And &lt;a href=&quot;https://pmc.ncbi.nlm.nih.gov/articles/PMC5541280/&quot;&gt;Papadimitriou 2017&lt;/a&gt; reviews several previous studies that find statistical errors behind official recommendations; correcting for these, adults should get 8000 IU/day.&lt;/p&gt;
&lt;p&gt;(On the other hand, &lt;a href=&quot;https://www.sciencedirect.com/science/article/abs/pii/S1011134415301561&quot;&gt;Krzyścin et al 2016&lt;/a&gt; estimates that the existing hunter-gatherer group, the Hadza, get 2000 IU of Vitamin D from sun exposure, and their food is a poor source of Vitamin D. Given that &lt;em&gt;existing&lt;/em&gt; hunter-gatherers live in the areas the colonialists &lt;em&gt;didn&#39;t&lt;/em&gt; want, ancient hunter-gatherers probably ate &amp;amp; got more Vitamin D. So, 2000 IU is a &lt;em&gt;lower&lt;/em&gt; bound on how much Vitamin D one should get — still several times more than the current official recommended dose of 400–800 IU/day.)&lt;/p&gt;
&lt;p&gt;So why are all the official sources still so paranoid about Vitamin D, and lowballing the recommendations? Well, alas, official policy is always a few decades behind the science in &lt;em&gt;any&lt;/em&gt; field. See: trans fats, open-label placebos, aerosol transmission of Covid-19, etc. And because something something incentives, it&#39;s &lt;em&gt;&amp;quot;rational&amp;quot;&lt;/em&gt; for government/insurers to be very risk-averse &amp;amp; slow to change (for better &amp;amp; worse).&lt;/p&gt;
&lt;p&gt;&lt;u&gt;Speaking of the Sun,&lt;/u&gt; why take supplements instead of just getting Vitamin D from sun exposure? Well, &lt;a href=&quot;https://pubmed.ncbi.nlm.nih.gov/29659012/&quot;&gt;skin cancer&lt;/a&gt;. But also: because Sun-Skin D varies greatly depending on the season, your latitude, and your skin type. There&#39;s less ultraviolet rays from the Sun in winter/fall, and at latitudes further from the equator. And &lt;a href=&quot;https://pmc.ncbi.nlm.nih.gov/articles/PMC10861575/&quot;&gt;the darker your skin is, the less Vitamin D your skin makes for the same amount of Sun exposure&lt;/a&gt;. As expected from the bio-physics of skin, Black adults have &lt;a href=&quot;https://pmc.ncbi.nlm.nih.gov/articles/PMC6075634/&quot;&gt;the highest prevalence of Vitamin D deficiency&lt;/a&gt; (82.1%!!), followed by Hispanic adults (62.9%). (But hey, at least Black adults have the lowest incidence of skin cancer. You win some you lose some.) The point is: speaking as someone with Southeast Asian skin, who&#39;s currently in Canada during winter... even if I stood outside naked for hours, I&#39;d get approximately &lt;em&gt;zero&lt;/em&gt; IU/day of Vitamin D from the Sun. Thus: supplements.&lt;/p&gt;
&lt;p&gt;&lt;u&gt;Direct effect on suicide:&lt;/u&gt; Finally, a meta-analysis &lt;em&gt;directly&lt;/em&gt; measuring the effect of Vitamin D on suicidal behaviour. &lt;a href=&quot;https://link.springer.com/content/pdf/10.1186/s12888-025-06613-w.pdf&quot;&gt;Yu et al 2025&lt;/a&gt;: “Vitamin D in patients with [suicidal behaviours] were significantly lower than in controls (standardized mean difference: –0.69, or a &#39;medium&#39; difference)”. Reminder that this paper &lt;em&gt;by itself&lt;/em&gt; only measures correlation, not causation — but combined with the above experiments of Vitamin D on depression, I think it&#39;s reasonable to guess it&#39;s partly causal.&lt;/p&gt;
&lt;p&gt;&lt;u&gt;To recap:&lt;/u&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Almost half of you have a Vitamin D insufficiency according to the official recommendation (800 IU/day).&lt;/li&gt;
&lt;li&gt;And those official recommendations are &lt;em&gt;way&lt;/em&gt; too low. The optimal amount of Vitamin D for depression is probably 4000 IU/day, with an effect around &lt;em&gt;twice&lt;/em&gt; that of the best antidepressant.&lt;/li&gt;
&lt;li&gt;Even the official &lt;em&gt;maximum&lt;/em&gt; safe dose (4000 IU/day) is below what your body can produce from the Sun in optimal conditions (10,000 IU/day). Recent randomized controlled trials confirm that 10,000 IU/day is, indeed, mostly safe.&lt;/li&gt;
&lt;li&gt;Reminder that official policy is often decades behind the science.&lt;/li&gt;
&lt;li&gt;Reminder that I&#39;m &lt;em&gt;not&lt;/em&gt; saying &amp;quot;take supplements &lt;em&gt;instead&lt;/em&gt; of antidepressants&amp;quot;; in fact the above meta-analysis shows you can effectively stack them!&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;Bonus: Vitamin D supplementation was found in several randomized controlled trials &lt;a href=&quot;https://pubmed.ncbi.nlm.nih.gov/39225947/&quot;&gt;to reduce mortality from Covid-19&lt;/a&gt;, though much less than official treatments like Paxlovid. Vitamin D also &lt;a href=&quot;https://www.sciencedirect.com/science/article/pii/S2161831322010274&quot;&gt;probably helps guard against influenza&lt;/a&gt; too, though the evidence is small &amp;amp; early.&lt;/p&gt;
&lt;hr /&gt;
&lt;p&gt;&lt;a id=&quot;conclusion&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h2&gt;Conclusion: &lt;em&gt;All this time, you lacked the Vitamin?&lt;/em&gt;&lt;/h2&gt;
&lt;p&gt;Scurvy is caused by a lack of Vitamin C. It&#39;s a condition that causes your wounds to re-open up &amp;amp; teeth to fall out. Scurvy used to kill &lt;em&gt;almost half(!)&lt;/em&gt; of all sailors on major expeditions; it&#39;s estimated millions died. It can be cured by eating lemons.&lt;/p&gt;
&lt;p&gt;Rickets is mostly caused by a lack of Vitamin D. It&#39;s a condition where kids&#39; bones go all soft and deformed. During the Industrial Revolution, up to 80% of kids suffered from it. It can be prevented with cod liver oil.&lt;/p&gt;
&lt;p&gt;Goiters is mostly caused by a lack of Iodine. It&#39;s a condition where the thyroid gland in your neck swells up painfully, to the size of an apple. During WWI, a &lt;em&gt;third&lt;/em&gt; of adult men had goiters. It can be prevented with iodized salt.&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2026-01/depression/The-Vitamin.png&quot; alt=&quot;Tumblr meme: &#39;All this time, you lacked The Vitamin? And yet you persisted?&#39;&quot; title=&quot;Tumblr meme: &#39;All this time, you lacked The Vitamin? And yet you persisted?&#39;&quot; /&gt;&lt;/p&gt;
&lt;p&gt;&lt;a href=&quot;https://ourworldindata.org/mental-health&quot;&gt;About 1 in 4 people are expected to have clinical depression sometime in their life&lt;/a&gt;. Depression is the #1 source of the global &amp;quot;burden from disease&amp;quot; in the mental health category, &lt;a href=&quot;https://ourworldindata.org/burden-of-disease#how-do-different-diseases-and-disabilities-contribute-towards-the-burden-of-disease&quot;&gt;and &lt;em&gt;that&lt;/em&gt; category is the #6 burden of disease in the world&lt;/a&gt;, above Alzheimer&#39;s, malaria, and sexually transmitted infections.&lt;/p&gt;
&lt;p&gt;(But honestly, did you need those stats? This is likely a lived experience for a lot of you reading this.)&lt;/p&gt;
&lt;p&gt;The effective altruists are all, &lt;a href=&quot;https://www.givewell.org/how-much-does-it-cost-to-save-a-life&quot;&gt;&amp;quot;woah for just $3000 you can prevent a child&#39;s death from malaria&amp;quot;&lt;/a&gt; — and that&#39;s great! save them kids! — but where&#39;s the fanfare for the accumulating evidence that, &amp;quot;woah with cheap daily supplements we can save millions from suicide &amp;amp; depressed lives&amp;quot;?&lt;/p&gt;
&lt;p&gt;Over and over again throughout history, some horrific thing that caused millions to suffer, turned out to be &amp;quot;yeah you were missing this one molecule lol&amp;quot;. To be clear: not everything is gonna be &lt;em&gt;that&lt;/em&gt; simple, and mental health is &lt;em&gt;not&lt;/em&gt; &amp;quot;just&amp;quot; chemistry. Also, all the numbers on this page have with large error bars &amp;amp; uncertainty, more research is needed.&lt;/p&gt;
&lt;p&gt;But, as of right now, I feel I can at least confidently claim the following:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Vitamin D and Omega-3 are both &lt;em&gt;at least on par&lt;/em&gt; with the median antidepressant (effect size ~= +0.3).&lt;/li&gt;
&lt;li&gt;The evidence is much stronger for Vitamin D; it&#39;s very plausibly at least &lt;em&gt;twice&lt;/em&gt; as good as antidepressants.&lt;/li&gt;
&lt;li&gt;Both supplements are cheap and safe, so what&#39;s the harm of trying? (positive &amp;quot;expected value&amp;quot; for this bet)&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;So:&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;MY RECOMMENDATIONS FOR RESEARCHERS:&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;More big, pre-registered, double-blind randomized controlled trials, please. (And specifically: testing &lt;em&gt;high&lt;/em&gt; doses.)&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;MY SPECIFIC RECOMMENDATIONS FOR &lt;em&gt;YOU:&lt;/em&gt;&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Go to a pharmacy, buy the following supplements over-the-counter, in whatever form you like: (I like the easy-to-swallow gel capsules)&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Vitamin D&lt;/strong&gt;
&lt;ul&gt;
&lt;li&gt;🌱 By default, Vitamin D supplements are derived from… (quick web search)… &lt;em&gt;the grease in sheep&#39;s wool?&lt;/em&gt; Huh. Also fish liver oil. Anyway, if you&#39;re vegan, make sure your bottle specifically says &amp;quot;vegan&amp;quot; or &amp;quot;from lichen/mushrooms&amp;quot;. (If you&#39;re vegetarian, the sheep&#39;s-wool Vitamin D is fine, they don&#39;t kill the sheep for it.)&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Omega-3 &lt;em&gt;where EPA is ~60% of the Omega-3 total.&lt;/em&gt;&lt;/strong&gt; For example, my 500mg Omega-3 capsules have 300mg EPA, 200mg DHA.
&lt;ul&gt;
&lt;li&gt;🌱 By default, Omega-3 supplements come from fish. If you&#39;re veg(etari)?an, there &lt;em&gt;are&lt;/em&gt; plant-based sources of Omega-3, but look carefully: most vegan Omega-3 supplements provide &lt;em&gt;more DHA than EPA&lt;/em&gt;, which the above studies suggest fully cancel out the antidepressant effect. &lt;em&gt;Double check the nutritional label to make sure it&#39;s ≥60% EPA.&lt;/em&gt; For example, &lt;a href=&quot;https://www.amazon.ca/Natures-Way-NutraVege-Plant-Based-Vegetarian/dp/B078T4DC2K?th=1&quot;&gt;this one is 300mg EPA + 200mg DHA&lt;/a&gt;. (&lt;em&gt;not&lt;/em&gt; an affiliate link)&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;strong&gt;Then, every day:&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;Take ~4000 IU of Vitamin D&lt;/strong&gt;
&lt;ul&gt;
&lt;li&gt;⚠️ be cautious if you have kidney stones, or are on medications that could interact with Vitamin D. &amp;quot;ask your doctor&amp;quot;.&lt;/li&gt;
&lt;li&gt;(4,000 IU is the official max safe dose)&lt;/li&gt;
&lt;li&gt;10,000 IU if you&#39;re feeling daring (a couple large controlled trials showed no major lasting adverse side effects).&lt;/li&gt;
&lt;li&gt;if you have darker skin / live in higher latitudes / it&#39;s winter, you definitely need &lt;em&gt;some&lt;/em&gt; form of vit D supplementation&lt;/li&gt;
&lt;li&gt;bonus: &lt;em&gt;may&lt;/em&gt; improve immune response to Covid &amp;amp; influenza?&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Take ~1500 mg of ≥60%-EPA Omega-3&lt;/strong&gt;
&lt;ul&gt;
&lt;li&gt;⚠️ be cautious if you&#39;re on blood thinners, or other medications that could interact with Omega-3. again, &amp;quot;ask your doctor&amp;quot;.&lt;/li&gt;
&lt;li&gt;(5000 mg/day is the official max safe dose)&lt;/li&gt;
&lt;li&gt;bonus: &lt;em&gt;may&lt;/em&gt; improve cognition?&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Don&#39;t quit your existing antidepressants if they&#39;re net-positive for you!&lt;/strong&gt;
&lt;ul&gt;
&lt;li&gt;you may also want to ask your doctor about &lt;em&gt;Amitriptyline&lt;/em&gt;, or those other best-effect-size antidepressants.&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;Can you get these doses of Vitamin D &amp;amp; Omega-3 through whole foods alone, no supplements? Probably, but it&#39;d be expensive &amp;amp; tedious: you&#39;d have to eat something like 2,000 calories of farmed salmon &lt;em&gt;a day&lt;/em&gt; to get 4,000 IU/day of Vitamin D. As for Omega-3, eating mostly oily fishes &lt;em&gt;would&lt;/em&gt; get you &amp;gt;1000mg of Omega-3, but they&#39;d be &lt;em&gt;more DHA than EPA&lt;/em&gt;, which the above studies suggest would cancel out the antidepressant effects.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;The effect sizes on depression:&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;The best antidepressant: &lt;strong&gt;+0.417&lt;/strong&gt;
&lt;ul&gt;
&lt;li&gt;like your mental health grade going from F to F+, or C to C+&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;1500mg of ≥60%-EPA Omega-3: &lt;strong&gt;+0.558&lt;/strong&gt;
&lt;ul&gt;
&lt;li&gt;like your mental health grade going from F to D–, or C to B–&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;4000 IU of Vitamin D: &lt;strong&gt;+0.98&lt;/strong&gt;
&lt;ul&gt;
&lt;li&gt;like your mental health grade going from F to D, or C to B&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;For completeness &amp;amp; comparison, here&#39;s the effect size of other things on depression:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Any &lt;a href=&quot;https://en.wikipedia.org/wiki/Dodo_bird_verdict&quot;&gt;mainstream &amp;quot;bona-fide&amp;quot; psychotherapy&lt;/a&gt; (CBT, Psychodynamic, Humanist, Solutions-Focused): &lt;strong&gt;+0.35&lt;/strong&gt;, source: &lt;a href=&quot;https://pmc.ncbi.nlm.nih.gov/articles/PMC5244449/&quot;&gt;Kamenov et al 2016&lt;/a&gt;
&lt;ul&gt;
&lt;li&gt;like going from C to C+&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;Aerobic/Cardio Exercise: &lt;strong&gt;+0.79&lt;/strong&gt;, source &lt;a href=&quot;https://pure-oai.bham.ac.uk/ws/portalfiles/portal/54143117/Ioannis_Morres_et_al_Aerobic_exercise_for_adult_patients_Depression_and_Anxiety_2018.pdf&quot;&gt;Ioannis et al 2018&lt;/a&gt;
&lt;ul&gt;
&lt;li&gt;like going from C to B–&lt;/li&gt;
&lt;li&gt;(dose: &amp;quot;45 minutes, at moderate intensity, three times/week&amp;quot; ⇒ ~20 min/day)&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;Good Sleep: &lt;strong&gt;+1.10(???)&lt;/strong&gt;, a &lt;em&gt;lot&lt;/em&gt; of interpretation &amp;amp; calculations, see footnote&lt;sup class=&quot;footnote-ref&quot;&gt;&lt;a href=&quot;https://blog.ncase.me/on-depression/#fn5&quot; id=&quot;fnref5&quot;&gt;[5]&lt;/a&gt;&lt;/sup&gt;
&lt;ul&gt;
&lt;li&gt;like going from C to B&lt;/li&gt;
&lt;li&gt;(dose: going from moderate insomnia to healthy sleep)&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;Bright Light Therapy: &lt;strong&gt;+0.487&lt;/strong&gt;, source &lt;a href=&quot;https://pmc.ncbi.nlm.nih.gov/articles/PMC11447633/&quot;&gt;Menegaz de Almeida et al 2025&lt;/a&gt;
&lt;ul&gt;
&lt;li&gt;(the above paper reports Odds Ratio of 2.42, &lt;a href=&quot;https://www.escal.site/&quot;&gt;which converts to&lt;/a&gt; Cohen&#39;s d effect size of +0.487)&lt;/li&gt;
&lt;li&gt;like going from C to C+&lt;/li&gt;
&lt;li&gt;I went &lt;a href=&quot;https://www.nytimes.com/wirecutter/reviews/best-light-therapy-lamp/&quot;&gt;with Wirecutter&#39;s recommendation&lt;/a&gt; for a UV-free 10,000 lux lamp.&lt;/li&gt;
&lt;li&gt;(dose: 10,000 lux, 30 min a day)&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;Mindfulness Meditation: &lt;strong&gt;+0.42&lt;/strong&gt;, source &lt;a href=&quot;https://www.frontiersin.org/journals/psychiatry/articles/10.3389/fpsyt.2019.00193/full&quot;&gt;Breedvelt et al 2019&lt;/a&gt;
&lt;ul&gt;
&lt;li&gt;like going from C to C+&lt;/li&gt;
&lt;li&gt;(dose: 7 weeks, &amp;quot;153 min each week&amp;quot; ⇒ ~20 min/day)&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2026-01/depression/on-depression-summary.png&quot; alt=&quot;Diagram of all the above estimated effect sizes&quot; title=&quot;Diagram of all the above estimated effect sizes&quot; /&gt;&lt;/p&gt;
&lt;p&gt;(And remember to add +0.4 for the power of hope, e.g. &amp;quot;placebo&amp;quot;! Also remember: you can &lt;em&gt;stack&lt;/em&gt; any of the above interventions to get an even larger effect! You can&#39;t just naively add up the effect sizes, but I&#39;d be surprised if the effect of {vitamin d + omega-3 + bright lamps + cardio + good sleep + meditation + therapy + antidepressants} &lt;em&gt;combined&lt;/em&gt; ends up being less than +2.00. Two letter grades up means going from D to B, or, theoretically, from clinically depressed to flourishing! For more papers &amp;amp; my working research notes on &amp;quot;best bang for buck on depression&amp;quot;, &lt;a href=&quot;https://docs.google.com/document/d/1hjNxJmNjoVktrJDFPtFH1hgzKmGpZkRFQ6Vwf0s7c6w/edit?tab=t.0#heading=h.e0blw3xq3cvm&quot;&gt;check out this Google Doc&lt;/a&gt;.)&lt;/p&gt;
&lt;p&gt;Also, remember that all the above estimates are uncertain. And that averages hide &lt;em&gt;variation&lt;/em&gt; in effect. That said, I still think the overall picture is still strong: there exist high bang-for-buck ways to reduce depression, which are &lt;em&gt;at least on par&lt;/em&gt; with drugs &amp;amp; therapy (and plausibly 2x better), that aren&#39;t (yet) common knowledge amongst policymakers &amp;amp; the public. And again, they&#39;re dirt cheap with minor-to-no adverse side effects. Moderate chance of a big win, for a known tiny cost. That&#39;s a positive &amp;quot;expected value&amp;quot; bet right there.&lt;/p&gt;
&lt;p&gt;I got onto this research rabbithole a few months ago while borrowing my housemate&#39;s ADHD meds, which I may or may not eventually collect into a &amp;quot;JOYMAXXING&amp;quot; informal meta-meta-analysis. (&lt;a href=&quot;https://youtu.be/JOj97Edna7k?si=xAUt3Kp54RyfGX50&amp;amp;t=334&quot;&gt;: See me yap about it on video as a cartoon cat&lt;/a&gt;.) But for this blog post, I wanted to dive deeper into Vitamin D and Omega-3, since their effect sizes are so huge, &lt;em&gt;and&lt;/em&gt; they&#39;re insultingly cheap &amp;amp; easy, compared to therapy or regular cardio.&lt;/p&gt;
&lt;p&gt;Stay safe this winter, keep away the seasonal depression. Get your supplements, and reach out to your friends &amp;amp; loved ones!&lt;/p&gt;
&lt;p&gt;💖,&lt;br /&gt;
~ Nicky Case&lt;/p&gt;
&lt;iframe src=&quot;https://ncase.me/ncase-credits/signup2.html&quot; frameborder=&quot;no&quot; width=&quot;640&quot; height=&quot;250&quot;&gt;&lt;/iframe&gt;
&lt;iframe src=&quot;https://ncase.me/ncase-credits/supporters/jan-2026.html&quot; frameborder=&quot;no&quot; width=&quot;640&quot; height=&quot;640&quot;&gt;&lt;/iframe&gt;&lt;hr class=&quot;footnotes-sep&quot; /&gt;
&lt;section class=&quot;footnotes&quot;&gt;
&lt;ol class=&quot;footnotes-list&quot;&gt;
&lt;li id=&quot;fn1&quot; class=&quot;footnote-item&quot;&gt;&lt;p&gt;I made up these ranges by requiring the standard letter grades F,D,C,B,A, to have their centers be -2,-1,0,+1,+2. Then, I made sure all in-between grades like C+ or A– had equal intervals. Each interval is +/- ⅙, or ⅓ wide:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;F---: -3.16 to -2.83&lt;/li&gt;
&lt;li&gt;F--: -2.82 to -2.50&lt;/li&gt;
&lt;li&gt;F–: -2.49 to -2.17&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;F: -2.16 to -1.83&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;F+: -1.82 to -1.50&lt;/li&gt;
&lt;li&gt;D–: -1.49 to -1.17&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;D: -1.16 to -0.83&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;D+: -0.82 to -0.50&lt;/li&gt;
&lt;li&gt;C–: -0.49 to -0.17&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;C: -0.16 to +0.17&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;C+: +0.18 to +0.50&lt;/li&gt;
&lt;li&gt;B–: +0.51 to +0.83&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;B: +0.84 to +1.17&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;B+: +1.18 to +1.50&lt;/li&gt;
&lt;li&gt;A–: +1.51 to +1.83&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;A: +1.84 to +2.17&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;A+: +2.18 to +2.50&lt;/li&gt;
&lt;li&gt;A++: +2.51 to +2.83&lt;/li&gt;
&lt;li&gt;A+++: +2.84 to +3.17&lt;/li&gt;
&lt;/ul&gt;
 &lt;a href=&quot;https://blog.ncase.me/on-depression/#fnref1&quot; class=&quot;footnote-backref&quot;&gt;↩︎&lt;/a&gt;&lt;/li&gt;
&lt;li id=&quot;fn2&quot; class=&quot;footnote-item&quot;&gt;&lt;p&gt;Median antidepressant Odds Ratio: 1.66, see below figure. This &lt;a href=&quot;https://www.escal.site/&quot;&gt;converts to&lt;/a&gt; Cohen&#39;s d of 0.279. &lt;a href=&quot;https://blog.ncase.me/on-depression/#fnref2&quot; class=&quot;footnote-backref&quot;&gt;↩︎&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li id=&quot;fn3&quot; class=&quot;footnote-item&quot;&gt;&lt;p&gt;See &lt;a href=&quot;https://academic.oup.com/ijnp/article-pdf/14/3/405/2688342/14-3-405.pdf&quot;&gt;Fountoulakis &amp;amp; Möller 2011&lt;/a&gt; Table 1 Row 2. Will talk more about this paper again in a few paragraphs. &lt;a href=&quot;https://blog.ncase.me/on-depression/#fnref3&quot; class=&quot;footnote-backref&quot;&gt;↩︎&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li id=&quot;fn4&quot; class=&quot;footnote-item&quot;&gt;&lt;p&gt;See Table 1. A follow-up paper by Kirsch in in 2008 found that the drug group improved by 9.60 (non-standardized) points, and the placebo group by 7.80 points. (So, placebo + time is 7.80/9.60 = 0.81 = 81% of the full effect.) The F&amp;amp;M recalculation found the drug group improved by 10.04 points, and the placebo by 7.85 points. (So, placebo + time is 7.85/10.04 = 0.78 = 78% of the full effect.) And rows 2 &amp;amp; 3 confirm that Kirsch was still right about the following: “The [total effect] for drug groups was 1.24 [C to B+] and that for placebo 0.92 [C to B]” and “The effect size concerning the difference between improvement in drug groups and improvement in placebo groups was 0.32 [like C to C+]”. &lt;a href=&quot;https://blog.ncase.me/on-depression/#fnref4&quot; class=&quot;footnote-backref&quot;&gt;↩︎&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li id=&quot;fn5&quot; class=&quot;footnote-item&quot;&gt;&lt;p&gt;&lt;a href=&quot;https://pmc.ncbi.nlm.nih.gov/articles/PMC10039857/&quot;&gt;Lee et al 2023&lt;/a&gt; reports the following effect sizes. Digital therapy for Insomnia → Sleep = 0.76, and Digital therapy for Insomnia → Depression = 0.42. Assuming the therapy &lt;em&gt;for insomnia specifically&lt;/em&gt; affects depression &lt;em&gt;only&lt;/em&gt; through better sleep (Digital therapy for Insomnia → Sleep → Depression), we can do an &lt;a href=&quot;https://cameron.econ.ucdavis.edu/e240a/ch04iv.pdf&quot;&gt;&amp;quot;Instrumental Variable&amp;quot;&lt;/a&gt; estimate of the effect of Sleep → Depression = 0.42 / 0.76 = 0.55. &lt;strong&gt;To be precise: this is saying, if you improve your sleep by 1 standard deviation, on average your depression improves by 0.55 standard deviations.&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;So: how many standard deviations is going from &amp;quot;moderate insomnia&amp;quot; to &amp;quot;healthy sleep&amp;quot;? The standard measure is the Insomnia Severity Index (ISI), which &lt;a href=&quot;https://www.sleepprimarycareresources.org.au/questionnaires/isi&quot;&gt;you can take online&lt;/a&gt;. A score of 0–7 means no insomnia, 8–14 is subclinical insomnia, 15–21 is clinical insomnia (moderate), 22–28 is clinical insomnia (severe). Let&#39;s be conservative and say we&#39;re just going from barely clinical to barely healthy: 15 to 7, or a reduction of 8 points. &lt;a href=&quot;https://pubmed.ncbi.nlm.nih.gov/19689221/&quot;&gt;Yang et al 2009&lt;/a&gt; says a 6-point reduction is 1.5 standard deviations, which means 4 points is 1 standard deviation. So a reduction of 8 points is 2 standard deviations. &lt;strong&gt;So, if you improve your sleep from insomniac to healthy, you improve by at least 8 points, which is 2 standard deviations, so your depression should improve by 2 × 0.55 standard deviations, or ~1.10.&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;Reminder that my estimate is &lt;em&gt;full&lt;/em&gt; of assumptions upon assumptions &amp;amp; these error bars will compound. But I&#39;d be surprised if the true causal effect of going from insomniac to healthy sleep isn&#39;t &lt;em&gt;at least&lt;/em&gt; a &amp;quot;large&amp;quot; +0.8 effect. &lt;a href=&quot;https://blog.ncase.me/on-depression/#fnref5&quot; class=&quot;footnote-backref&quot;&gt;↩︎&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ol&gt;
&lt;/section&gt;
</content>
        </entry>

    

        
        
        <entry>
            <title>a human poem on humans liking AI poems over human poems</title>
            <link href="https://blog.ncase.me/poem/"/>
            <updated>2026-01-07T00:00:00Z</updated>
            <id>https://blog.ncase.me/poem/</id>
            <content xml:lang="en" type="html">&lt;p&gt;Oh! To compete!&lt;br /&gt;
With an autocomplete!&lt;br /&gt;
In a recent RCT&lt;sup class=&quot;footnote-ref&quot;&gt;&lt;a href=&quot;https://blog.ncase.me/poem/#fn1&quot; id=&quot;fnref1&quot;&gt;[1]&lt;/a&gt;&lt;/sup&gt;&lt;br /&gt;
They found that good ol&#39; GPT&lt;br /&gt;
Could beat!&lt;br /&gt;
A human like me.&lt;/p&gt;
&lt;p&gt;Oh! Dare I ask!&lt;br /&gt;
At what special human task!&lt;br /&gt;
Poetry, and humans guessed&lt;br /&gt;
worse than chance, at the test&lt;br /&gt;
of who&#39;s human, and confessed&lt;br /&gt;
to like AI above the rest:&lt;br /&gt;
Obsolete!&lt;br /&gt;
A human like me.&lt;/p&gt;
&lt;p&gt;Oh! What a shame!&lt;br /&gt;
Who&#39;s the scapegoat I shall blame!&lt;br /&gt;
The masses, their dumb asses?&lt;br /&gt;
Postmodern poets who don&#39;t rhyme?&lt;br /&gt;
The study, it&#39;s too cruddy?&lt;br /&gt;
Sam Altman tryna make a dime?&lt;/p&gt;
&lt;p&gt;&lt;code&gt;Or, let&#39;s go back in time?&lt;/code&gt;&lt;/p&gt;
&lt;p&gt;Copernicus! You &lt;em&gt;did&lt;/em&gt; frick us!&lt;br /&gt;
WE ARE NOT THE CENTER.&lt;br /&gt;
Darwin-Wallace! &lt;em&gt;Apes&lt;/em&gt;, you call us?!&lt;br /&gt;
WE ARE NOT THE CENTER.&lt;br /&gt;
Turing-Gödel! Turning hurdles!&lt;br /&gt;
Understanding is compression.&lt;sup class=&quot;footnote-ref&quot;&gt;&lt;a href=&quot;https://blog.ncase.me/poem/#fn2&quot; id=&quot;fnref2&quot;&gt;[2]&lt;/a&gt;&lt;/sup&gt;&lt;br /&gt;
Intelligence is search.&lt;sup class=&quot;footnote-ref&quot;&gt;&lt;a href=&quot;https://blog.ncase.me/poem/#fn3&quot; id=&quot;fnref3&quot;&gt;[3]&lt;/a&gt;&lt;/sup&gt;&lt;br /&gt;
Existential-o depression&lt;br /&gt;
from each scientific lurch.&lt;/p&gt;
&lt;p&gt;WE ARE NOT THE CENTER.&lt;br /&gt;
THERE IS NOTHING SPECIAL ABOUT BEING:&lt;/p&gt;
&lt;p&gt;A human like me.&lt;/p&gt;
&lt;p&gt;.&lt;/p&gt;
&lt;p&gt;.&lt;/p&gt;
&lt;p&gt;.&lt;/p&gt;
&lt;p&gt;O... kay....?&lt;/p&gt;
&lt;p&gt;Pretty erudite, for a Luddite&lt;br /&gt;
I was playing. Science is a&#39;ight, a&#39;ight?&lt;br /&gt;
I like not dying of cholera or smallpox&lt;br /&gt;
or burying half our kids in a small box&lt;br /&gt;
(as they did before germ theory&lt;sup class=&quot;footnote-ref&quot;&gt;&lt;a href=&quot;https://blog.ncase.me/poem/#fn4&quot; id=&quot;fnref4&quot;&gt;[4]&lt;/a&gt;&lt;/sup&gt;)&lt;br /&gt;
It&#39;s erm, eerie,&lt;br /&gt;
how the past wasn&#39;t health-compliant.&lt;br /&gt;
I&#39;d have to be an ungrateful brat&lt;br /&gt;
to piss on the shoulders of giants.&lt;/p&gt;
&lt;p&gt;And yet, the ape-part of me&lt;br /&gt;
the part that lives on a Disc World&lt;br /&gt;
under a dome speckled with Gods,&lt;br /&gt;
the part that hates poetry that doesn&#39;t rhyme,&lt;br /&gt;
the part that lives&lt;br /&gt;
&lt;em&gt;in the center&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;That part cries out!&lt;br /&gt;
It&#39;s got to shout!&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Oh! Myth is dead.&lt;br /&gt;
It&#39;s ok to feel bereft.&lt;br /&gt;
Oh! The ineffable!	&lt;br /&gt;
&lt;em&gt;It&#39;s been fully eff&#39;d!&lt;/em&gt;&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;Once there was a time when Art was the pinnacle&lt;br /&gt;
of human glory. Then a billion dollars and a nickel&lt;br /&gt;
later, a next-token-predictor can write poetry the fickle&lt;br /&gt;
humans prefer over human poetry. It tickles&lt;br /&gt;
my ironic nerves that we now take superior pride in,&lt;br /&gt;
what,&lt;br /&gt;
Counting the R&#39;s in &amp;quot;strawberry&amp;quot;?&lt;br /&gt;
Playing a pixel puzzle game?&lt;sup class=&quot;footnote-ref&quot;&gt;&lt;a href=&quot;https://blog.ncase.me/poem/#fn5&quot; id=&quot;fnref5&quot;&gt;[5]&lt;/a&gt;&lt;/sup&gt;&lt;br /&gt;
Drawing hands that aren&#39;t scary?&lt;br /&gt;
&lt;em&gt;That&#39;s&lt;/em&gt; our claim to fame?&lt;/p&gt;
&lt;p&gt;(bish, &lt;em&gt;i&lt;/em&gt; can&#39;t draw hands either 😭)&lt;/p&gt;
&lt;p&gt;And sure, this current boom may fail&lt;br /&gt;
We&#39;ve hit diminishing returns on scale&lt;sup class=&quot;footnote-ref&quot;&gt;&lt;a href=&quot;https://blog.ncase.me/poem/#fn6&quot; id=&quot;fnref6&quot;&gt;[6]&lt;/a&gt;&lt;/sup&gt;&lt;br /&gt;
Backprop, big data and a VC whale&lt;br /&gt;
May not be enough this time.&lt;/p&gt;
&lt;p&gt;&lt;code&gt;Or, let&#39;s go forth in time?&lt;/code&gt;&lt;/p&gt;
&lt;p&gt;Not enough &lt;em&gt;this&lt;/em&gt; time.&lt;/p&gt;
&lt;p&gt;But it&#39;s &lt;em&gt;only&lt;/em&gt; a matter of time.&lt;/p&gt;
&lt;p&gt;We will not be the center.&lt;/p&gt;
&lt;p&gt;We never were.&lt;/p&gt;
&lt;p&gt;And that&#39;s fine.&lt;/p&gt;
&lt;p&gt;They raised us to love being special&lt;br /&gt;
(then complain that &amp;quot;narcissism&amp;quot; is up&lt;sup class=&quot;footnote-ref&quot;&gt;&lt;a href=&quot;https://blog.ncase.me/poem/#fn7&quot; id=&quot;fnref7&quot;&gt;[7]&lt;/a&gt;&lt;/sup&gt;)&lt;br /&gt;
But growing up means realizing it&#39;s okay&lt;br /&gt;
to not be special&lt;br /&gt;
to not be the center of Attention and/or Creation&lt;br /&gt;
to not have poetry that scans good&lt;br /&gt;
Fuck it, I&#39;m having fun&lt;br /&gt;
Once we automate it all away&lt;br /&gt;
And there&#39;s nothing we &lt;em&gt;need&lt;/em&gt; to do&lt;br /&gt;
And there&#39;s nothing we &lt;em&gt;need&lt;/em&gt; from each other&lt;br /&gt;
We&#39;ll be forced to answer,&lt;br /&gt;
scared&lt;br /&gt;
sacred&lt;br /&gt;
What do we &lt;em&gt;want&lt;/em&gt; to do?&lt;br /&gt;
What do we &lt;em&gt;want&lt;/em&gt; to be for each other?&lt;/p&gt;
&lt;p&gt;&lt;em&gt;Whose center&lt;/em&gt;&lt;br /&gt;
do you want to be?&lt;/p&gt;
&lt;p&gt;&amp;quot;A human likes &lt;em&gt;me?&amp;quot;&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;.&lt;/p&gt;
&lt;p&gt;.&lt;/p&gt;
&lt;p&gt;.&lt;/p&gt;
&lt;p&gt;Wait, was &lt;em&gt;that&lt;/em&gt; my so-called wisdom?&lt;br /&gt;
A saccharine cliché?&lt;br /&gt;
&amp;quot;The real fully-automated space communism&lt;br /&gt;
was all the friends we made along the way&amp;quot;?&lt;br /&gt;
What about AI existential risk?&lt;br /&gt;
What was that about the world being a disk?&lt;br /&gt;
What about AI being used to rope ya&lt;br /&gt;
into a totalitarian dystopia&lt;br /&gt;
and oop, yeah&lt;br /&gt;
well&lt;br /&gt;
Were you expecting anything actually useful from a poem?&lt;br /&gt;
Art? Being &lt;em&gt;useful?&lt;/em&gt; lol&lt;br /&gt;
This is why humans prefer AI poetry&lt;br /&gt;
Shit, I can&#39;t even have an uplifting moment&lt;br /&gt;
Without undercutting it with ironic cynicism&lt;br /&gt;
(Hey! Maybe that&#39;s what&#39;ll stop rogue AI)&lt;br /&gt;
(By being trained on internet data)&lt;br /&gt;
(If it becomes self-aware it&#39;ll just get depressed)&lt;/p&gt;
&lt;p&gt;Fuck it, I&#39;m having fun&lt;/p&gt;
&lt;p&gt;Let&#39;s end this with a &lt;em&gt;true&lt;/em&gt; human poem, that I &lt;em&gt;know&lt;/em&gt; a corporate chatbot will never make: a poem, of copyrighted characters and sexual content.&lt;/p&gt;
&lt;p&gt;Behold, Art:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;There once was a mouse named Mickey&lt;br /&gt;
Who whipped out his mousey dickey&lt;br /&gt;
And started humping&lt;br /&gt;
Mr. Xi Jinping&lt;br /&gt;
While reading the Tiananmen Square Wiki&lt;/p&gt;
&lt;p&gt;(don&#39;t worry, it&#39;s consensual)&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;Oh! What a purge!&lt;br /&gt;
Of many a mental urge!&lt;br /&gt;
And my ape, who wants to sing the mythical&lt;br /&gt;
And my brain, who wants to eff the ineffable&lt;br /&gt;
And my heart, who&#39;s torn about being ethical&lt;br /&gt;
And the machine, maybe at the pinnacle&lt;br /&gt;
We&#39;ll all merge?...&lt;/p&gt;
&lt;p&gt;A human-like me.&lt;/p&gt;
&lt;p&gt;&lt;span&gt; &lt;br /&gt; &lt;br /&gt; &lt;br /&gt; &lt;br /&gt; &lt;br /&gt; &lt;br /&gt; &lt;br /&gt; &lt;br /&gt; &lt;br /&gt; &lt;br /&gt; &lt;br /&gt; &lt;br /&gt; &lt;br /&gt; &lt;br /&gt; &lt;br /&gt; &lt;br /&gt; &lt;br /&gt; &lt;br /&gt; &lt;br /&gt; &lt;br /&gt;&lt;/span&gt;&lt;/p&gt;
&lt;iframe src=&quot;https://ncase.me/ncase-credits/signup2.html&quot; frameborder=&quot;no&quot; width=&quot;640&quot; height=&quot;250&quot;&gt;&lt;/iframe&gt;
&lt;iframe src=&quot;https://ncase.me/ncase-credits/supporters/jan-2026.html&quot; frameborder=&quot;no&quot; width=&quot;640&quot; height=&quot;640&quot;&gt;&lt;/iframe&gt;
&lt;hr /&gt;
&lt;hr class=&quot;footnotes-sep&quot; /&gt;
&lt;section class=&quot;footnotes&quot;&gt;
&lt;ol class=&quot;footnotes-list&quot;&gt;
&lt;li id=&quot;fn1&quot; class=&quot;footnote-item&quot;&gt;&lt;p&gt;Well, &amp;quot;recent&amp;quot;. From &lt;a href=&quot;https://www.nature.com/articles/s41598-024-76900-1&quot;&gt;a 2024 paper&lt;/a&gt;: &amp;quot;participants performed below chance levels in identifying AI-generated poems [and] were more likely to judge AI-generated poems as human-authored than actual human-authored poems.&amp;quot; The human poets ranged from Shakespeare to Whitman to Plath. The AI poet was ChatGPT &lt;em&gt;3.5&lt;/em&gt;, with no prompt engineering, or feedback &amp;amp; iteration. &lt;a href=&quot;https://blog.ncase.me/poem/#fnref1&quot; class=&quot;footnote-backref&quot;&gt;↩︎&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li id=&quot;fn2&quot; class=&quot;footnote-item&quot;&gt;&lt;p&gt;&lt;strong&gt;&amp;quot;Understanding is Compression&amp;quot;&lt;/strong&gt; is an idea that&#39;s been around for centuries, if not exactly in those words. Ockham&#39;s Razor says that given two theories that explain the same thing, we should pick the simpler one. Einstein said &lt;em&gt;&amp;quot;A theory is more impressive the greater the simplicity of its premises, the more different things it relates, and the more expanded its area of application.”&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;And now, this idea is finding good use in AI! Neural networks trained with regularization (rewarding simplicity), and Auto-encoders (compress large input → small embedding → decompress back to original input), both lead to AI that&#39;s more robust &amp;amp; generalizes better.&lt;/p&gt;
&lt;p&gt;Hat tip to these papers: &lt;a href=&quot;https://spaces-cdn.owlstown.com/blobs/4pab25rnepck8d3lp5lq2jk5ynx7&quot;&gt;Understanding as Compression&lt;/a&gt; by Daniel A. Wilkenfeld, a delightful accessable read. &lt;a href=&quot;https://arxiv.org/pdf/1310.8599&quot;&gt;Information compression, intelligence, computing, and mathematics&lt;/a&gt; by J Gerard Wolff, founder of the SP (Simplicity-Power) Theory of Intelligence. &lt;a href=&quot;https://arxiv.org/pdf/2407.07723&quot;&gt;Understanding is Compression&lt;/a&gt; by Li, Huang, Wang, Hu, Wyeth, Bu, Yu, Gao, Liu &amp;amp; Li, which show that LLMs can compress text (and even images/audio) &lt;em&gt;better&lt;/em&gt; than standard compression algorithms! &lt;a href=&quot;https://blog.ncase.me/poem/#fnref2&quot; class=&quot;footnote-backref&quot;&gt;↩︎&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li id=&quot;fn3&quot; class=&quot;footnote-item&quot;&gt;&lt;p&gt;&lt;strong&gt;&amp;quot;Problem-solving is (heuristic) search&amp;quot;&lt;/strong&gt; is also an old idea:&lt;/p&gt;
&lt;p&gt;Turing &amp;amp; Champernowne designed &lt;a href=&quot;https://en.wikipedia.org/wiki/Turochamp&quot;&gt;the first chess AI&lt;/a&gt; in 1948, which searched every possible move &amp;amp; counter-move, then selected the best move, based on a &amp;quot;heuristic&amp;quot; (rule-of-thumb) based on the value of pieces, safety &amp;amp; mobility of the pieces, etc.&lt;/p&gt;
&lt;p&gt;Simon &amp;amp; Newell created the &lt;a href=&quot;https://en.wikipedia.org/wiki/General_Problem_Solver&quot;&gt;General Problem Solver&lt;/a&gt; in 1957, which could take &lt;em&gt;any&lt;/em&gt; formal game/system, like mazes or Sudokus or geometry, and search through possible moves (&amp;quot;operators&amp;quot;) to get to the solution. Because of the exponential explosion in possible states, their AI tried to narrow down its search using a means-ends heuristic: first try moves that &lt;em&gt;directly&lt;/em&gt; get you closer to your goal, backtrack only when that fails.&lt;/p&gt;
&lt;p&gt;Skipping decades ahead, DeepMind released &lt;a href=&quot;https://en.wikipedia.org/wiki/AlphaZero&quot;&gt;AlphaZero&lt;/a&gt; in 2017, which beat human grandmasters at chess, go, and shogi (&amp;quot;Japanese Chess&amp;quot;). The core of AlphaZero is the Monte Carlo Tree Search, which, once again, is &amp;quot;just&amp;quot; search through possible moves &amp;amp; counter-moves. But this time, instead of being hard-coded like in past chess AIs, the heuristic is a neural network that&#39;s learnt from scratch! &lt;a href=&quot;https://blog.ncase.me/poem/#fnref3&quot; class=&quot;footnote-backref&quot;&gt;↩︎&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li id=&quot;fn4&quot; class=&quot;footnote-item&quot;&gt;&lt;p&gt;For &lt;em&gt;thousands&lt;/em&gt; of years, across &lt;em&gt;all&lt;/em&gt; human cultures: &lt;a href=&quot;https://ourworldindata.org/child-mortality#key-insights&quot;&gt;the percent of kids who died before age 15 was around 50%&lt;/a&gt;. This only changed around ~1850, when the work of Jon Snow &amp;amp; Louis Pasteur (&amp;amp; many others) proved that diseases spread through &amp;quot;germs&amp;quot; that were invisible to the naked eye. Only after this scientific discovery, public health infrastructure &lt;em&gt;actually worked&lt;/em&gt;. Global child mortality plummeted from 50% then to &amp;quot;only&amp;quot; 4% now, and &amp;quot;only&amp;quot; 0.3% in the best countries. To quote a wise philosopher: &lt;a href=&quot;https://youtu.be/l4bW99gtXR8?si=HBuGgQPQPSKxltjD&amp;amp;t=187&quot;&gt;&amp;quot;YEAH SCIENCE, BITCH!&amp;quot;&lt;/a&gt; &lt;a href=&quot;https://blog.ncase.me/poem/#fnref4&quot; class=&quot;footnote-backref&quot;&gt;↩︎&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li id=&quot;fn5&quot; class=&quot;footnote-item&quot;&gt;&lt;p&gt;The &lt;a href=&quot;https://arcprize.org/&quot;&gt;Abstraction &amp;amp; Reasoning Corpus (ARC)&lt;/a&gt; is one of the few famous tests for AI that is (currently) unbeaten. You know &lt;a href=&quot;https://en.wikipedia.org/wiki/Raven%27s_Progressive_Matrices#Problem_structure&quot;&gt;those IQ tests&lt;/a&gt; where you look at some shapes, figure out the pattern, then select the shape that continues the pattern? ARC is like those, but 1) in pixel-game format, and 2) designed to be &amp;quot;easy for Humans, hard for AI&amp;quot;.&lt;/p&gt;
&lt;p&gt;The reason they&#39;re hard for current AI is because — unlike AI excelling at math word problems or reciting facts from textbooks — there &lt;em&gt;is no common pattern&lt;/em&gt; in the ARC puzzles, and &lt;em&gt;no way to memorize the answers&lt;/em&gt;, because the non-profit behind ARC uses a private test that&#39;s never been put online. (&lt;a href=&quot;https://arcprize.org/play&quot;&gt;but they have a public version humans can play with, and AIs can train with&lt;/a&gt;.) &lt;a href=&quot;https://blog.ncase.me/poem/#fnref5&quot; class=&quot;footnote-backref&quot;&gt;↩︎&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li id=&quot;fn6&quot; class=&quot;footnote-item&quot;&gt;&lt;p&gt;Toby Ord, co-founder of Giving What We Can and author of &lt;em&gt;The Precipice&lt;/em&gt; — a book on existential risk that gave AI the highest probability of all things that could cause human extinction this century — more recently, Ord is collecting stats that show, contrary to AI bulls, modern LLM improvements are coming at unsustainably exponential costs:&lt;/p&gt;
&lt;p&gt;&lt;a href=&quot;https://www.tobyord.com/writing/inference-scaling-and-the-log-x-chart&quot;&gt;Inference Scaling and the Log-x Chart&lt;/a&gt; shows how AI companies lie with statistics by making their &lt;em&gt;x&lt;/em&gt; axis exponential, and &lt;em&gt;not&lt;/em&gt; the y axis.&lt;/p&gt;
&lt;p&gt;&lt;a href=&quot;https://www.tobyord.com/writing/hourly-costs-for-ai-agents&quot;&gt;Are the Costs of AI Agents Also Rising Exponentially?&lt;/a&gt; shows that, unlike humans — whose performance is linear: we can do 10× as much work in 10× as much time — LLM agents hit diminishing returns &lt;em&gt;fast&lt;/em&gt;. &lt;a href=&quot;https://blog.ncase.me/poem/#fnref6&quot; class=&quot;footnote-backref&quot;&gt;↩︎&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li id=&quot;fn7&quot; class=&quot;footnote-item&quot;&gt;&lt;p&gt;While factchecking my poem, I found a 2024 meta-analysis that showed young people&#39;s &amp;quot;narcissism&amp;quot; has actually been &lt;em&gt;declining&lt;/em&gt; since 2008, possibly due to the financial crisis! &lt;a href=&quot;https://onlinelibrary.wiley.com/doi/pdf/10.1111/jopy.12982&quot;&gt;See Figure 5&lt;/a&gt;. I guess the misperception that &amp;quot;narcissism is rising&amp;quot; is due to narcissistic traits being &lt;em&gt;more visible &amp;amp; rewarded&lt;/em&gt; in a modern attention economy, not due to actual generational changes.&lt;/p&gt;
&lt;p&gt;Also, reminder that:&lt;/p&gt;
&lt;p&gt;1) &lt;a href=&quot;https://onlinelibrary.wiley.com/doi/pdf/10.1111/jopy.12982&quot;&gt;The &amp;quot;narcissism scale&amp;quot; is a flawed measure&lt;/a&gt; that mixes up healthy &amp;amp; unhealthy traits,&lt;/p&gt;
&lt;p&gt;2) Don&#39;t confuse &amp;quot;younger people score higher on narcissism&amp;quot; (age effect) with &amp;quot;each generation is getting more narcissistic&amp;quot; (cohort effect). &lt;a href=&quot;https://journals.sagepub.com/doi/10.1177/25152459251342750&quot;&gt;That&#39;s a fallacy&lt;/a&gt;, like confusing &amp;quot;younger people are shorter in height&amp;quot; and &amp;quot;each generation is getting shorter in height&amp;quot;.&lt;/p&gt;
&lt;p&gt;3) People with &lt;a href=&quot;https://en.wikipedia.org/wiki/Narcissistic_personality_disorder&quot;&gt;narcissistic personality disorder&lt;/a&gt; aren&#39;t cartoon villains, they&#39;re actual people. &lt;a href=&quot;https://blog.ncase.me/poem/#fnref7&quot; class=&quot;footnote-backref&quot;&gt;↩︎&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ol&gt;
&lt;/section&gt;
</content>
        </entry>

    

        
        
        <entry>
            <title>Connect your past, present &amp; future, with lines on a graph</title>
            <link href="https://blog.ncase.me/lines-on-a-graph/"/>
            <updated>2026-01-01T00:00:00Z</updated>
            <id>https://blog.ncase.me/lines-on-a-graph/</id>
            <content xml:lang="en" type="html">&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2026-01/lines.png&quot; alt=&quot;Catgirl riding a line, going up&quot; title=&quot;Catgirl riding a line, going up&quot; /&gt;&lt;/p&gt;
&lt;p&gt;So you have a Dream.&lt;/p&gt;
&lt;p&gt;Like all worthy Dreams, it may take years or decades to achieve. Alas, the human motivation system &lt;em&gt;can’t&lt;/em&gt; learn from rewards with a year-long delay.&lt;/p&gt;
&lt;p&gt;So, you&#39;re instead advised to “trust the process”, and convert your long-term goal into regular short-term micro-goals:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Write a novel ⇒ write 500 words a day.&lt;/li&gt;
&lt;li&gt;Cure your social phobia ⇒ talk with one stranger a day.&lt;/li&gt;
&lt;li&gt;Get married in 10 years ⇒ get 2,000 &lt;a href=&quot;https://colah.github.io/personal/micromarriages/&quot;&gt;micromarriages&lt;/a&gt; a week.&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;This &lt;em&gt;is&lt;/em&gt; great advice, and I endorse it, but… it feels incomplete? If every day I just focus on this day, I feel like I&#39;m in a hamster wheel, hitting my same micro-goal over and over again, but for what? Intellectually, I know &amp;quot;for what&amp;quot; — the Dream — but I don&#39;t &lt;em&gt;feel&lt;/em&gt; it anymore.&lt;/p&gt;
&lt;p&gt;So, here&#39;s a solution that reliably helps me &lt;em&gt;feel&lt;/em&gt; the connection between my past, present &amp;amp; future: &lt;strong&gt;lines on a graph!&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;Let&#39;s look at 3 examples: project management, money, and weight loss.&lt;/p&gt;
&lt;hr /&gt;
&lt;h2&gt;🚀 Example 1: Projects&lt;/h2&gt;
&lt;p&gt;The good news is I’m my own boss. The bad news is my boss is a moron.&lt;/p&gt;
&lt;p&gt;Specifically, my boss (me) is bad at project management. She (I) still is (am), to be honest. But the tool that&#39;s helped me the most is the &lt;a href=&quot;https://en.wikipedia.org/wiki/Burndown_chart&quot;&gt;Burndown Chart&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;How it works: First, give your project (or &amp;quot;sprint&amp;quot;) a deadline, and break it down into small tasks. Give each task a number of &amp;quot;points&amp;quot; it&#39;s worth (I just make points = hours). Then each day, do tasks. Your burndown chart software will then show you, in a single graph:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Your past: how many points/hours you&#39;ve gone through&lt;/li&gt;
&lt;li&gt;Your present: how many points/hours you need to do &lt;em&gt;today&lt;/em&gt; to stay on track&lt;/li&gt;
&lt;li&gt;Your future: when you&#39;ll finish your project!&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;( &lt;a href=&quot;https://commons.wikimedia.org/wiki/File:SampleBurndownChart.svg&quot;&gt;example chart by User:PabloStraub&lt;/a&gt; ⤵ )&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2026-01/burndown.png&quot; alt=&quot;Chart of ideal vs actual hours left of work to do, over time&quot; title=&quot;Chart of ideal vs actual hours left of work to do, over time&quot; /&gt;&lt;/p&gt;
&lt;p&gt;When I shipped my latest project, &lt;a href=&quot;https://aisafety.dance/p3/&quot;&gt;a 20,000-word-with-40-illustrations behemoth&lt;/a&gt;, I used a variant called a Burnup Chart, which also tracks total estimated time &lt;em&gt;over time&lt;/em&gt;, to see how badly I mis-estimated how long tasks would take. Here&#39;s my actual chart:&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2026-01/burnup.png&quot; alt=&quot;Chart of my ideal hours done, actual hours done, and estimated total hours, over time&quot; title=&quot;Chart of my ideal hours done, actual hours done, and estimated total hours, over time&quot; /&gt;&lt;/p&gt;
&lt;p&gt;( As you can see: I started off estimating that project would take ~25 hours, but by the end it turned out to be ~75h. I was off by a factor of three, yeesh. But thanks to this Burnup Chart design, &lt;em&gt;I could account for my own mis-accounting&lt;/em&gt;, and see where the blue &amp;amp; red trendlines would intersect, to get a better estimate of when I&#39;d actually finish my project. And I did, in fact, finish by Dec 1st 2025, as planned! )&lt;/p&gt;
&lt;p&gt;The best part about Burndown/Burnup charts? &lt;strong&gt;I can see how different decisions will affect the future, &lt;em&gt;today.&lt;/em&gt;&lt;/strong&gt; For example, I can try deleting a task, or taking an extra break day, or increasing hours worked a day, and &lt;em&gt;immediately&lt;/em&gt; see how that&#39;d affect when I&#39;ll finish my project.&lt;/p&gt;
&lt;p&gt;By tying my past-present-future together into lines on a graph, not only am I more motivated, I also get &lt;em&gt;instant feedback&lt;/em&gt; from the future.&lt;/p&gt;
&lt;p&gt;(&lt;a href=&quot;https://docs.google.com/spreadsheets/d/1AyFS1ufnYWcGjjC5KxvSrfb04f64UYwNChvqlR7cNwI/edit?usp=sharing&quot;&gt;Here&#39;s a blank template&lt;/a&gt;, and &lt;a href=&quot;https://docs.google.com/spreadsheets/d/11faTjGrXp_pUn-NzBG-ioYPew-IDMVS5RViqYPH_R8U/edit?usp=sharing&quot;&gt;here&#39;s an actual example of when I used it&lt;/a&gt;.)&lt;/p&gt;
&lt;h2&gt;💸 Example 2: Money&lt;/h2&gt;
&lt;p&gt;Another example of &amp;quot;see your effect on the future, today&amp;quot; — just enter your current income &amp;amp; spending into a spreadsheet, multiply by the compounding interest of your savings/investments, and voilà: (example data, not my real finances)&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2026-01/money1.png&quot; alt=&quot;Spreadsheet of income, expenses, and savings over time&quot; title=&quot;Spreadsheet of income, expenses, and savings over time&quot; /&gt;&lt;/p&gt;
&lt;p&gt;Again, this lets you play with decisions like &amp;quot;what if I tried to increase income by this much&amp;quot;, &amp;quot;what if I tried to reduce my rent by that much&amp;quot;, and you get &lt;em&gt;instant&lt;/em&gt; future-feedback.&lt;/p&gt;
&lt;p&gt;But more importantly, it helps you viscerally &lt;em&gt;see&lt;/em&gt; the connection between your short-term micro-goals, and long-term Dream. &amp;quot;Put $10 a day into savings&amp;quot; is doable but doesn&#39;t sound like much, but — just do the math — and you&#39;ll find it&#39;s enough to &lt;em&gt;retire a millionaire at age 63, already adjusted for inflation, starting from zero money!&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;(yes, for people living paycheck-to-paycheck or less, &amp;quot;put aside $10 a day&amp;quot; isn&#39;t feasible. but for most middle-class people and up, it should be.)&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2026-01/money2.png&quot; alt=&quot;Spreadsheet showing how one can become a millionaire by retirement age, just by saving $10 a day into a low-fee diversified index fund&quot; title=&quot;Spreadsheet showing how one can become a millionaire by retirement age, just by saving $10 a day into a low-fee diversified index fund&quot; /&gt;&lt;/p&gt;
&lt;p&gt;( Quick aside, &lt;a href=&quot;https://blog.ncase.me/30/#tip_3&quot;&gt;the best advice for almost all laypeople&lt;/a&gt;: ditch the fancy BS mutual funds, and forget about day-trading stocks. Just put your money in a very-low-fee diversified index fund. If that feels icky because you&#39;re investing in capitalism, remember, if you put your money in a savings account, your bank will &lt;em&gt;still&lt;/em&gt; invest it, they&#39;ll just give &lt;em&gt;you&lt;/em&gt; nothing but crumbs. Also: you can pick a socially-responsible index fund that excludes weaponry &amp;amp; fossil fuels. )&lt;/p&gt;
&lt;p&gt;( Here&#39;s &lt;a href=&quot;https://docs.google.com/spreadsheets/d/1-Bqc9K_WiBVUD5dhS0bGmwZsw07e4boe3DZB7bDU5tk/edit?usp=sharing&quot;&gt;the spreadsheets for both the above money examples&lt;/a&gt; )&lt;/p&gt;
&lt;h2&gt;🏋️‍♀️ Example 3: Weight&lt;/h2&gt;
&lt;p&gt;Finally, the most cliché New Year&#39;s Resolution: losing weight. Again, use a spreadsheet with lines on a graph: (my actual data)&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2026-01/weight.png&quot; alt=&quot;Graph of my weight over time, with trendline, and example of one day&#39;s data&quot; title=&quot;Graph of my weight over time, with trendline, and example of one day&#39;s data&quot; /&gt;&lt;/p&gt;
&lt;p&gt;You can see, at a glance, the connection between:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Your past: weight over time&lt;/li&gt;
&lt;li&gt;Your present: your target caloric deficit for &lt;em&gt;today&lt;/em&gt;&lt;/li&gt;
&lt;li&gt;Your future: when you&#39;re likely to hit your weight goal&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;( I find that visualizing the trendline is emotionally helpful, because weight is &lt;em&gt;so variable&lt;/em&gt; from day to day. Seeing the trendline helps reassure me, ignore the daily noise, I&#39;m on the right track. )&lt;/p&gt;
&lt;p&gt;And again, you can play with &lt;em&gt;what if&#39;s&lt;/em&gt;, and get instant future-feedback.
A daily micro-goal like &amp;quot;250 calorie deficit&amp;quot; (one less cookie a day) — do the math&lt;sup class=&quot;footnote-ref&quot;&gt;&lt;a href=&quot;https://blog.ncase.me/lines-on-a-graph/#fn1&quot; id=&quot;fnref1&quot;&gt;[1]&lt;/a&gt;&lt;/sup&gt; — becomes &amp;quot;lose 50lb (23kg) in two years&amp;quot; (enough to get you from moderately-Obese to &amp;quot;Healthy&amp;quot; BMI!)&lt;/p&gt;
&lt;p&gt;( &lt;strong&gt;IMPORTANT:&lt;/strong&gt; a) Yes both BMI and the Calorie-In-Calorie-Out model are &lt;em&gt;very&lt;/em&gt; incomplete and flawed, b) Being Overweight &amp;amp; Obese Class I (25 &amp;lt; BMI &amp;lt; 35) is probably fine health-wise; previous studies didn&#39;t control for exercise levels &amp;amp; cardiovascular fitness (&lt;a href=&quot;https://dip2023.s3.amazonaws.com/Lecture%204/Barry-2014.pdf&quot;&gt;see Figures 3 &amp;amp; 4&lt;/a&gt;), c) May I interest you in the &lt;a href=&quot;https://slimemoldtimemold.com/2022/07/12/lose-10-6-pounds-in-four-weeks-with-this-one-weird-trick-discovered-by-local-slime-hive-mind-doctors-grudgingly-respect-them-hope-to-become-friends/&quot;&gt;Potato Diet&lt;/a&gt;? Potatoes fill you up very fast (high protein &amp;amp; high fiber if you keep the skins), have almost all the daily nutrients you need, and are versatile foods you can cook in many ways to keep the diet interesting. I did the Potato Diet for a month and lost 10 pounds, and I was &lt;em&gt;already&lt;/em&gt; &amp;quot;Normal&amp;quot; BMI! d) Or, ask your doctor about semaglutide. (&lt;a href=&quot;https://www.nejm.org/doi/full/10.1056/NEJMoa2032183&quot;&gt;see Figure 1&lt;/a&gt;) )&lt;/p&gt;
&lt;p&gt;&lt;a href=&quot;https://docs.google.com/spreadsheets/d/1N9tBR_qn7J4kBPcGkeG-S_kf6mHUlPoJs44unSS_msA/edit?usp=sharing&quot;&gt;Here&#39;s my weight-tracking Google Sheets template!&lt;/a&gt;&lt;/p&gt;
&lt;p&gt;Anyway, let’s wrap up:&lt;/p&gt;
&lt;hr /&gt;
&lt;h2&gt;🚀 Get started with lines today!&lt;/h2&gt;
&lt;blockquote&gt;
&lt;p&gt;&amp;quot;Huh, what an interesting &amp;amp; inspiring article! I&#39;ll try this later&amp;quot;&lt;br /&gt;
(one year later)&lt;br /&gt;
&amp;quot;Oh no I forgot.&amp;quot;&lt;/p&gt;
&lt;p&gt;~ you, if you don&#39;t implement this &lt;em&gt;right now&lt;/em&gt;.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;So before you close this tab, at least ask yourself:&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;1) What&#39;s a Dream you have, that you&#39;ve wanted for a long while, yet haven&#39;t done it yet?&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;No shame here, it can be as wacky or tacky as your heart desires. For example:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;📕 Write the next Great Canadian Novel&lt;/li&gt;
&lt;li&gt;🧠 Master the secrets of the universe&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;No, seriously, stop &amp;amp; think of a Dream you have. Silently think it and/or write it down, before continuing.&lt;/p&gt;
&lt;p&gt;Got one?&lt;/p&gt;
&lt;p&gt;Good. Next step:&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;2) How can you convert your long-term Dream into a regular micro-goal?&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;What makes a good micro-goal:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;You can do it regularly &amp;amp; often. (e.g. daily, weekly)&lt;/li&gt;
&lt;li&gt;There&#39;s a clear, specific pass/fail. (e.g. &amp;quot;write 500 words&amp;quot; vs &amp;quot;do some writing&amp;quot;)&lt;/li&gt;
&lt;li&gt;It &lt;em&gt;causes, not just correlates with&lt;/em&gt;, your long-term goal. (to mitigate &lt;a href=&quot;https://en.wikipedia.org/wiki/Goodhart%27s_law&quot;&gt;Goodhart’s Law&lt;/a&gt;, where “what gets measured gets gamed”)&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;Running with the above examples:&lt;/p&gt;
&lt;p&gt;📕 &amp;quot;Write a Great novel&amp;quot; ⇒ &lt;strong&gt;Write/edit 500 words a day.&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;500 words a day is easy. You probably write that much alone in text messages.&lt;/li&gt;
&lt;li&gt;A novel is around 50,000 words.&lt;sup class=&quot;footnote-ref&quot;&gt;&lt;a href=&quot;https://blog.ncase.me/lines-on-a-graph/#fn2&quot; id=&quot;fnref2&quot;&gt;[2]&lt;/a&gt;&lt;/sup&gt; At 500 words a day, you can finish the first draft in 100 days, by mid-April. Then, &lt;em&gt;editing&lt;/em&gt; 500 words a day, you can create a polished draft by mid-July. Then, editing 500 words a day &lt;em&gt;again&lt;/em&gt;, you can create a good final version by mid-October.&lt;/li&gt;
&lt;li&gt;So: by writing/editing a mere 500 words a day, you can &lt;em&gt;publish a novel every year&lt;/em&gt;, with two months to spare!
&lt;ul&gt;
&lt;li&gt;(The exact numbers vary if you want to write longer books, or do more revisions, but the general result holds: you &lt;em&gt;could&lt;/em&gt; complete a polished book every 1 or 2 years, by writing/editing only 500 to 1000 words a day.)&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;Your first novel is very unlikely to be Great, but with more practice &amp;amp; feedback, you maximize your chances of writing a truly Great work. Persist! Remember that Slaughterhouse-Five was Vonnegut&#39;s &lt;em&gt;sixth&lt;/em&gt; novel! (As for GenAI? The people will be thirsty for quality, amidst the coming flood of AI slop.)&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;🧠 &amp;quot;Master &lt;s&gt;the secrets of the universe&lt;/s&gt; Physics&amp;quot; ⇒ &lt;strong&gt;Study 1 hour a day.&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;1 hour a day is easy-ish. You probably waste that much time a day on Internet Attention Sinks. (I know I do)&lt;/li&gt;
&lt;li&gt;Here&#39;s a study plan, at 1 hour a day:
&lt;ul&gt;
&lt;li&gt;Day 1, Day 2: Watch one of &lt;a href=&quot;https://theoreticalminimum.com/courses&quot;&gt;Leonard Susskind&#39;s free online Stanford University lectures&lt;/a&gt;. Each lecture is around 2 hours, so watch half today, half tomorrow. Write down your notes &amp;amp; questions.&lt;/li&gt;
&lt;li&gt;Day 3, 4, 5: Ask &lt;a href=&quot;https://claude.ai/&quot;&gt;Claude&lt;/a&gt; or &lt;a href=&quot;https://gemini.google.com/&quot;&gt;Gemini&lt;/a&gt; your questions, to clarify your understanding. (I don&#39;t trust ChatGPT or Grok.) What about LLM hallucinations? As of Jan 1st 2026, I find that large language models are pretty good on educational topics, &lt;em&gt;as long as they&#39;re undergraduate-level or below&lt;/em&gt;.&lt;/li&gt;
&lt;li&gt;Day 6, 7, 8: Web search for &amp;quot;[topic] practice problems with answers&amp;quot;, then &lt;em&gt;actually do them&lt;/em&gt;. If you get stuck, ask your chatbot of choice for help &lt;em&gt;but not full solutions&lt;/em&gt;. A gym coach shouldn&#39;t lift the weights &lt;em&gt;for&lt;/em&gt; you.&lt;/li&gt;
&lt;li&gt;Day 9, 10: Make &lt;a href=&quot;https://apps.ankiweb.net/&quot;&gt;Anki flashcards&lt;/a&gt; to remember the core ideas &amp;amp; formulas, easy-ish-ly and effectively. (&lt;a href=&quot;https://www.youtube.com/watch?v=eVajQPuRmk8&quot;&gt;8 min YouTube video&lt;/a&gt; on how &amp;amp; why spaced repetition works)&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;So that&#39;s 10 days = 2 weeks with weekends off, per Susskind lecture. There&#39;s 10 lectures per course × 6 courses = 60 lectures. At 2 weeks per lecture × 60 lectures = 120 weeks, or 2 years.&lt;/li&gt;
&lt;li&gt;So: you can get &lt;em&gt;a college-level understanding of Newtonian physics, Electrodynamics, Quantum mechanics, Special &amp;amp; General relativity, Cosmology, and Entropy&lt;/em&gt;... at the cost of &lt;em&gt;1 hour per weekday&lt;/em&gt; (and $20/month for a chatbot)… spread across 2 years, which is &lt;em&gt;half&lt;/em&gt; the time of a 4-year college degree.&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;( It is not exactly a well-kept secret that the modern university system is horribly broken. &lt;a href=&quot;https://www.smbc-comics.com/comic/college-level-mathematics&quot;&gt;Relevant SMBC&lt;/a&gt;: )&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2026-01/smbc_university.png&quot; alt=&quot;Comic showing how a university degree costs as much as hiring 3 PhDs to tutor you full time + buying several baby tigers&quot; title=&quot;Comic showing how a university degree costs as much as hiring 3 PhDs to tutor you full time + buying several baby tigers&quot; /&gt;&lt;/p&gt;
&lt;p&gt;( If you don&#39;t have the pre-requisite math skills for Susskind&#39;s lectures, add an extra five months of &amp;quot;study 1 hour a day&amp;quot;, and use Khan Academy for &lt;a href=&quot;https://www.youtube.com/playlist?list=PL7AF1C14AF1B05894&quot;&gt;high-school algebra&lt;/a&gt; (~12 hours of lectures), and 3Blue1Brown for &lt;a href=&quot;https://www.youtube.com/playlist?list=PLZHQObOWTQDMsr9K-rj53DwVRMYO3t5Yr&quot;&gt;calculus&lt;/a&gt; (~3 hours) and &lt;a href=&quot;https://www.youtube.com/playlist?list=PLZHQObOWTQDPD3MizzM2xVFitgF8hE_ab&quot;&gt;linear algebra&lt;/a&gt; (~3 hours). So that&#39;s an extra 18h of lectures, and given our above study plan where we tackle 2h of lectures per 2 weeks, that&#39;s an extra 18 weeks, or about five months, of self-study. Again, at just 1 hour a day. What a steal of a deal! )&lt;/p&gt;
&lt;p&gt;Okay, back on track:&lt;/p&gt;
&lt;p&gt;Step 1, know your long-term Dream. Step 2, break it down into a sustainable regular micro-goal. This step &lt;em&gt;is&lt;/em&gt; tricky; if you&#39;re stuck, paste a link to this blog post into Claude (with web search turned on), tell &#39;em your Dream, and ask Claude to help you break down your Dream into sustainable micro-goals, with inspiration from the above two examples.&lt;/p&gt;
&lt;p&gt;After that, your final step:&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;3) DRAW LINES ON A GRAPH&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;I strongly recommend using a spreadsheet program like Google Sheets, Excel, or Numbers. It needs more setup than an app, but you get full freedom to customize it however you like! If you need help setting up a spreadsheet, look up YouTube or consult Claude. Again, you can also get started with my templates:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;🚀 Project Burnup: &lt;a href=&quot;https://docs.google.com/spreadsheets/d/1AyFS1ufnYWcGjjC5KxvSrfb04f64UYwNChvqlR7cNwI/edit?usp=sharing&quot;&gt;template&lt;/a&gt;, &lt;a href=&quot;https://docs.google.com/spreadsheets/d/11faTjGrXp_pUn-NzBG-ioYPew-IDMVS5RViqYPH_R8U/edit?usp=sharing&quot;&gt;example with my actual data&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;💸 Money: &lt;a href=&quot;https://docs.google.com/spreadsheets/d/1-Bqc9K_WiBVUD5dhS0bGmwZsw07e4boe3DZB7bDU5tk/edit?usp=sharing&quot;&gt;templates with examples&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;🏋️‍♀️ Calorie Tracker: &lt;a href=&quot;https://docs.google.com/spreadsheets/d/1N9tBR_qn7J4kBPcGkeG-S_kf6mHUlPoJs44unSS_msA/edit?usp=sharing&quot;&gt;template&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;Alternatively, you can use these apps:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;🃏 &lt;a href=&quot;https://trello.com/&quot;&gt;Trello&lt;/a&gt; for productivity (with &lt;a href=&quot;https://trello.com/power-ups/search?q=burndown&quot;&gt;a Burndown Chart add-on&lt;/a&gt;)&lt;/li&gt;
&lt;li&gt;🐝 &lt;a href=&quot;https://www.beeminder.com/&quot;&gt;Beeminder&lt;/a&gt; for any quantifiable goal. (this app helps you set up &amp;quot;commitments with a sting&amp;quot;! every time you slip behind your goal, Beeminder will take some of your money.)&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;(If you don&#39;t want to do this final step today, &lt;em&gt;set an exact day &amp;amp; hour in your Calendar to do this&lt;/em&gt;, so you don&#39;t forget to do it!)&lt;/p&gt;
&lt;p&gt;. . .&lt;/p&gt;
&lt;p&gt;And that&#39;s all the steps!&lt;/p&gt;
&lt;p&gt;Try it out for one area of your life, then slowly expand to others, one Dream at a time.&lt;/p&gt;
&lt;hr /&gt;
&lt;h2&gt;🤔 Q &amp;amp; A&lt;/h2&gt;
&lt;p&gt;&lt;strong&gt;Q:&lt;/strong&gt; If it really &lt;em&gt;is&lt;/em&gt; that trivially easy to write one book a year, become a millionaire by middle age, go from Obese to &amp;quot;Healthy&amp;quot; BMI, and gain university-level mastery of hard skills &amp;amp; topics… why isn&#39;t everybody already doing this? Heck, why aren&#39;t &lt;em&gt;you&lt;/em&gt; doing this, Nicky. Why aren&#39;t &lt;em&gt;you&lt;/em&gt; already a millionaire polymath scientist-artist supermodel bestselling author?&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;A:&lt;/strong&gt; lol i dunno. i guess everyone&#39;s only living at 2% of their true potential&lt;/p&gt;
&lt;p&gt;( in 2026, i resolve to live up to &lt;em&gt;5%&lt;/em&gt; of my true potential! )&lt;/p&gt;
&lt;p&gt;( edit: ok, actual answer to this paradox, I think, is something like &amp;quot;&lt;a href=&quot;https://en.wikipedia.org/wiki/Hyperbolic_discounting&quot;&gt;hyperbolic discounting&lt;/a&gt; sucks&amp;quot;. Less jargon-y: human brains tend to overweigh the short-term over long-term, because the short-term is more immediately visible. My hope is the &amp;quot;lines on a graph&amp;quot; trick helps work around this brain-glitch, by making the long-term more vivid and &lt;em&gt;here&lt;/em&gt;. )&lt;/p&gt;
&lt;hr /&gt;
&lt;p&gt;“Write a novel a year” sounds impossible. Most bestselling authors aren’t that prolific. “Write 500 words a day” sounds trivial. You do that unthinkingly between texts &amp;amp; DMs &amp;amp; Slacks.&lt;/p&gt;
&lt;p&gt;Yet &lt;em&gt;they’re the same goal&lt;/em&gt;, just on different time-scales.&lt;/p&gt;
&lt;p&gt;That’s the magic of micro-goals and lines-on-a-graph, seeing your past + present + future at a single glance: you can connect the impossible Dream with the possible Daily, see your effect on the future &lt;em&gt;today&lt;/em&gt;, and &lt;em&gt;browse the multiverse of possible futures&lt;/em&gt; for the one your heart desires.&lt;/p&gt;
&lt;p&gt;📈 Happy Line-making, everyone!&lt;br /&gt;
~ Nicky Case&lt;/p&gt;
&lt;hr /&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2026-01/nyee.png&quot; alt=&quot;illustration of a catgirl collapsed in exhaustion, in front of a giant party-ful &#39;2026&#39; sign&quot; title=&quot;illustration of a catgirl collapsed in exhaustion, in front of a giant party-ful &#39;2026&#39; sign&quot; /&gt;&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;P.S:&lt;/strong&gt; One of my Dreams this year is to write &lt;em&gt;one blog post a week.&lt;/em&gt; I hope you found value in this first entry of my &amp;quot;one post a week&amp;quot; challenge for 2026!&lt;/p&gt;
&lt;p&gt;Let&#39;s do the math: I expect my posts to be 2,000 words on average. (this one&#39;s 2,600 words long.) So if I write &amp;amp; polish 500 words a day, I can get 1 post done per 4 weekdays, giving me a good buffer. 2,000 words per week × 52 weeks per year = 104,000 words. That&#39;s the size of &lt;em&gt;two books&lt;/em&gt;.&lt;/p&gt;
&lt;p&gt;If you&#39;d like to keep up on my &amp;quot;books disguised a blog&amp;quot;, sign up for my low-volume, monthly newsletter! Every end of month, I&#39;ll share links to the 4 or 5 posts I wrote — this way, you can skip the ones that don&#39;t interest you:&lt;/p&gt;
&lt;iframe src=&quot;https://ncase.me/ncase-credits/signup2.html&quot; frameborder=&quot;no&quot; width=&quot;640&quot; height=&quot;250&quot;&gt;&lt;/iframe&gt;
&lt;p&gt;I also reflected on 2025 and wrote more on my resolutions for 2026, &lt;a href=&quot;https://www.patreon.com/posts/h-n-y-2026-from-147107425&quot;&gt;in this public Patreon post&lt;/a&gt;. You can support my research &amp;amp; writings (always free, online, and open-source!) via &lt;a href=&quot;https://www.patreon.com/ncase&quot;&gt;Patreon&lt;/a&gt;, &lt;a href=&quot;https://ko-fi.com/nickycase&quot;&gt;Ko-Fi&lt;/a&gt;, or &lt;a href=&quot;https://www.paypal.com/paypalme/nickycase&quot;&gt;PayPal&lt;/a&gt; ⤵&lt;/p&gt;
&lt;iframe src=&quot;https://ncase.me/ncase-credits/supporters/jan-2026.html&quot; frameborder=&quot;no&quot; width=&quot;640&quot; height=&quot;640&quot;&gt;&lt;/iframe&gt;
&lt;p&gt;Bless your cotton socks!  Happy new year 2026!  🎉 🎊 🎉&lt;/p&gt;
&lt;hr class=&quot;footnotes-sep&quot; /&gt;
&lt;section class=&quot;footnotes&quot;&gt;
&lt;ol class=&quot;footnotes-list&quot;&gt;
&lt;li id=&quot;fn1&quot; class=&quot;footnote-item&quot;&gt;&lt;p&gt;A simple (but inaccurate) rule-of-thumb is Wishnofsky&#39;s Rule: 1 pound of fat ~= 3500 kcal. (&lt;a href=&quot;https://pmc.ncbi.nlm.nih.gov/articles/PMC4035446/&quot;&gt;This rule is inaccurate &amp;amp; there are better models, but honestly it&#39;s not &lt;em&gt;too&lt;/em&gt; far off.&lt;/a&gt;) In other words, with a calorie deficit of 250 kcal a day, you&#39;ll have a deficit of 3500 kcal in 14 days, or burning 1 pound of fat per 2 weeks. So, over 2 years, that&#39;s ~104 weeks, or burning 104/2 = 52 pounds (23kg). Given &lt;a href=&quot;https://commons.wikimedia.org/wiki/File:BMI_chart.png&quot;&gt;the standard BMI model &lt;/a&gt;(which, yes, is also inaccurate but let&#39;s roll with this for now), 52lb/23kg is enough to take you from moderately Obese to &amp;quot;Healthy&amp;quot;. &lt;a href=&quot;https://blog.ncase.me/lines-on-a-graph/#fnref1&quot; class=&quot;footnote-backref&quot;&gt;↩︎&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li id=&quot;fn2&quot; class=&quot;footnote-item&quot;&gt;&lt;p&gt;Random web sites claim that most published books &amp;quot;should&amp;quot; be around 80,000 to 100,000 words, with Young Adult having a smaller word count, and 40,000 at minimum to count as a novel, not &amp;quot;novella&amp;quot;. I&#39;m picking 50,000 to make the math easy to follow.&lt;/p&gt;
&lt;p&gt;Some Great novels around 50,000 words: Kurt Vonnegut&#39;s &lt;em&gt;Slaughterhouse-Five&lt;/em&gt;, Douglas Adams&#39;s &lt;em&gt;The Hitchhiker’s Guide to the Galaxy&lt;/em&gt;, F. Scott Fitzgerald&#39;s &lt;em&gt;The Great Gatsby&lt;/em&gt;. 50k words is plenty! &lt;a href=&quot;https://blog.ncase.me/lines-on-a-graph/#fnref2&quot; class=&quot;footnote-backref&quot;&gt;↩︎&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ol&gt;
&lt;/section&gt;
</content>
        </entry>

    

        
        
        <entry>
            <title>Signal Boosts for Autumn 2025: much ado about the Algorithm</title>
            <link href="https://blog.ncase.me/signal-boost-autumn-2025/"/>
            <updated>2025-10-30T00:00:00Z</updated>
            <id>https://blog.ncase.me/signal-boost-autumn-2025/</id>
            <content xml:lang="en" type="html">&lt;p&gt;Hey folks! Things are still in the works. In the meantime, here&#39;s some recent-ish stuff I found valuable, and I hope you do too. 💖 (Sorry this post is, uh, over 10,000 words. Guess I&#39;m making up for not posting a Signal Boost in almost a year!)&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;Much ado about The Algorithm&lt;/strong&gt;
&lt;ul&gt;
&lt;li&gt;🎙️ An interview series with experts on this new algorithmic era &lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#interview&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;😑 New study: It&#39;s probably not the algorithm&#39;s fault, we just suck &lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#we-suck&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;🙃 Gradual Disempowerment: how even “dumb” AI could take over humanity &lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#gradual&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;🤝 Bridging-Based Ranking: reverse political polarization by reversing the algorithm &lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#bridging&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;🧐 Zero-Knowledge IDs: How to prove yourself without doxxing yourself &lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#zkp&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;🧠 Cyborgism: How to make AI that enhances us, not replaces us &lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#cyborg&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;🍻 The 6Pack of Digital Democracy &lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#6pack&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Fun Stuff!&lt;/strong&gt;
&lt;ul&gt;
&lt;li&gt;🕵️ Clues by Sam: a daily deductive detective game &lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#clues&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;🐦 Snakebird: cute birbs, cruel puzzles &lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#snakebird&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;🦉 ZeWei&#39;s Multiverse Tour Guide Adventures Continue &lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#zewei&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;🤓 Emnerson: goofy internet gal &lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#emnerson&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;🧇 &amp;quot;blahaj goes to waffle house at 3am&amp;quot; &lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#blahaj&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;SPOOKS FOR HALLOWEEN&lt;/strong&gt;
&lt;ul&gt;
&lt;li&gt;👻 BOO! a silly music video &amp;amp; music album &lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#boo&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;🩸 Bury Your Gays by Chuck Tingle &lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#bury-your-gays&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;hr /&gt;
&lt;h2&gt;Much ado about The Algorithm&lt;/h2&gt;
&lt;p&gt;&lt;a id=&quot;interview&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h3&gt;🎙️ Interview series with experts on this new algorithmic era&lt;/h3&gt;
&lt;p&gt;My friend Tobias Rose-Stockwell — author of the book &lt;a href=&quot;https://www.outragemachine.org/&quot;&gt;Outrage Machine&lt;/a&gt; — just launched a new interview series!  In this series, he&#39;ll yap with experts on the simple question of &amp;quot;seriously wtf is going on, what do we do when algorithms install depression &amp;amp; extremism into our kids, scientists can&#39;t agree if AI is hype of the end-of-humanity, also holy shit politics??&amp;quot;&lt;/p&gt;
&lt;p&gt;Simple question!&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Episode 1: Jonathan Haidt&lt;/strong&gt;, psychologist and author of the bestselling books &lt;a href=&quot;https://en.wikipedia.org/wiki/The_Righteous_Mind&quot;&gt;The Righteous Mind&lt;/a&gt;&lt;sup class=&quot;footnote-ref&quot;&gt;&lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#fn1&quot; id=&quot;fnref1&quot;&gt;[1]&lt;/a&gt;&lt;/sup&gt; and &lt;a href=&quot;https://en.wikipedia.org/wiki/The_Anxious_Generation&quot;&gt;The Anxious Generation&lt;/a&gt;.&lt;/p&gt;
&lt;iframe width=&quot;640&quot; height=&quot;360&quot; src=&quot;https://www.youtube-nocookie.com/embed/Dh4g0gxirSQ?rel=0&quot; title=&quot;YouTube video player&quot; frameborder=&quot;0&quot; allow=&quot;accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share&quot; referrerpolicy=&quot;strict-origin-when-cross-origin&quot; allowfullscreen=&quot;&quot;&gt;&lt;/iframe&gt;
&lt;p&gt;&lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#interview-with-jon&quot;&gt;: click to expand — my notes from the Jon episode&lt;/a&gt;&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Episode 2: Tim Urban&lt;/strong&gt;, blogger behind &lt;a href=&quot;https://waitbutwhy.com/&quot;&gt;Wait But Why&lt;/a&gt; and author of &lt;a href=&quot;https://waitbutwhy.com/whatsourproblem&quot;&gt;What&#39;s Our Problem?&lt;/a&gt;&lt;/p&gt;
&lt;iframe width=&quot;640&quot; height=&quot;360&quot; src=&quot;https://www.youtube-nocookie.com/embed/eGBhUXmxSbI?rel=0&quot; title=&quot;YouTube video player&quot; frameborder=&quot;0&quot; allow=&quot;accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share&quot; referrerpolicy=&quot;strict-origin-when-cross-origin&quot; allowfullscreen=&quot;&quot;&gt;&lt;/iframe&gt;
&lt;p&gt;&lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#interview-with-tim&quot;&gt;: click to expand — my notes from the Tim episode&lt;/a&gt;&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Upcoming interviews:&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;Tristan Harris&lt;/strong&gt;, ex-Google employee most famous for calling out Google, Apple &amp;amp; Facebook for using dark design patterns to keep us (literally, not just metaphorically) addicted — and promoting specific better designs that respect human attention &amp;amp; autonomy. (See: &lt;a href=&quot;https://www.slideshare.net/slideshow/google-deck-on-digital-wellbeing-a-call-to-minimize-distraction-and-respect-users-attention/109671786&quot;&gt;Tristan&#39;s leaked internal Google presentation (2013)&lt;/a&gt;.)&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Steven Pinker&lt;/strong&gt;, who I do not like because he keeps strawmanning/weakmanning all his opponents, and IMO is wrong on cognitive science (his own field), and most of the other fields he&#39;s written on, especially AI Safety. But he &lt;em&gt;is&lt;/em&gt; a leading voice &amp;amp; bestselling author, so, well, whatever.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Esther Perel&lt;/strong&gt;, psychotherapist best known for her talks &amp;amp; books about &lt;em&gt;SEX!&lt;/em&gt; and relationships.&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;Anyway if the above interests you, check out my friend&#39;s new interview series!&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;On &lt;a href=&quot;https://www.youtube.com/@IntoTheMachineShow&quot;&gt;YouTube&lt;/a&gt; / &lt;a href=&quot;https://tobias.substack.com/&quot;&gt;Substack&lt;/a&gt; / &lt;a href=&quot;https://podcasts.apple.com/ca/podcast/into-the-machine-with-tobias-rose-stockwell/id1824137015&quot;&gt;Apple Podcasts&lt;/a&gt;&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;Be one of the first thousand fans, it&#39;s &lt;em&gt;very&lt;/em&gt; early days for this project.&lt;/p&gt;
&lt;p&gt;&lt;a id=&quot;we-suck&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h3&gt;😑 New study: It&#39;s probably not the algorithm&#39;s fault, we just suck&lt;/h3&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2025-10/its_you.png&quot; alt=&quot;Screenshot from Bojack Horseman: Todd simply tells Bojack, It&#39;s you.&quot; title=&quot;Screenshot from Bojack Horseman: Todd simply tells Bojack, _&amp;quot;It&#39;s you.&amp;quot;_&quot; /&gt;&lt;/p&gt;
&lt;p&gt;Tobias sent me &lt;a href=&quot;https://arxiv.org/abs/2508.03385&quot;&gt;this paper&lt;/a&gt; (&lt;a href=&quot;https://arstechnica.com/science/2025/08/study-social-media-probably-cant-be-fixed/&quot;&gt;Ars Technica lay-summary&lt;/a&gt;), which hurts him because it contradicts one of his core theses. +1 for Tobias&#39;s intellectual honesty, showing the counter-evidence.&lt;/p&gt;
&lt;p&gt;Anyway, Tobias&#39;s thesis (a common one among researchers) is that social media breeds extremism because of their algorithmic feeds. &lt;em&gt;Enraging is engaging&lt;/em&gt;. The data shows it: anger is &lt;em&gt;the&lt;/em&gt; best way to make a headline go viral.&lt;sup class=&quot;footnote-ref&quot;&gt;&lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#fn2&quot; id=&quot;fnref2&quot;&gt;[2]&lt;/a&gt;&lt;/sup&gt; The algorithms, trained to predict us, train &lt;em&gt;us&lt;/em&gt; to become &lt;em&gt;predictable&lt;/em&gt;.&lt;/p&gt;
&lt;p&gt;But this &lt;em&gt;new&lt;/em&gt; paper suggests that, no, it really is just our fault, no algorithm needed. The paper uses an interesting method: they simulate a social network, where each person&lt;sup class=&quot;footnote-ref&quot;&gt;&lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#fn3&quot; id=&quot;fnref3&quot;&gt;[3]&lt;/a&gt;&lt;/sup&gt; is simulated using LLMs&lt;sup class=&quot;footnote-ref&quot;&gt;&lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#fn4&quot; id=&quot;fnref4&quot;&gt;[4]&lt;/a&gt;&lt;/sup&gt;, deciding what to re-share and who to follow, etc. &amp;quot;Agent-based modelling&amp;quot; for social science isn&#39;t new, but using LLMs to do it is!&lt;/p&gt;
&lt;p&gt;(But you should be rightfully skeptical: &lt;em&gt;can&lt;/em&gt; LLMs simulate people accurately? &lt;a href=&quot;https://arxiv.org/abs/2411.10109&quot;&gt;This Stanford preprint&lt;/a&gt; finds that LLMs can replicate 1,000 real people&#39;s answers 85% as accurately &lt;em&gt;as the people themselves 2 weeks later&lt;/em&gt;. So, not perfect, but not bad!)&lt;/p&gt;
&lt;p&gt;Back to the paper: they use LLM agents to simulate a &amp;quot;baseline&amp;quot; social network&lt;sup class=&quot;footnote-ref&quot;&gt;&lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#fn5&quot; id=&quot;fnref5&quot;&gt;[5]&lt;/a&gt;&lt;/sup&gt;, then simulate proposed solutions to social media — e.g. chronological feeds, hide engagement metrics, bridging-based algorithms, etc — and measure they&#39;re effect on political polarization, attention-economy inequality, etc.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;The headline result: it&#39;s our fault.&lt;/strong&gt; Even a no-algorithm, chronological feed leads to insular echo chambers &amp;amp; polarization.&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2025-10/paper-1.png&quot; alt=&quot;chart from paper, see below paragraph for explanation&quot; title=&quot;chart from paper, see below paragraph for explanation&quot; /&gt;&lt;/p&gt;
&lt;p&gt;Figure 3: The &amp;quot;E-I Index&amp;quot; is a measure of outgroup (external) to ingroup (internal) follows. 0 (upwards) is less polarization, -1 (downwards) means more polarization. &lt;em&gt;No intervention reduces polarization that much.&lt;/em&gt; Chronological&#39;s practically the same as baseline. The best intervention is Bridging, but even then not by much.&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2025-10/paper-2.png&quot; alt=&quot;chart from paper, see below paragraph for explanation&quot; title=&quot;chart from paper, see below paragraph for explanation&quot; /&gt;&lt;/p&gt;
&lt;p&gt;Figure 4: Correlation between how extremist a poster is, and how many followers/reposts they receive. Chronological feeds are &lt;em&gt;much worse&lt;/em&gt; than baseline. Bridging is quite a bit better, actually. (Wait, how can a no-algorithm feed be &lt;em&gt;worse?&lt;/em&gt; My guess in footnote:&lt;sup class=&quot;footnote-ref&quot;&gt;&lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#fn6&quot; id=&quot;fnref6&quot;&gt;[6]&lt;/a&gt;&lt;/sup&gt;)&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2025-10/paper-3.png&quot; alt=&quot;chart from paper, see below paragraph for explanation&quot; title=&quot;chart from paper, see below paragraph for explanation&quot; /&gt;&lt;/p&gt;
&lt;p&gt;Figure 5: The &amp;quot;Gini coefficient&amp;quot; is a measure of inequality. Chronological feeds lead to the least amount of inequality in followers/reposts. Bridging leads to the most, but not that much higher than baseline.&lt;/p&gt;
&lt;p&gt;= = =&lt;/p&gt;
&lt;p&gt;On one hand, these results aren&#39;t &lt;em&gt;too&lt;/em&gt; surprising. Puritan-era Salem didn&#39;t have algorithms or mass media, yet they had literal witch hunts. Tumblr &amp;amp; 4chan in the early 2010&#39;s didn&#39;t have algorithmic feeds – (4chan doesn&#39;t even have a &amp;quot;reblogs&amp;quot; or &amp;quot;followers&amp;quot;) – yet not only were these sites hotbeds of extremism, it &lt;em&gt;escaped containment&lt;/em&gt; and affected US politics, which affected world politics. And today, Bluesky has a chronological feed by default&lt;sup class=&quot;footnote-ref&quot;&gt;&lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#fn7&quot; id=&quot;fnref7&quot;&gt;[7]&lt;/a&gt;&lt;/sup&gt;, yet it&#39;s... well. It&#39;s light-blue Tumblr. Donald Trump&#39;s Truth Social also has a chronological feed.&lt;sup class=&quot;footnote-ref&quot;&gt;&lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#fn8&quot; id=&quot;fnref8&quot;&gt;[8]&lt;/a&gt;&lt;/sup&gt;&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;So: no algorithm is needed for poisonous politics. &amp;quot;It&#39;s you.&amp;quot;&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;Well, damn. I was hoping there&#39;d be an easy technical fix, and the problem wasn&#39;t, &amp;quot;human nature is fundamentally corrupt&amp;quot;.&lt;/p&gt;
&lt;p&gt;On the other hand, the paper &lt;em&gt;did&lt;/em&gt; find a hopeful result – even though their own Abstract downplays it because, I don&#39;t know, academic humility and/or pessimism sells? They found that &lt;strong&gt;&amp;quot;bridging-based algorithms&amp;quot;&lt;/strong&gt;, like &lt;a href=&quot;https://asteriskmag.com/issues/08/the-making-of-community-notes&quot;&gt;Birdwatch/Community Notes&lt;/a&gt; or &lt;a href=&quot;https://en.wikipedia.org/wiki/Pol.is&quot;&gt;Pol.is&lt;/a&gt;, &lt;em&gt;does&lt;/em&gt; cut down polarization! (as explained in the above Figures)&lt;/p&gt;
&lt;p&gt;Bridging algorithms, instead of showing you what {people you agree with} agree on, show you what {people who disagree} &lt;em&gt;still&lt;/em&gt; agree on. It automatically finds &amp;amp; highlights common ground, by design! (More explanation later in this Signal Boost. &lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#bridging&quot;&gt;↪&lt;/a&gt;)&lt;/p&gt;
&lt;p&gt;Now, there &lt;em&gt;is&lt;/em&gt; a small cost to using bridging algorithms, in that it &lt;em&gt;slightly&lt;/em&gt; increases follower-inequality, but honestly, it&#39;s a &lt;em&gt;very&lt;/em&gt; small difference, and you could just offset that by making your algorithm give a bit more weight to smaller creators.&lt;/p&gt;
&lt;p&gt;Let&#39;s leave on that hopeful note.&lt;/p&gt;
&lt;p&gt;= = =&lt;/p&gt;
&lt;p&gt;Other caveats I want to mention:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Though LLM agents have been validated against human data (85% as accurate as people themselves 2 weeks later), because this simulation runs LLM agents against each other for dozens of rounds, small biases in LLMs can accumulate. (see: &lt;a href=&quot;https://www.astralcodexten.com/p/the-claude-bliss-attractor&quot;&gt;Scott Alexander&#39;s article on why Claude Finds God and DALL-E Goes Racist&lt;/a&gt;)&lt;/li&gt;
&lt;li&gt;The study might being over-estimating polarization, because the LLM &amp;quot;knows&amp;quot; polarization happens on social media, and thus acts it out. (it&#39;s a text &lt;em&gt;predictor&lt;/em&gt;, remember?) Then again, the study could be &lt;em&gt;under&lt;/em&gt;-estimating polarization, because these LLMs have been fine-tuned to be &amp;quot;friendly&amp;quot;, and thus will be kinder &amp;amp; more curious than the average internet user. Maybe these two errors roughly cancel out.&lt;/li&gt;
&lt;li&gt;Although the &lt;em&gt;algorithmic feed&lt;/em&gt; may not be the culprit, other design decisions could have led to higher assholery. For example, the low-friction Retweet lets impulsive emotions win the stage. And the Quote-Retweet (and Tumblr&#39;s Reblog-with-Commentary) is the perfect way to dunk on someone and get the &amp;quot;last&amp;quot; word.&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#notes-on-polarization&quot;&gt;: click to expand — why polarization isn&#39;t always bad, and the 3 kinds of &amp;quot;polarization&amp;quot;&lt;/a&gt;&lt;/p&gt;
&lt;p&gt;= = =&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Link to &lt;a href=&quot;https://arxiv.org/abs/2508.03385&quot;&gt;paper by Larooij &amp;amp; Törnberg&lt;/a&gt;, &lt;a href=&quot;https://arstechnica.com/science/2025/08/study-social-media-probably-cant-be-fixed/&quot;&gt;press article&lt;/a&gt;&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;&lt;a id=&quot;gradual&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h3&gt;🙃 Gradual Disempowerment: how even “dumb” AI could take over humanity&lt;/h3&gt;
&lt;p&gt;There&#39;s two main tribes in the AI Risk world: 1) those who believe the main threat is that &amp;quot;super-intelligent&amp;quot; AI goes rogue &amp;amp; take over humanity, and 2) those who believe the main threat is that &amp;quot;dumb&amp;quot; AI amplifies economic inequality and digital authoritarianism.&lt;/p&gt;
&lt;p&gt;The recent Gradual Disempowerment paper asks: &lt;em&gt;why not both?&lt;/em&gt;  &lt;strong&gt;&amp;quot;Dumb&amp;quot; AI could take over humanity &lt;em&gt;through&lt;/em&gt; our normal cultural, economic, and political incentives.&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;Here&#39;s how AI slop could disempower us in Culture, Economics, and States:&lt;/p&gt;
&lt;p&gt;= = =&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Culture:&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2025-10/spaghetti.jpg&quot; alt=&quot;Thumbnails of AI-generated videos of Will Smith eating spaghetti (huge progress from 2023 to 2024)&quot; title=&quot;Thumbnails of AI-generated videos of Will Smith eating spaghetti (huge progress from 2023 to 2024)&quot; /&gt;&lt;/p&gt;
&lt;p&gt;(&lt;a href=&quot;https://www.youtube.com/watch?v=vbWe5k4fFWE&quot;&gt;AI will smith eating spaghetti&lt;/a&gt; in 2023 vs 2024. jfc &lt;a href=&quot;https://www.youtube.com/watch?v=bXKkZh2UEEA&quot;&gt;2025&lt;/a&gt; is scarily realistic)&lt;/p&gt;
&lt;p&gt;This is the one we&#39;re all, unfortunately, most familiar with.&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;First, we made chronological internet feeds.&lt;/li&gt;
&lt;li&gt;Then, when there was too much content, we invented Newgrounds-style &amp;amp; Reddit-style voting to find the gems.&lt;/li&gt;
&lt;li&gt;Then, to counter spam (&amp;amp; to make investors happy), we put opaque machine-learning algorithms in charge of boosting content. Importantly: these algorithms are &lt;em&gt;grown, not designed.&lt;/em&gt; To over-emphasize: &lt;em&gt;nobody understands how these algorithms really work,&lt;/em&gt; not even the engineers or the algorithm itself.&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;So, human autonomy is now &lt;em&gt;near-fully removed&lt;/em&gt; from the &amp;quot;consumer&amp;quot; side of modern culture: our media diet is filtered through an algorithm that &lt;em&gt;literally nobody understands&lt;/em&gt;. We&#39;re &lt;em&gt;already&lt;/em&gt; disempowered here.&lt;/p&gt;
&lt;p&gt;But wait, it gets worse! To stay in the game, creators are forced to re-design their content &lt;em&gt;for the algorithm&lt;/em&gt;. The &lt;em&gt;algorithm itself&lt;/em&gt; is the primary audience now. Some of it&#39;s not too bad: tacky clickbait thumbnails, a supercut in the first 30 seconds. But some of it&#39;s pretty bad: outrage-entrepreneurship, straight-up lies &amp;amp; (engaging) bullshit.&lt;/p&gt;
&lt;p&gt;But wait, it gets &lt;em&gt;even worse!&lt;/em&gt; As the space gets more competitive, creators will be pressured to make more at lower cost. Well, what&#39;s the harm in a few AI-generated images? Or AI video clips? C&#39;mon, I &lt;em&gt;have&lt;/em&gt; to let an AI pick the best title &amp;amp; thumbnail. And maybe AI voiceover? Sure, an AI&#39;s writing the script, but I&#39;m still generating the idea &amp;amp; outline! Okay, just the idea. Okay, AI can handle that too.&lt;/p&gt;
&lt;p&gt;OpenAI &amp;amp; Meta (Facebook) recently announced &lt;a href=&quot;https://www.vox.com/technology/464097/meta-openai-sora-slop-ai&quot;&gt;their own AI-video versions of TikTok&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;To recap: we&#39;re already disempowered on the &amp;quot;consumer&amp;quot; side of culture. Soon, we may be disempowered on the &amp;quot;creator&amp;quot; side of culture. &lt;strong&gt;We&#39;ll have AIs making content for AIs, humans in the backseat.&lt;/strong&gt; We won&#39;t even remember there used to be a steering wheel.&lt;/p&gt;
&lt;p&gt;(Oh and &lt;em&gt;then&lt;/em&gt; there&#39;s the AI Companions. &lt;a href=&quot;https://globaldialogues.ai/updates/global-dialogues-4-human-ai-relationships&quot;&gt;Source for all the following stats&lt;/a&gt;: ~40% of people use AI for emotional support at least once a week.&lt;sup class=&quot;footnote-ref&quot;&gt;&lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#fn9&quot; id=&quot;fnref9&quot;&gt;[9]&lt;/a&gt;&lt;/sup&gt; ~17% accept AI-Human romance, and ~11% would &lt;em&gt;personally consider&lt;/em&gt; dating an AI. To be clear, I&#39;m no prude: &lt;em&gt;I&lt;/em&gt; use Claude as an &amp;quot;AI Life Coach&amp;quot; on a near-daily basis, and I&#39;ve done months-long romantic roleplay with AI characters.)&lt;/p&gt;
&lt;p&gt;(But, consider the slippery slope: AI Companions can be more attentive and less demanding than a human could ever be, so you&#39;ll slowly drift towards AI friends &amp;amp; lovers like you slowly drift towards 1, 2, 3+ hours on Discord and Instagram. Then, your human-interaction skills atrophy, so you tend more towards AI Companions, so your human-skills atrophy more, repeat.)&lt;/p&gt;
&lt;p&gt;(Conclusion: not only will &lt;em&gt;mass media culture&lt;/em&gt; be non-human, even &lt;em&gt;person-to-person culture&lt;/em&gt; will be non-human.)&lt;/p&gt;
&lt;p&gt;= = =&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Economics:&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;I&#39;m a programmer. So let&#39;s start with how programming, as a career, could fall to Death By Slop:&lt;/p&gt;
&lt;p&gt;To stay competitive in the marketplace, we&#39;ll be incentivized to rely more and more on LLM coders. At first, senior developers do better than ever, thanks to these LLM coders doing all the grunt work! But junior developers can&#39;t get a job doing that grunt work anymore, which also means juniors can&#39;t get the experience to &lt;em&gt;become&lt;/em&gt; seniors. So, when the seniors die out or cognitively age out, there is no new guard to take over. And/or, the more that seniors depend on LLM coders, the more their &lt;em&gt;own&lt;/em&gt; skills atrophy.&lt;/p&gt;
&lt;p&gt;(&lt;a href=&quot;https://www.brainonllm.com/&quot;&gt;A recent preprint&lt;/a&gt; by MIT researchers found that, at least for essay-writing, LLM users&#39; skills &lt;em&gt;do&lt;/em&gt; atrophy, and even &amp;quot;consistently underperformed at neural, linguistic, and behavioral levels&amp;quot;.)&lt;/p&gt;
&lt;p&gt;Either way, humanity loses the ability to even &lt;em&gt;check&lt;/em&gt; if the LLM-written code is safe, and not – say – making the bioprinting lab&#39;s password &amp;quot;hunter2&amp;quot;.&lt;/p&gt;
&lt;p&gt;Now consider Slop coming for management positions. I mean, it&#39;s mostly emails &amp;amp; Slacks &amp;amp; meetings anyway, right? A CEO would love to get rid of the middlemen, accumulate the extra money for themselves. Until the shareholders vote to get rid of the human CEO, a million-dollar money sink, and put GPT-CEO in charge (with some human stooge as CEO on paper, just for legal reasons). The shareholders are probably AI themselves at this point; stock-trading is &lt;em&gt;already&lt;/em&gt; almost entirely algorithmic. (Which may or may not have led to the glitch of &lt;a href=&quot;https://en.wikipedia.org/wiki/2010_flash_crash&quot;&gt;the 2010 Flash Crash&lt;/a&gt; — as AIs get put in more positions of economic power, &lt;em&gt;and the failure modes of AIs are correlated&lt;/em&gt;, such sudden all-at-once failures become more likely.)&lt;/p&gt;
&lt;p&gt;&lt;em&gt;But, at least the consumer can vote with their dollar?&lt;/em&gt; Do you know how Amazon makes most of their profits? It&#39;s not the marketplace, or the books, or the streaming service. It&#39;s their cloud compute.&lt;sup class=&quot;footnote-ref&quot;&gt;&lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#fn10&quot; id=&quot;fnref10&quot;&gt;[10]&lt;/a&gt;&lt;/sup&gt; Yes, the big tech company makes most of its profits &lt;em&gt;selling to other&lt;/em&gt; big tech companies. And the &lt;em&gt;biggest&lt;/em&gt; company in the world right now? NVIDIA, the chip manufacturer, got to the top of the world by &lt;em&gt;selling to other tech companies&lt;/em&gt;.&lt;/p&gt;
&lt;p&gt;If dollars are votes, humans &lt;em&gt;already&lt;/em&gt; lack majority vote for the world&#39;s top companies.&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2025-10/bloomberg.webp&quot; alt=&quot;no alt text set, shame on me&quot; title=&quot;no alt text set, shame on me&quot; /&gt;&lt;/p&gt;
&lt;p&gt;(&lt;a href=&quot;https://www.bloomberg.com/news/features/2025-10-07/openai-s-nvidia-amd-deals-boost-1-trillion-ai-boom-with-circular-deals&quot;&gt;image source&lt;/a&gt;)&lt;/p&gt;
&lt;p&gt;Imagine a world, where AI-run companies buy from &amp;amp; sell to AI-run companies. Almost all human-run shops and products get immediately outcompeted. Voting with your dollar means nothing, if you even &lt;em&gt;have&lt;/em&gt; dollars to vote with, after losing your job years ago. Maybe there&#39;ll be a Universal Basic Income? I&#39;d hope so – but what incentive do the business leaders &amp;amp; politicians have to keep their promise to implement UBI, when we&#39;re &lt;em&gt;already&lt;/em&gt; disempowered?&lt;/p&gt;
&lt;p&gt;In sum: Labour gets automated away. Management gets automated away. Capital is almost-entirely owned by companies, fictional legal persons, which no longer have real persons at the wheel.&lt;/p&gt;
&lt;p&gt;That&#39;s how we get disempowered. Not a coup, &amp;quot;just&amp;quot; dumb AI + dumb incentives.&lt;/p&gt;
&lt;p&gt;= = =&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;States:&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;C&#39;mon, you already knew politicians don&#39;t write their own speeches. And soon, maybe not even &lt;em&gt;any human&lt;/em&gt; writes those speeches. An AI, with access to the heartbeat of the data of the swing voters, could give a politician superhuman ability to win the polls. And if they refuse? Well, they&#39;ll be outvoted by someone else who takes the AI boost. Why stop at speeches? Why not have AI optimize your platform, campaign promises, specific policies, &lt;em&gt;law?&lt;/em&gt; Why not have an AI run the entire government through teleprompted meat-puppets in suits?&lt;/p&gt;
&lt;p&gt;(Assuming voting decisions are even human at this point; an AI-mediated Culture will show voters what the Algorithm &lt;em&gt;wants&lt;/em&gt; the voters to see, and they&#39;ll vote accordingly. &lt;em&gt;But the human spirit will rebel if democracy&#39;s on the line&lt;/em&gt;, you think? Sure, an AI could predict which humans are rebellious, then give them the opposite content to reverse-psychology them into the correct actions.)&lt;/p&gt;
&lt;p&gt;As for law &lt;em&gt;enforcement&lt;/em&gt;, we don&#39;t have enough police to patrol all the streets, let alone the Wild West of the internet. So, let&#39;s put 24/7 cameras everywhere, with AIs to monitor &amp;amp; report crime. The officers the AI dispatches &lt;em&gt;are&lt;/em&gt; human... until we can manufacture Robo-Cops who never sleep, never lose their cool, and – importantly – never unionize or complain about lack of pay. This same law-enforcement AI could track people online for acting or thinking suspiciously. And if you don&#39;t use the internet at all? Why, that&#39;s the most suspicious action of all.&lt;/p&gt;
&lt;p&gt;Actually, I guess we can just use self-piloted drones with a gun attached as a Robo-Cop. And war robot. And neighbouring-country-conquest robot.&lt;/p&gt;
&lt;p&gt;Oh btw, people &lt;em&gt;already&lt;/em&gt; trust AI chatbots more than their elected representatives, civil servants, and even faith/community leaders. (&lt;a href=&quot;https://blog.cip.org/p/when-ai-acts-for-you-or-as-you&quot;&gt;source&lt;/a&gt;)&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2025-10/Pasted%20image%2020251029164330.png&quot; alt=&quot;Chart of people&#39;s trust in various entities. Family doctors rank #1, AI chatbots #2, government &amp;amp; faith leaders below that.&quot; title=&quot;Chart of people&#39;s trust in various entities. Family doctors rank #1, AI chatbots #2, government &amp;amp; faith leaders below that.&quot; /&gt;&lt;/p&gt;
&lt;p&gt;In &lt;em&gt;EVERY&lt;/em&gt; continent, more people Agree than Disagree that AI could make better decisions than their government representatives: (though &amp;quot;Unsure&amp;quot; is a large portion) (&lt;a href=&quot;https://blog.cip.org/p/when-ai-acts-for-you-or-as-you&quot;&gt;source&lt;/a&gt;)&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2025-10/Pasted%20image%2020251029164718.png&quot; alt=&quot;Chart showing trust in AI &amp;gt; government for every continent&quot; title=&quot;Chart showing trust in AI &amp;gt; government for every continent&quot; /&gt;&lt;/p&gt;
&lt;p&gt;(Maybe Antarctica is the exception)&lt;/p&gt;
&lt;p&gt;Point is: worldwide, more people than not &lt;em&gt;already would&lt;/em&gt; trust AI with governance over the current human bastards. To be fair, I can&#39;t blame &#39;em. Our leaders suck. To quote &lt;a href=&quot;https://www.youtube.com/watch?v=fwzbIUffcR4&quot;&gt;the first song&lt;/a&gt; from OK Go&#39;s first new album in over 10 years:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;🎵 Still, no stochastic parrot has yet called&lt;br /&gt;
🎵 On his nation to knock back bleach&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;= = =&lt;/p&gt;
&lt;p&gt;Now, although the slope is slippery, I &lt;em&gt;do&lt;/em&gt; like my LLM coding assistant, and I&#39;m sympathetic to folks with AI friends and AI romances.&lt;/p&gt;
&lt;p&gt;Still, the Gradual Disempowerment paper made a powerful case for a new possible type of AI takeover — not by a super-intelligent AI seeking power, but humans being lazy &amp;amp; greedy, giving away more and more of our autonomy to AIs. And if you try to opt out of the race, you just get trampled.&lt;/p&gt;
&lt;p&gt;The real point of the paper: AI alignment isn&#39;t enough, we need &lt;em&gt;human&lt;/em&gt; alignment.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Link to &lt;a href=&quot;https://gradual-disempowerment.ai/&quot;&gt;Gradual Disempowerment&lt;/a&gt;, &lt;a href=&quot;https://arxiv.org/abs/2501.16946&quot;&gt;full paper on arXiv&lt;/a&gt;&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;My one critique of the paper is it doesn&#39;t even &lt;em&gt;try&lt;/em&gt; to hint at solutions? Way to leave a girl high &amp;amp; dry, y&#39;all. But personally, I think the &lt;a href=&quot;https://vitalik.eth.limo/general/2025/01/05/dacc2.html&quot;&gt;d/acc&lt;/a&gt; and &lt;a href=&quot;https://vitalik.eth.limo/general/2024/08/21/plurality.html&quot;&gt;Plurality&lt;/a&gt; approaches are roughly correct: we need &amp;quot;human alignment&amp;quot; approaches that &lt;em&gt;scales with&lt;/em&gt; improving tech, instead of getting obsolete. If all that sounds vague as hell, that&#39;s what the next 4 sections are for, to give you 4 concrete examples of how tech can work &lt;em&gt;with, not against&lt;/em&gt; human autonomy:&lt;/p&gt;
&lt;ol&gt;
&lt;li&gt;🤝 Reverse political polarization by reversing the algorithm &lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#bridging&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;🧐 Cryptography that lets you prove yourself without doxxing yourself &lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#zkp&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;🧠 Cyborgism: AI that &lt;em&gt;enhances&lt;/em&gt; us, not replaces us &lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#cyborg&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;🍻 The 6Pack of Digital Democracy &lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#6pack&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;/ol&gt;
&lt;p&gt;Let&#39;s dive in:&lt;/p&gt;
&lt;p&gt;&lt;a id=&quot;bridging&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h3&gt;🤝 BellKor &amp;amp; Birdwatch &amp;amp; Bridging: reverse political polarization by reversing the algorithm&lt;/h3&gt;
&lt;p&gt;Fun story: &lt;a href=&quot;https://en.wikipedia.org/wiki/Netflix_Prize&quot;&gt;in 2009, Netflix launched a million-dollar prize&lt;/a&gt; for a recommendation algorithm that could beat their own by 10%. There was a winner! Netflix paid out the million dollars! Then they didn&#39;t use that algorithm &amp;amp; just wrote their own lol&lt;/p&gt;
&lt;p&gt;Okay &lt;a href=&quot;https://www.techdirt.com/2012/04/13/why-netflix-never-implemented-algorithm-that-won-netflix-1-million-challenge/&quot;&gt;that&#39;s not the full story&lt;/a&gt;. The winning &amp;quot;algorithm&amp;quot;, BellKor&#39;s Pragmatic Chaos, was actually a collection of 107 different algorithms. Almost all of these algorithms only added an extra ~0.1% accuracy to the final collection. But, &lt;em&gt;one&lt;/em&gt; of these algorithms, called &lt;strong&gt;&lt;a href=&quot;https://datajobs.com/data-science-repo/Recommender-Systems-%5BNetflix%5D.pdf&quot;&gt;Matrix Factorization&lt;/a&gt;&lt;/strong&gt;, was responsible for almost all of the accuracy boost! &lt;em&gt;That&lt;/em&gt; was (part of) what Netflix kept in their new algorithm, and even to this day, it&#39;s the core of most recommendation systems online.&lt;/p&gt;
&lt;p&gt;But how does Matrix Factorization work? Well it&#39;s &amp;quot;elegant&amp;quot; in that &lt;strong&gt;it&#39;s only 2 lines of math&lt;/strong&gt;, but because academic writing &amp;amp; math notation sucks, it still took me 30 minutes with Claude&#39;s help to understand.&lt;/p&gt;
&lt;p&gt;Anyway, here&#39;s my attempt at explaining the algorithm:&lt;/p&gt;
&lt;p&gt;= = =&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Step 1) Predict each user&#39;s rating of an item, as the sum of four things:&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;a) &lt;strong&gt;How much a user&#39;s preferences align with this item&#39;s features&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;Example: If the algorithm knows I love horror, and it knows Movie X is horror, my preferences align perfectly with the item&#39;s features, so alignment = 1. (If I hate horror, alignment = -1, if I&#39;m indifferent, alignment = 0.) Now, here&#39;s the neat part: &lt;em&gt;you do not need to hard-code the preferences/features!&lt;/em&gt;  The algorithm learns by itself which factors best predict ratings. (So instead of Horror, the algorithm just &amp;quot;thinks&amp;quot; of it as Factor #42 or something.)&lt;/p&gt;
&lt;p&gt;b) &lt;strong&gt;The item&#39;s &amp;quot;bias&amp;quot;&lt;/strong&gt;: how well-rated an item is, &lt;em&gt;independent of how much it aligns with user preferences.&lt;/em&gt; You can &lt;em&gt;roughly&lt;/em&gt; think of this as an item&#39;s &amp;quot;quality&amp;quot;.&lt;/p&gt;
&lt;p&gt;Example: Users like me prefer supernatural slasher horror-comedies, and the movies &lt;em&gt;Final Destination 4&lt;/em&gt; and &lt;em&gt;Final Destination: Bloodlines&lt;/em&gt; both align fully with our preferences. However, &lt;em&gt;Bloodlines&lt;/em&gt; is rated higher than &lt;em&gt;4&lt;/em&gt; whether or not one likes supernatural slasher horror-comedies, because it&#39;s just a higher-quality film.&lt;/p&gt;
&lt;p&gt;c) &lt;strong&gt;The user&#39;s &amp;quot;bias&amp;quot;&lt;/strong&gt;. How friendly/critical this user is in their ratings, i.e. when they rate 5-out-of-10, is that &amp;quot;average&amp;quot; or &amp;quot;really bad&amp;quot;?&lt;/p&gt;
&lt;p&gt;d) &lt;strong&gt;The global &amp;quot;bias&amp;quot;.&lt;/strong&gt; How friendly/critical are users in general.&lt;/p&gt;
&lt;p&gt;To recap:&lt;/p&gt;
&lt;p&gt;Predicted rating&lt;br /&gt;
= Global bias&lt;br /&gt;
+ User bias&lt;br /&gt;
+ Item bias&lt;br /&gt;
+ User preferences aligning with item features&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2025-10/formula_1.png&quot; alt=&quot;The official formula, annotated explanation&quot; title=&quot;The official formula, annotated explanation&quot; /&gt;&lt;/p&gt;
&lt;p&gt;= = =&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Step 2) To train the algorithm, minimize:&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;The gap (&lt;strong&gt;&amp;quot;error&amp;quot;&lt;/strong&gt;) between {predicted ratings} &amp;amp; {actual ratings}&lt;br /&gt;
+ The &lt;strong&gt;&amp;quot;complexity&amp;quot;&lt;/strong&gt; of your algorithm.  (to &amp;quot;keep it simple stupid&amp;quot;, Occam&#39;s Razor)&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2025-10/formula_2.png&quot; alt=&quot;The official formula, annotated explanation&quot; title=&quot;The official formula, annotated explanation&quot; /&gt;&lt;/p&gt;
&lt;p&gt;= = =&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Step 3) To find recommendations, maximize:&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;The same equation as in Step 1!&lt;/p&gt;
&lt;p&gt;Predicting how you&#39;d rate this item you&#39;ve not seen before&lt;br /&gt;
= Global bias&lt;br /&gt;
+ &lt;strong&gt;Your&lt;/strong&gt; bias&lt;br /&gt;
+ Item&#39;s bias&lt;br /&gt;
+ How much &lt;strong&gt;your&lt;/strong&gt; preferences align with this item&#39;s features&lt;/p&gt;
&lt;p&gt;Million-dollar prize, two lines of math, that&#39;s $500,000 per equation!&lt;/p&gt;
&lt;p&gt;= = =&lt;/p&gt;
&lt;p&gt;In hindsight, there was a possible downside.&lt;/p&gt;
&lt;p&gt;By giving people &lt;em&gt;what they&#39;re already into&lt;/em&gt;, you nudge them into staying the same, avoiding exploration &amp;amp; growth. For movies &amp;amp; TV shows, this isn&#39;t &lt;em&gt;too&lt;/em&gt; bad: Netflix predicts I like horror, so it gives me more horror, so I become even more into horror, repeat forever.&lt;/p&gt;
&lt;p&gt;But when this algorithm got applied to social media, &lt;em&gt;specifically politics &amp;amp; news&lt;/em&gt;, it promotes polarization, and reduces cross-tribe win-win understanding. YouTube predicts I like left-libertarian content, so it gives me more pro-left-libertarian content, so I become even more left-libertarian, repeat forever. I &lt;em&gt;can&lt;/em&gt; seek out social-conservative and Marxist-leftist and Yarvin-autocrat stuff, (and I do for &lt;s&gt;masochism&lt;/s&gt; research), but the algorithms put up friction for that, while &amp;quot;see what I&#39;ll already agree with&amp;quot; is the WD-40 easy-glide default.&lt;/p&gt;
&lt;p&gt;(Counter-argument: see above paper &lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#we-suck&quot;&gt;↪&lt;/a&gt;, maybe it&#39;s not the algorithms&#39; fault, we just suck.)&lt;/p&gt;
&lt;p&gt;= = =&lt;/p&gt;
&lt;p&gt;A few programmers at Twitter (back when it was called Twitter) worried about this, too.&lt;/p&gt;
&lt;p&gt;They worried how this million-dollar algorithm, which worked for getting niche movies to niche audiences, could fracture democracy with an infinite fractal of niche echo chambers within echo chambers. Meanwhile, news sites tried &amp;quot;factcheckers&amp;quot;, but people hated &amp;amp; distrusted them. After all, who watches the watchers, who factchecks the factcheckers?&lt;/p&gt;
&lt;p&gt;So the Twitter programmers wondered — could they make a &amp;quot;fact-checking&amp;quot; service &lt;em&gt;run by the public themselves&lt;/em&gt;, by &lt;em&gt;using&lt;/em&gt; an edit of that million-dollar algorithm, for good?&lt;/p&gt;
&lt;p&gt;They did, &lt;a href=&quot;https://asteriskmag.com/issues/08/the-making-of-community-notes&quot;&gt;they called it &amp;quot;Birdwatch&amp;quot;&lt;/a&gt;&lt;sup class=&quot;footnote-ref&quot;&gt;&lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#fn11&quot; id=&quot;fnref11&quot;&gt;[11]&lt;/a&gt;&lt;/sup&gt;, and here&#39;s how it worked:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Let anyone submit &amp;quot;notes&amp;quot; to factcheck viral tweets.&lt;/li&gt;
&lt;li&gt;People rate how helpful those notes are.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;You keep the algorithm&#39;s Step 1 &amp;amp; Step 2 the same&lt;/strong&gt;, so it can learn what people&#39;s preferences &amp;amp; notes&#39; features are.
&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;The user/item factors:&lt;/strong&gt; Again, we do &lt;em&gt;not&lt;/em&gt; need to hard-code the factors. But in practice, the algorithm learns that the #1 factor that most predicts people&#39;s ratings, is the left-right political spectrum. Adding extra factors like &amp;quot;authoritarian-libertarian&amp;quot; doesn&#39;t improve prediction much.&lt;sup class=&quot;footnote-ref&quot;&gt;&lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#fn12&quot; id=&quot;fnref12&quot;&gt;[12]&lt;/a&gt;&lt;/sup&gt; (Which is surprising given past research shows the general public&#39;s politics is at &lt;em&gt;least&lt;/em&gt; 2D.&lt;sup class=&quot;footnote-ref&quot;&gt;&lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#fn13&quot; id=&quot;fnref13&quot;&gt;[13]&lt;/a&gt;&lt;/sup&gt; Maybe the Birdwatch community is just weird.)&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;The item&#39;s bias:&lt;/strong&gt; Instead of being &amp;quot;movie quality&amp;quot;, it&#39;s now &amp;quot;note quality&amp;quot;, &lt;em&gt;independent&lt;/em&gt; of how much it aligns with users&#39; politics. IMPORTANTLY: note quality is &lt;em&gt;not&lt;/em&gt; just a note&#39;s average helpfulness-rating. &lt;em&gt;Average rating&lt;/em&gt; will be skewed by the &lt;em&gt;average&lt;/em&gt; Birdwatch-rater&#39;s political preferences. This setup gets us a note&#39;s quality &lt;em&gt;regardless&lt;/em&gt; of politics.&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;The key difference is WE REVERSE STEP 3&lt;/strong&gt;: Instead of taking user preferences (ie politics) into account, we highlight the best notes &lt;em&gt;NOT&lt;/em&gt; taking user preferences into account! That means: this highlights notes that &lt;em&gt;people across the political spectrum&lt;/em&gt; agree is helpful. Common ground, by algorithmic design!&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;Here&#39;s an example of a factcheck note, that&#39;s rated as helpful by people across the political spectrum, because it&#39;s very specific &amp;amp; easily verifiable:&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2025-10/Pasted%20image%2020251030152607.png&quot; alt=&quot;Original poster posts a photo of a car on fire, claiming it&#39;s outside the Supreme Court in Washington DC. Below it, Birdwatch selected a factcheck: No, that&#39;s a photo from a 2010 protest in Toronto, here&#39;s the video the screenshot&#39;s from: (link)&quot; title=&quot;Original poster posts a photo of a car on fire, claiming it&#39;s outside the Supreme Court in Washington DC. Below it, Birdwatch selected a factcheck: &amp;quot;No, that&#39;s a photo from a 2010 protest in Toronto, here&#39;s the video the screenshot&#39;s from: (link)&amp;quot;&quot; /&gt;&lt;/p&gt;
&lt;p&gt;(The Birdwatch creators made sure to never call the notes &amp;quot;factchecks&amp;quot;, and instead said &amp;quot;Readers added context you may want to know&amp;quot;... but c&#39;mon, they&#39;re factchecks.)&lt;/p&gt;
&lt;p&gt;Here&#39;s a graph of all the notes from their pilot program. Each dot is a note. X-axis is the &amp;quot;item factor&amp;quot; ~= &amp;quot;how politically left-right coded the note is&amp;quot;, and the Y-axis is the &amp;quot;item bias&amp;quot; ~= &amp;quot;how helpful the note is &lt;em&gt;regardless&lt;/em&gt; of politics&amp;quot;.&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2025-10/birdwatch.png&quot; alt=&quot;Figure 2 from the Birdwatch paper, explanation below&quot; title=&quot;Figure 2 from the Birdwatch paper, explanation below&quot; /&gt;&lt;/p&gt;
&lt;p&gt;As you can see in the dense yellow circles, most people submit partisan slop. But at the top, there&#39;s the few gems folks across the spectrum agree are helpful! (And at the bottom... well, farting in an elevator &amp;quot;pisses off both sides&amp;quot;. &amp;quot;Pissing off both sides&amp;quot; doesn&#39;t mean you&#39;re &lt;em&gt;useful&lt;/em&gt;.) As for the diamond shape, don&#39;t worry about it, it&#39;s an artifact of how the &amp;quot;minimize complexity&amp;quot; math works in Step 2.&lt;/p&gt;
&lt;p&gt;There&#39;s a few extra complexities, but that&#39;s the heart of the Birdwatch algorithm! The result? It&#39;s, as far as I know, the &lt;em&gt;only&lt;/em&gt; fact-checking service that gets net-positive ratings (more &amp;quot;helpful&amp;quot; than &amp;quot;unhelpful&amp;quot;) from Democrats, Independents, and Republicans alike! In this polarized era, that&#39;s no mean feat.&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2025-10/birdwatch-2.png&quot; alt=&quot;Figure 4 from the paper, showing how every party rates Birdwatch as net-helpful.&quot; title=&quot;Figure 4 from the paper, showing how every party rates Birdwatch as net-helpful.&quot; /&gt;&lt;/p&gt;
&lt;p&gt;(Wait, if the notes are chosen to be helpful &lt;em&gt;regardless of politics&lt;/em&gt;, why do Democrats like them more than Republicans, even while Republicans still find them net-helpful? I&#39;m not sure, but I&#39;d guess the left-right axis &lt;em&gt;among Birdwatch raters&lt;/em&gt; is slightly different from the left-right axis &lt;em&gt;among party-registered voters&lt;/em&gt;. Remember, the algorithm &lt;em&gt;learns&lt;/em&gt; what the left-right axis is, it&#39;s not hard-coded; and it shouldn&#39;t be, since what &amp;quot;left-right&amp;quot; means changes across time &amp;amp; cultures.)&lt;/p&gt;
&lt;p&gt;Unfortunately, this &amp;quot;bridging&amp;quot; algorithm was only used for the factcheck-notes, &lt;em&gt;not&lt;/em&gt; Twitter&#39;s algorithmic feed. As mentioned in the paper two sections ago, &lt;strong&gt;bridging-based algorithms were the &lt;em&gt;only&lt;/em&gt; design choice we know of (so far) that reverses polarization &amp;amp; extremism.&lt;/strong&gt; So, implementing bridging &lt;em&gt;for the main feeds themselves&lt;/em&gt; could be a big win for digital democracy. And, good news? I heard that 𝕏 may adopt bridging for their main feed, and other platforms will adopt bridging-based algorithm, &lt;em&gt;at least&lt;/em&gt; for a similar factchecking system.&lt;/p&gt;
&lt;p&gt;...my &lt;em&gt;source&lt;/em&gt; on this info? I dunno, I heard it somewhere. Why are you factchecking me?&lt;/p&gt;
&lt;p&gt;= = =&lt;/p&gt;
&lt;p&gt;(Aside: Birdwatch inspires me – what other small alterations can we make to the million-dollar Matrix Factorization algorithm? What if you recommended high-quality content that &lt;em&gt;anti-aligns&lt;/em&gt; with your usual preferences? Or recommend content that aligns with &lt;em&gt;all but one&lt;/em&gt; of your preferences? e.g. I like animation &amp;amp; musicals &amp;amp; monsters &amp;amp; queer found-family coming-out allegories &amp;amp; muscular women, but I mildly dislike K-Pop =&amp;gt; the algorithm recommends K-Pop Demon Hunters =&amp;gt; I now mildly like K-Pop.)&lt;/p&gt;
&lt;p&gt;= = =&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Links:&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;a href=&quot;https://datajobs.com/data-science-repo/Recommender-Systems-%5BNetflix%5D.pdf&quot;&gt;The winners of the Netflix Prize explain Matrix Factorization&lt;/a&gt; (technical)&lt;/li&gt;
&lt;li&gt;&lt;a href=&quot;https://asteriskmag.com/issues/08/the-making-of-community-notes&quot;&gt;An interview with the creators of Birdwatch&lt;/a&gt; (lay-friendly)&lt;/li&gt;
&lt;li&gt;&lt;a href=&quot;https://arxiv.org/pdf/2210.15723&quot;&gt;The original Birdwatch paper&lt;/a&gt; (technical)&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;a id=&quot;zkp&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h3&gt;🧐 PolyLog&#39;s explanation of Zero-Knowledge Proofs&lt;/h3&gt;
&lt;p&gt;FINALLY. Thanks to PolyLog&#39;s video, I &lt;em&gt;finally&lt;/em&gt; understand one of the coolest recent discoveries in computer science: that you can prove you have a solution to a problem, &lt;em&gt;without revealing any info whatsoever about your solution.&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Here&#39;s &lt;a href=&quot;https://www.youtube.com/watch?v=Otvcbw6k4eo&quot;&gt;their 20-min video&lt;/a&gt; on Zero-Knowledge Proofs (ZKPs) ⤵&lt;/strong&gt;  I won&#39;t try to out-do their explanation in this blog post, you can just watch it.  They show you how you can prove you &lt;em&gt;have&lt;/em&gt; a valid solution to a Sudoku puzzle, &lt;em&gt;without revealing ANY info about your solution&lt;/em&gt;. (And this method works &lt;em&gt;in general&lt;/em&gt; for any mathematical/computable proof!)&lt;/p&gt;
&lt;iframe width=&quot;640&quot; height=&quot;360&quot; src=&quot;https://www.youtube-nocookie.com/embed/Otvcbw6k4eo?rel=0&quot; title=&quot;YouTube video player&quot; frameborder=&quot;0&quot; allow=&quot;accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share&quot; referrerpolicy=&quot;strict-origin-when-cross-origin&quot; allowfullscreen=&quot;&quot;&gt;&lt;/iframe&gt;
&lt;p&gt;(&lt;strong&gt;&lt;a href=&quot;https://www.youtube.com/@PolylogCS&quot;&gt;check out PolyLog&#39;s channel&lt;/a&gt;&lt;/strong&gt;, their other videos are pretty good, too!)&lt;/p&gt;
&lt;p&gt;But why am I boosting this video, in relation to &amp;quot;The Algorithm&amp;quot; and solutions to &amp;quot;Gradual Disempowerment&amp;quot;? Because: Zero-Knowledge Proofs is a big win for privacy in a digital democracy. How? Because they allow you to &lt;em&gt;prove yourself&lt;/em&gt;, &lt;em&gt;without doxxing yourself&lt;/em&gt;. For example:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Do you want an online petition or discussion board, to authenticate that users are actually a resident of some area, or affiliated with some institution? BAM! ZKPs let you prove &amp;quot;yes I&#39;m a resident of X&amp;quot; or &amp;quot;yes I&#39;m with institution Y&amp;quot; &lt;em&gt;while revealing no other info, like your name or ID.&lt;/em&gt;&lt;/li&gt;
&lt;li&gt;Are you worried that governments &lt;a href=&quot;https://en.wikipedia.org/wiki/Online_Safety_Act_2023&quot;&gt;like the UK&lt;/a&gt; are forcing age authentication, under the disguise of &amp;quot;protecting children&amp;quot;, but it&#39;s secretly the first step to creating a China-style tech-powered surveillance state while accusing all opponents of being enablers for pedophiles &amp;amp; terrorists? BAM! ZKPs let you prove &amp;quot;yes I&#39;m over 18&amp;quot; &lt;em&gt;without even revealing your exact age.&lt;/em&gt;&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;Of course, &lt;a href=&quot;https://vitalik.eth.limo/general/2025/06/28/zkid.html&quot;&gt;ZKPs won&#39;t solve all privacy issues&lt;/a&gt;, and institutions can just &lt;em&gt;lie&lt;/em&gt; about using ZKPs — but that&#39;s more reason to make the core idea of ZKPs accessible, and not seem like incomprehensible math magic! Hence, why I&#39;m so glad for PolyLog&#39;s lay-friendly explainer. ZKPs can and should be a core tool in our privacy toolbelt, for digital democracies.&lt;/p&gt;
&lt;p&gt;(Current brain status: I have a technical understanding of how ZKPs work &lt;em&gt;in general&lt;/em&gt;, and a rough idea of how ZKPs work &lt;em&gt;for authentication specifically&lt;/em&gt;... but I&#39;m still wrapping my head around the mathematical details: polynomial commitments, elliptic curves, homeomorphic encryption, etc. Once I &lt;em&gt;really&lt;/em&gt; understand those, I may make a video explainer on all this. Maybe.)&lt;/p&gt;
&lt;p&gt;&lt;a id=&quot;cyborg&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h3&gt;🧠 Geoffrey Litt&#39;s writings: AI should enhance us, not replace us&lt;/h3&gt;
&lt;p&gt;Problem 1: &lt;em&gt;Truly general&lt;/em&gt; AI could automate huge sectors of labour all at once, not &amp;quot;just&amp;quot; the piece-by-piece automation we&#39;ve dealt with in the past. And as history shows, &amp;quot;suddenly mass unemployment&amp;quot; almost never ends well for a country.&lt;/p&gt;
&lt;p&gt;Problem 2: If we hand over control to AI &lt;em&gt;without&lt;/em&gt; fully solving the AI Value Alignment problem, we&#39;d be passing control of humanity over to entities that neither love us nor hate us, we&#39;re just &lt;em&gt;numbers&lt;/em&gt; to make &lt;em&gt;go up&lt;/em&gt;.&lt;/p&gt;
&lt;p&gt;An idea to solve both problems at the same time: &lt;strong&gt;instead of making AI to replace humans, let&#39;s make AI to &lt;em&gt;enhance&lt;/em&gt; humans?&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2025-10/NickyCase0018.png&quot; alt=&quot;Drawing of a goofy bicycle, saying, A computer should be a bicycle for the mind.&quot; title=&quot;Drawing of a goofy bicycle, saying, &amp;quot;A computer should be a bicycle for the mind&amp;quot;.&quot; /&gt;&lt;/p&gt;
&lt;p&gt;&lt;em&gt;(slide from &lt;a href=&quot;https://blog.ncase.me/the-creative-cyborg/&quot;&gt;my talk at the final XOXO&lt;/a&gt;, quote is &lt;a href=&quot;http://web.archive.org/web/20240902210345/https://miro.medium.com/v2/resize:fit:1400/format:webp/0%2ASgDLtymZpgcXEA7A&quot;&gt;from Steve Jobs&lt;/a&gt;)&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;This way, Human+AI combos can still be economically competitive in our Darwinian marketplace, yet keep humanity&#39;s values &amp;amp; autonomy at the centre of our tools. This general idea is called &lt;strong&gt;&lt;a href=&quot;https://www.lesswrong.com/posts/bxt7uCiHam4QXrQAA/cyborgism&quot;&gt;&amp;quot;cyborgism&amp;quot;&lt;/a&gt;&lt;/strong&gt; in the AI Alignment community. (I mean, it&#39;s just tool use, but &amp;quot;cyborg&amp;quot; and &amp;quot;mental prosthetics&amp;quot; sounds cooler.)&lt;/p&gt;
&lt;p&gt;Wait – won&#39;t this increase inequality, with the richest getting the most AI-gains? Well, consider books: books &lt;em&gt;do&lt;/em&gt; enhance the reader, so by default books &lt;em&gt;would have&lt;/em&gt; increased inequality by enhancing those who can already afford the most books. But the solution wasn&#39;t to ban books, but to create free universal libraries! Likewise, to share the AI-enhancement gains, we should at least seriously consider &lt;em&gt;free, open-source, verifiable, publicly owned &amp;amp; evaluated&lt;/em&gt; AI tools, to help &lt;em&gt;everyone&lt;/em&gt; augment their own human autonomy and skills.&lt;/p&gt;
&lt;p&gt;(crucially: The AI tools should be &lt;em&gt;non-agentic&lt;/em&gt;, no goal-seeking of its own. And we should remove risky capabilities like bioweapon knowledge from them. See: &lt;a href=&quot;https://vitalik.eth.limo/general/2025/01/05/dacc2.html&quot;&gt;d/acc&lt;/a&gt;)&lt;/p&gt;
&lt;p&gt;I&#39;ve been banging on this idea of Cyborgism for almost 8 years now. (&lt;a href=&quot;https://jods.mitpress.mit.edu/pub/issue3-case/release/6&quot;&gt;My 2018 article in an MIT Journal&lt;/a&gt;, and &lt;a href=&quot;https://blog.ncase.me/the-creative-cyborg/&quot;&gt;my 2024 XOXO talk&lt;/a&gt;.) But it&#39;s only been an idea. What about an &lt;em&gt;actual implementation,&lt;/em&gt; at least a proof-of-concept?&lt;/p&gt;
&lt;p&gt;Here, I&#39;d like to highlight a couple articles (with &lt;em&gt;actual working prototypes&lt;/em&gt;) from programmer Geoffrey Litt!&lt;/p&gt;
&lt;p&gt;👉 &lt;strong&gt;&lt;a href=&quot;https://www.geoffreylitt.com/2025/07/27/enough-ai-copilots-we-need-ai-huds&quot;&gt;&amp;quot;Enough AI copilots! We need AI HUDs&amp;quot;&lt;/a&gt;&lt;/strong&gt; 👈&lt;/p&gt;
&lt;p&gt;Most LLM-based coding tools right now (GitHub Copilot, Claude Code, AMP Code, etc) all put the LLM in the role of an independent agent. But that&#39;s &lt;em&gt;not&lt;/em&gt; how most automation that improves knowledge-work, works. Examples:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Spellcheck, mercifully, isn&#39;t Clippy heckling you with &amp;quot;HEY DID YOU MEAN TO WRITE &#39;SPELLCHECK&#39; NOT &#39;SPELCHECK&#39;???&amp;quot; Spellcheck just lets you &lt;em&gt;see&lt;/em&gt; something underlined in red, lists suggestions, and lets &lt;em&gt;you&lt;/em&gt; decide when &amp;amp; how to deal with it.&lt;/li&gt;
&lt;li&gt;Data visualization isn&#39;t Clippy yelling &amp;quot;HEY YOU&#39;RE LOSING WEIGHT AT 1 LB PER 10 DAYS&amp;quot;. Spreadsheet apps just let you &lt;em&gt;see&lt;/em&gt; your data as a line chart, then lets &lt;em&gt;you&lt;/em&gt; decide how to interpret &amp;amp; act on it.&lt;/li&gt;
&lt;li&gt;Illustration apps don&#39;t have Clippy yelling &amp;quot;HEY THE CHARACTER&#39;S EYES ARE UNEVEN AND ALSO THE COLOUR VALUES ARE MUDDY&amp;quot;, the program just lets you &lt;em&gt;see&lt;/em&gt; asymmetry &amp;amp; values by toggling Flip Horizontal and Greyscale, then lets &lt;em&gt;you&lt;/em&gt; decide how to fix it, or leave it as an intentional artistic choice.&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;Geoffrey&#39;s analogy: these aren&#39;t &amp;quot;copilots&amp;quot;, they&#39;re Heads-Up Displays (HUDs), like Tony Stark&#39;s helmet – &lt;strong&gt;they &lt;em&gt;augment&lt;/em&gt; your senses, while keeping the autonomy in &lt;em&gt;your&lt;/em&gt; hands.&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;So instead of making Clippy the Coder — which will put junior devs out of a job, and de-skill senior devs — how about we make AI-powered &amp;quot;HUDs&amp;quot;, to let you &amp;quot;just see&amp;quot; what a program does, then lets &lt;em&gt;you&lt;/em&gt; decide what to do about it?&lt;/p&gt;
&lt;p&gt;Well, Geoff made a proof-of-concept for that. Here&#39;s what it looks like:&lt;/p&gt;
&lt;p&gt;&lt;video width=&quot;640&quot; controls=&quot;&quot; src=&quot;https://blog.ncase.me/content/stuff/2025-10/geoff-demo.mp4&quot; aria-label=&quot;no alt text set, shame on me&quot;&gt;&lt;/video&gt;&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;With the debugger, I have a HUD! I have new senses, I can see how my program runs. The HUD extends beyond the narrow task of fixing the bug. I can ambiently build up my own understanding, spotting new problems and opportunities.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;Geoffrey explains it more in &lt;a href=&quot;https://www.geoffreylitt.com/2025/07/27/enough-ai-copilots-we-need-ai-huds&quot;&gt;his post&lt;/a&gt;. Point is: less Clippy!&lt;/p&gt;
&lt;p&gt;👉 &lt;strong&gt;&lt;a href=&quot;https://www.geoffreylitt.com/2023/03/25/llm-end-user-programming&quot;&gt;Malleable software in the age of LLMs&lt;/a&gt;&lt;/strong&gt; 👈&lt;/p&gt;
&lt;p&gt;For all the badly-written vibe code out there, I still have a soft spot for LLM coding, because it &lt;em&gt;could&lt;/em&gt; make a half-century-long dream finally come true: software that is &lt;em&gt;fully modifiable &amp;amp; customizable&lt;/em&gt; by a layperson user. (Fight for the users!&lt;sup class=&quot;footnote-ref&quot;&gt;&lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#fn14&quot; id=&quot;fnref14&quot;&gt;[14]&lt;/a&gt;&lt;/sup&gt;)&lt;/p&gt;
&lt;p&gt;There&#39;s been varying degrees of success with this idea in the past: HyperCard, spreadsheets, the &lt;em&gt;original&lt;/em&gt; design for the World Wide Web was meant to be so that anyone could &lt;em&gt;write&lt;/em&gt; a website &lt;em&gt;as easily as reading one&lt;/em&gt;.&lt;sup class=&quot;footnote-ref&quot;&gt;&lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#fn15&quot; id=&quot;fnref15&quot;&gt;[15]&lt;/a&gt;&lt;/sup&gt;&lt;/p&gt;
&lt;p&gt;But to this day, there&#39;s no way for an average layperson to modify their software the way they can modify a recipe. Users can&#39;t just say, &amp;quot;Huh. Binaural beats sounds interesting. Okay Clippy, modify my pomodoro app so that it plays a pure Beta wave for focus during 25-minute work-sprints, then a random song from my Bandcamp library during 5-minute breaks.&amp;quot;&lt;/p&gt;
&lt;p&gt;Wait, didn&#39;t I &lt;em&gt;just&lt;/em&gt; rail against Clippy&#39;s? Do I contradict myself? Very well, I contradict myself. I contain &lt;s&gt;hypocrisies&lt;/s&gt; multitudes.&lt;/p&gt;
&lt;p&gt;Okay, I haven&#39;t figured out my contradictions yet. But both &amp;quot;More HUDs, less Clippy&amp;quot;, and &amp;quot;Clippy helps you truly &lt;em&gt;own&lt;/em&gt; your software&amp;quot;, point at the same principle: &lt;strong&gt;fight for the users.&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Tools should help us automate everything that gets in the way of our self-expression, &lt;em&gt;not automate away the self-expression itself&lt;/em&gt;.&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;Anyway Geoffrey goes more into &lt;a href=&quot;https://www.geoffreylitt.com/2023/03/25/llm-end-user-programming&quot;&gt;his post&lt;/a&gt;. And he&#39;s made this cool prototype of &amp;quot;end-user modify everything&amp;quot;, too:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;A few years ago, I developed an end-user programming system called &lt;a href=&quot;https://www.geoffreylitt.com/wildcard/&quot;&gt;Wildcard&lt;/a&gt; which would let people customize any website through a spreadsheet interface. For example, in this short demo you can see a user sorting articles on Hacker News in a different order, and then adding read times to the articles in the page, all by manipulating a spreadsheet synced with the webpage.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;&lt;video width=&quot;640&quot; controls=&quot;&quot; src=&quot;https://blog.ncase.me/content/stuff/2025-10/wildcard.mp4&quot; aria-label=&quot;no alt text set, shame on me&quot;&gt;&lt;/video&gt;&lt;/p&gt;
&lt;p&gt;The problem is &amp;quot;the user needs to be able to write small spreadsheet formulas to express computations. This is a lot easier than learning a full-fledged programming language, but it’s still a barrier to initial usage.&amp;quot;&lt;/p&gt;
&lt;p&gt;I don&#39;t know of a demo of it yet, but imagine there was an extension like GreaseMonkey or Stylus, except you &amp;quot;code&amp;quot; the JS or CSS you want by &lt;em&gt;writing natural language&lt;/em&gt; in your browser! &lt;strong&gt;To be clear this is a security nightmare &amp;amp; would be like handing a nuke to a toddler&lt;/strong&gt;, but, &amp;quot;look to where my finger is pointing, not the tip of my finger itself&amp;quot;&lt;sup class=&quot;footnote-ref&quot;&gt;&lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#fn16&quot; id=&quot;fnref16&quot;&gt;[16]&lt;/a&gt;&lt;/sup&gt; — I&#39;m trying to point to a future where &lt;em&gt;we fully own our tools, before they fully own us.&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;Fight for the users!&lt;/p&gt;
&lt;p&gt;= = =&lt;/p&gt;
&lt;p&gt;(Aside: What about AI to help augment our &lt;em&gt;emotional&lt;/em&gt; intelligence? Could just be as simple as an LLM-augmented diary that helps you recognize emotional patterns, debug cognitive distortions, and asks you helpful questions to help you figure it out for yourself. {In Soviet Russia, LLM prompts &lt;em&gt;you?&lt;/em&gt;}  That&#39;s what an ideal AI Friend – heck, even ideal human friends – should be: someone that makes you stronger and better &lt;em&gt;even when they&#39;re not around&lt;/em&gt;. Not someone who fosters dependence and personal-character atrophy.)&lt;/p&gt;
&lt;p&gt;= = =&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Links &amp;amp; Related:&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;&lt;a href=&quot;https://www.geoffreylitt.com/#writing&quot;&gt;Geoffrey Litt&#39;s blog&lt;/a&gt;&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;janus &amp;amp; Nicholas Kees&#39;s &lt;a href=&quot;https://www.lesswrong.com/posts/bxt7uCiHam4QXrQAA/cyborgism&quot;&gt;Cyborgism manifesto&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Vitalik Buterin&#39;s &lt;a href=&quot;https://vitalik.eth.limo/general/2025/02/28/aihumans.html&quot;&gt;&amp;quot;AI as the engine, humans as the steering wheel&amp;quot;&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href=&quot;https://www.lesswrong.com/posts/Gi8NP9CMwJMMSCWvc/ai-for-epistemics-hackathon&quot;&gt;The AI for Epistemics hackathon&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;My article in an MIT Journal: &lt;a href=&quot;https://jods.mitpress.mit.edu/pub/issue3-case/release/6&quot;&gt;How To Become A Centaur&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;My talk at XOXO: &lt;a href=&quot;https://blog.ncase.me/the-creative-cyborg/&quot;&gt;The Creative Cyborg&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;a id=&quot;6pack&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h3&gt;🍻 The 6Pack of Digital Democracy&lt;/h3&gt;
&lt;p&gt;Okay, a bit tacky to Signal Boost this project, since I&#39;m involved in it. But I joined this project &lt;em&gt;because&lt;/em&gt; I sincerely think it&#39;s one of the better bets for Human-AI Alignment, and the person who founded it has been a role model of mine for years, and she has both a great track record &amp;amp; actual influence in the world.&lt;/p&gt;
&lt;p&gt;&lt;a href=&quot;https://en.wikipedia.org/wiki/Audrey_Tang&quot;&gt;Audrey Tang&lt;/a&gt; is the Digital Minister of Taiwan. &lt;em&gt;She&lt;/em&gt; is the reason Taiwan is at the world&#39;s frontier of experiments in digital democracy, using humanely-designed tech to make government more open, accessible, accountable, responsive to the public&#39;s needs, and many other buzzwords that &lt;em&gt;would&lt;/em&gt; be empty if it weren&#39;t by Audrey Girlboss Tang, who frikkin&#39; &lt;em&gt;delivers&lt;/em&gt;.&lt;/p&gt;
&lt;p&gt;Anyway, her new thing is &lt;strong&gt;&lt;a href=&quot;https://6pack.care/&quot;&gt;6pack.care&lt;/a&gt;&lt;/strong&gt; (in collab with Caroline Green), a plan for human-AI alignment that (hopefully) works in both the short term (reversing democratic decay &amp;amp; extremism) and long term (humans living alongside powerful AIs).&lt;/p&gt;
&lt;p&gt;(I think Audrey&#39;s plan of making a long-term Alignment plan &amp;quot;pay dividends&amp;quot; in the short-term, is brilliant politics. The way she explained it: if you successfully advocated for reinforcing cockpit doors before 9/11, or pandemic resilience before Covid-19, your reward is... you see nothing happen. &amp;quot;No one cares about the bomb that didn’t go off, only the one that did.&amp;quot;&lt;sup class=&quot;footnote-ref&quot;&gt;&lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#fn17&quot; id=&quot;fnref17&quot;&gt;[17]&lt;/a&gt;&lt;/sup&gt; So 6pack&#39;s plan is to address problems people care about &lt;em&gt;now&lt;/em&gt; — algorithm-driven extremism, LLM-induced psychosis, social media mental health crises, etc — that also happen to scale to full Human-AI alignment.)&lt;/p&gt;
&lt;p&gt;Here&#39;s an overview of the 6 items of the 6pack of Digital Democracy, illustrated by... &lt;em&gt;me!&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2025-10/Overview_HiRez_NoHeader_Credits.png&quot; alt=&quot;Infographic of the 6Pack, in the form of a six-pack of beer: Actually listen to people, Actually keep promises, We check the process, We check the results, As win-win as possible, As local as possible.&quot; title=&quot;Infographic of the 6Pack, in the form of a six-pack of beer: Actually listen to people, Actually keep promises, We check the process, We check the results, As win-win as possible, As local as possible.&quot; /&gt;&lt;/p&gt;
&lt;p&gt;(I&#39;ve been commissioned to make 7 public-domain infographics in total, 1 for the Overview, 6 for each thing in the 6pack.)&lt;/p&gt;
&lt;p&gt;The 6pack is closely related to Vitalik Buterin&#39;s &lt;a href=&quot;https://vitalik.eth.limo/general/2025/02/28/aihumans.html&quot;&gt;&amp;quot;AI as the engine, humans as the steering wheel&amp;quot;&lt;/a&gt;, and the general &lt;a href=&quot;https://vitalik.eth.limo/general/2025/01/05/dacc2.html&quot;&gt;d/acc&lt;/a&gt; and &lt;a href=&quot;https://vitalik.eth.limo/general/2024/08/21/plurality.html&quot;&gt;Plurality&lt;/a&gt; movements. Here are the common shared principles, with specific tools:&lt;/p&gt;
&lt;ol&gt;
&lt;li&gt;&lt;strong&gt;Keep human values at the centre.&lt;/strong&gt;
&lt;ul&gt;
&lt;li&gt;concretely: bridging-based algorithms, linear/quadratic/score voting, frequent feedback loops, crowdsourced constitution/evals, &amp;quot;humans as steering wheel&amp;quot;, etc&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Set it up so that human power &lt;em&gt;scales with&lt;/em&gt; AI power, not get left behind.&lt;/strong&gt;
&lt;ul&gt;
&lt;li&gt;concretely: cyborgism (non-agentic AIs to augment human cognition, emotion, and collaboration), scaleable oversight, &amp;quot;AI as engine&amp;quot;, etc&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Make sure the tools are widely distributed, so that &lt;em&gt;power&lt;/em&gt; is distributed.&lt;/strong&gt;
&lt;ul&gt;
&lt;li&gt;concretely: decentralized algorithms like web-of-trust, peer-to-peer networks, or... &lt;em&gt;siggghh&lt;/em&gt;, blockchain. also, open-source software &lt;em&gt;and&lt;/em&gt; hardware, etc&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;/ol&gt;
&lt;p&gt;If that all still seems like jargon word salad to you... well, I got hired to illustrate 6 more pages for this thing, to explain each &amp;quot;pack&amp;quot; in lay-friendly detail. Stay tuned!&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;&lt;a href=&quot;https://6pack.care/&quot;&gt;Read the 6pack.care manifesto &amp;amp; outline online&lt;/a&gt;&lt;/strong&gt;&lt;br /&gt;
&lt;em&gt;(full book is supposed to come out in March 2026)&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;(P.S: Audrey doesn&#39;t want me calling it &amp;quot;The 6Pack of Digital Democracy&amp;quot;, and she&#39;s right, that&#39;s &lt;em&gt;not&lt;/em&gt; accurate — the 6pack is broader than that — but until we can come up with something catchier than &amp;quot;The 6Pack of Human-AI Multi-Agent Value Alignment&amp;quot; I&#39;m going to use the inaccurate alliteration, at least in this informal blog post.)&lt;/p&gt;
&lt;hr /&gt;
&lt;h2&gt;Fun Stuff!&lt;/h2&gt;
&lt;p&gt;&lt;a id=&quot;clues&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h3&gt;🕵️ Clues by Sam: a daily deductive detective game&lt;/h3&gt;
&lt;p&gt;I&#39;ve been hooked on this free daily puzzle game for the last month.&lt;/p&gt;
&lt;p&gt;Here&#39;s the setup. You&#39;re a detective. There are 20 people. Each one is either Innocent or Criminal. Everyone tells the truth, even Criminals.&lt;/p&gt;
&lt;p&gt;You start with &lt;em&gt;one&lt;/em&gt; person&#39;s clue:&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2025-10/clues-1.png&quot; alt=&quot;Screenshot from Clues by Sam, starting board&quot; title=&quot;Screenshot from Clues by Sam, starting board&quot; /&gt;&lt;/p&gt;
&lt;p&gt;Given the clues already on the board, you have to deduce a new Innocent or Criminal. &lt;em&gt;You cannot guess – the game knows which people&#39;s status is logically deduce-able at any time.&lt;/em&gt; Only when you correctly deduce a new person, do you get a new clue.&lt;/p&gt;
&lt;p&gt;For example:&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2025-10/clues-2.png&quot; alt=&quot;Screenshot from Clues by Sam, some deductions made&quot; title=&quot;Screenshot from Clues by Sam, some deductions made&quot; /&gt;&lt;/p&gt;
&lt;p&gt;And then:&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2025-10/clues-3.png&quot; alt=&quot;Screenshot from Clues by Sam, more deductions made&quot; title=&quot;Screenshot from Clues by Sam, more deductions made&quot; /&gt;&lt;/p&gt;
&lt;p&gt;And so on, until you&#39;ve figured out everybody!&lt;/p&gt;
&lt;p&gt;Like Sudoku, the first few times are tough, then you start learning some generalizable tricks — (in particular, the infamous &lt;a href=&quot;https://www.theguardian.com/science/2016/mar/28/did-you-solve-it-the-logic-question-almost-everyone-gets-wrong&quot;&gt;&amp;quot;if X then Y, if not-X then Y, therefore Y&amp;quot;&lt;/a&gt;) — and like Sudoku, after a while it &lt;em&gt;does&lt;/em&gt; get a bit repetitive, but it&#39;s still a nice mini-challenge to play on break.&lt;/p&gt;
&lt;p&gt;If that piques your interest, try out the 5-minute tutorial and get puzzlin&#39;, detective!&lt;/p&gt;
&lt;p&gt;🕵️‍♀️ &lt;strong&gt;&lt;a href=&quot;https://cluesbysam.com/&quot;&gt;Clues By Sam&lt;/a&gt;&lt;/strong&gt; 🕵️‍♂️&lt;/p&gt;
&lt;p&gt;&lt;a id=&quot;snakebird&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h3&gt;🐦 Snakebird: cute birbs, cruel puzzles&lt;/h3&gt;
&lt;p&gt;I actually played &amp;amp; finished this game at the start of 2025, when my laptop was taken at the US-Canada border and I couldn&#39;t do any work until I got a new laptop. (Customs &amp;amp; Border Patrol &lt;em&gt;did&lt;/em&gt; return my old laptop...  4 months later...)&lt;/p&gt;
&lt;p&gt;Anyway, I bring that up because Snakebird&#39;s is a great way to keep your mind off frustrating stuff you can do nothing about, and keep your mind on &lt;em&gt;very&lt;/em&gt; frustrating stuff you &lt;em&gt;can&lt;/em&gt; do something about!&lt;/p&gt;
&lt;p&gt;Snakebird, despite looking like a cutesy mobile game, is &lt;em&gt;infamous&lt;/em&gt; in the indie puzzle game community for being sadistically tough. It has a mid-3-star rating on the Apple App Store, because people complain about getting stuck on Level &lt;em&gt;Two&lt;/em&gt;. As for me, I&#39;m a puzzle aficionado, but it still took me a &lt;em&gt;month&lt;/em&gt; of playing ~1 hour a day to beat all 52 levels.&lt;/p&gt;
&lt;p&gt;(There &lt;em&gt;is&lt;/em&gt; an &amp;quot;easier&amp;quot; version called &lt;a href=&quot;https://noumenongames.com/game/snakebird_primer&quot;&gt;Snakebird Primer&lt;/a&gt;, and &lt;a href=&quot;https://www.nintendo.com/us/store/products/snakebird-complete-switch/&quot;&gt;Snakebird Complete&lt;/a&gt; contains both Primer + Original. I haven&#39;t tried them.)&lt;/p&gt;
&lt;p&gt;But while the game&#39;s tough, it&#39;s fair! It&#39;s always a matter of logical insight, not moon-logic clues or tedious trial-and-error. &amp;quot;Aha!&amp;quot;, not &amp;quot;oh that&#39;s bullshit&amp;quot;. (Okay, except for Level 26, that &lt;em&gt;was&lt;/em&gt; bullshit.)&lt;/p&gt;
&lt;p&gt;But... what &lt;em&gt;is&lt;/em&gt; Snakebird?&lt;/p&gt;
&lt;p&gt;It&#39;s like Snake meets Tetris: you slither &amp;amp; collect fruit to get longer like in Snake, but your snakes fall &lt;em&gt;instantly&lt;/em&gt; like in Tetris. Shenanigans ensue:&lt;/p&gt;
&lt;iframe width=&quot;640&quot; height=&quot;360&quot; src=&quot;https://www.youtube-nocookie.com/embed/WJJa3hMbs4s?rel=0&quot; title=&quot;YouTube video player&quot; frameborder=&quot;0&quot; allow=&quot;accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share&quot; referrerpolicy=&quot;strict-origin-when-cross-origin&quot; allowfullscreen=&quot;&quot;&gt;&lt;/iframe&gt;
&lt;p&gt;&amp;quot;That doesn&#39;t look too bad&amp;quot;, you think. Ha ha. Ha ha ha. Oh sweet summer child&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;&lt;a href=&quot;https://noumenongames.com/game/snakebird&quot;&gt;Snakebird&lt;/a&gt;&lt;/strong&gt;, available on iOS, Android, Mac, Windows, Linux&lt;/li&gt;
&lt;li&gt;&lt;a href=&quot;https://noumenongames.com/game/snakebird_primer&quot;&gt;Snakebird Primer&lt;/a&gt; also on all those platforms&lt;/li&gt;
&lt;li&gt;&lt;a href=&quot;https://www.nintendo.com/us/store/products/snakebird-complete-switch/&quot;&gt;Snakebird Complete&lt;/a&gt; only on Nintendo Switch&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;(hat tip: I first heard about Snakebird from Game Maker&#39;s Tool Kit&#39;s video on &lt;a href=&quot;https://www.youtube.com/watch?v=zsjC6fa_YBg&quot;&gt;how to design a tough-but-fair puzzle games&lt;/a&gt;)&lt;/p&gt;
&lt;p&gt;&lt;a id=&quot;zewei&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h3&gt;🦉 ZeWei&#39;s Multiverse Tour Guide Adventures Continue&lt;/h3&gt;
&lt;p&gt;&lt;a href=&quot;https://blog.ncase.me/signal-boosts-oct-2024/#zewei&quot;&gt;Last year&lt;/a&gt;, I signal-boosted the nonbinary furry EDM musician ZeWei&#39;s animated mockumentary of a linguist trapped in an alien world! Well, it&#39;s a full series now!&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Episode 2: &amp;quot;It&#39;s not colonialism, it&#39;s tourism!&amp;quot; /s (17 min)&lt;/strong&gt;&lt;/p&gt;
&lt;iframe width=&quot;640&quot; height=&quot;360&quot; src=&quot;https://www.youtube-nocookie.com/embed/WScut_fmFzw?rel=0&quot; title=&quot;YouTube video player&quot; frameborder=&quot;0&quot; allow=&quot;accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share&quot; referrerpolicy=&quot;strict-origin-when-cross-origin&quot; allowfullscreen=&quot;&quot;&gt;&lt;/iframe&gt;
&lt;p&gt;In this episode, our linguist protagonist is worried their world&#39;s language &amp;amp; culture is erasing those of the worlds they travel to. Just when you think the story&#39;s headed towards the standard &amp;quot;we must preserve the noble savages&#39; linguistic diversity&amp;quot;, the native of this other world calls them on their bullshit, that, no, &lt;em&gt;they have autonomy&lt;/em&gt;, they &lt;em&gt;chose&lt;/em&gt; to drop the worst parts of their language &amp;amp; culture, and carefully &lt;em&gt;chose&lt;/em&gt; parts of other cultures to adopt, and merged them into something uniquely their own. They don&#39;t owe it to anyone to &amp;quot;preserve&amp;quot; themselves like a living fossil. They don&#39;t live for historians or researchers or nostalgia, they live &lt;em&gt;for themselves&lt;/em&gt;.&lt;/p&gt;
&lt;p&gt;(But, yes... something &lt;em&gt;was&lt;/em&gt; lost. &amp;quot;It&#39;s complicated.&amp;quot;)&lt;/p&gt;
&lt;p&gt;This episode resonated with me, since I was born in Singapore (a very post-British-colonial city-state), then I immigrated to Vancouver Canada halfway during my childhood. My backstory is a mix of Western &amp;amp; Eastern influences. I&#39;d like to believe I&#39;ve chosen the better parts of both cultures, and dropped the unhealthy crap from both — but who knows — if I lost something, &lt;em&gt;would I even recognize the loss?&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Episode 3: &amp;quot;The Lore Episode&amp;quot; (30 min!!!)&lt;/strong&gt;&lt;/p&gt;
&lt;iframe width=&quot;640&quot; height=&quot;360&quot; src=&quot;https://www.youtube-nocookie.com/embed/RbEFhJ2z-7U?rel=0&quot; title=&quot;YouTube video player&quot; frameborder=&quot;0&quot; allow=&quot;accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share&quot; referrerpolicy=&quot;strict-origin-when-cross-origin&quot; allowfullscreen=&quot;&quot;&gt;&lt;/iframe&gt;
&lt;p&gt;This one&#39;s less head-y, and a lot more establishing this story&#39;s world. We finally get to see the other characters &amp;quot;behind the camera&amp;quot; of this mockumentary series! There&#39;s also a prophecy.&lt;/p&gt;
&lt;p&gt;Less for me to hook onto in this episode, but it&#39;s clearly setting up for a big series arc. Looking forward to seeing where the arc bends next!&lt;/p&gt;
&lt;p&gt;&lt;a id=&quot;emnerson&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h3&gt;🤓 Emnerson: goofy internet gal&lt;/h3&gt;
&lt;p&gt;A new, up-and-coming YouTuber who makes a variety of cool digital projects!&lt;/p&gt;
&lt;p&gt;The way I first learnt about her YouTube channel, is because she won Captain Disillusion&#39;s &amp;quot;unblur this image&amp;quot; challenge, cracking it within 20 minutes:&lt;/p&gt;
&lt;iframe width=&quot;640&quot; height=&quot;360&quot; src=&quot;https://www.youtube-nocookie.com/embed/Yi-Vea2AUEU?rel=0&quot; title=&quot;YouTube video player&quot; frameborder=&quot;0&quot; allow=&quot;accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share&quot; referrerpolicy=&quot;strict-origin-when-cross-origin&quot; allowfullscreen=&quot;&quot;&gt;&lt;/iframe&gt;
&lt;p&gt;Here&#39;s her latest video on merging YouTube channels&#39; thumbnails, and seeing what patterns pop up:&lt;/p&gt;
&lt;iframe width=&quot;640&quot; height=&quot;360&quot; src=&quot;https://www.youtube-nocookie.com/embed/g-1kUvJqagw?rel=0&quot; title=&quot;YouTube video player&quot; frameborder=&quot;0&quot; allow=&quot;accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share&quot; referrerpolicy=&quot;strict-origin-when-cross-origin&quot; allowfullscreen=&quot;&quot;&gt;&lt;/iframe&gt;
&lt;p&gt;Also, she happens to be trans, and I want to help boost my fellow up-and-coming trans creators. Everyone including me is very jealous of how fast she&#39;s transitioned in a few months and she is very pretty.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Check her out! &lt;a href=&quot;https://www.youtube.com/@emnersonn&quot;&gt;Emnerson&#39;s YouTube channel&lt;/a&gt;&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;&lt;a id=&quot;blahaj&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h3&gt;🧇 blahaj goes to waffle house at 3am&lt;/h3&gt;
&lt;p&gt;Cute as hell. Watch it, and subscribe to a new, up-and-coming animator!&lt;/p&gt;
&lt;iframe width=&quot;640&quot; height=&quot;360&quot; src=&quot;https://www.youtube-nocookie.com/embed/X1beEuBV7M0?rel=0&quot; title=&quot;YouTube video player&quot; frameborder=&quot;0&quot; allow=&quot;accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share&quot; referrerpolicy=&quot;strict-origin-when-cross-origin&quot; allowfullscreen=&quot;&quot;&gt;&lt;/iframe&gt;
(they also did all the original music, and are based in Taiwan)
&lt;p&gt;&lt;strong&gt;Link: &lt;a href=&quot;https://www.youtube.com/@AtogaCreative&quot;&gt;Atoga&#39;s YouTube channel&lt;/a&gt;&lt;/strong&gt;&lt;/p&gt;
&lt;hr /&gt;
&lt;h2&gt;SPOOKY STUFF for HALLOWEEN&lt;/h2&gt;
&lt;p&gt;&lt;a id=&quot;boo&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h3&gt;👻 BOO! a silly music video &amp;amp; music album&lt;/h3&gt;
&lt;p&gt;A 80&#39;s-throwback musical comedy of a tiny ghost trying to be spooky:&lt;/p&gt;
&lt;iframe width=&quot;640&quot; height=&quot;360&quot; src=&quot;https://www.youtube-nocookie.com/embed/NZx9fAk22Qg?rel=0&quot; title=&quot;YouTube video player&quot; frameborder=&quot;0&quot; allow=&quot;accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share&quot; referrerpolicy=&quot;strict-origin-when-cross-origin&quot; allowfullscreen=&quot;&quot;&gt;&lt;/iframe&gt;
&lt;p&gt;(The creator, &lt;a href=&quot;https://www.youtube.com/@Piemations/videos&quot;&gt;Piemations&lt;/a&gt;, usually does animated stuff! They made Sheriff Hayseed, Bird Town, Suction Cup Man, Mike &amp;amp; Zach, and How To Be Cool, if you recognize any of those titles.)&lt;/p&gt;
&lt;p&gt;Oh, and the song &amp;quot;BOO!&amp;quot; is just one track from his band&#39;s new album, &lt;a href=&quot;https://friesontheside.bandcamp.com/album/musical-scares&quot;&gt;Musical Scares&lt;/a&gt;. The track Big Doggie is my favourite. Comedy aside, it &lt;em&gt;actually bops&lt;/em&gt;. The intro drop hits so hard, I must&#39;ve restarted the song a dozen times just to hear that drop again. And I haven&#39;t heard the phrase &amp;quot;14 werewolves&amp;quot; in years, got &lt;em&gt;instant&lt;/em&gt; psychic damage from that.&lt;/p&gt;
&lt;iframe style=&quot;border: 0; width: 640px; height: 472px;&quot; src=&quot;https://bandcamp.com/EmbeddedPlayer/album=2742795446/size=large/bgcol=333333/linkcol=e32c14/artwork=small/transparent=true/&quot; seamless=&quot;&quot;&gt;&lt;a href=&quot;https://friesontheside.bandcamp.com/album/musical-scares&quot;&gt;Musical Scares by Fries On The Side&lt;/a&gt;test&lt;/iframe&gt;
&lt;a id=&quot;bury-your-gays&quot;&gt;&lt;/a&gt;
&lt;h3&gt;🩸 Bury Your Gays by Chuck Tingle&lt;/h3&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2025-10/Pasted%20image%2020251029165354.png&quot; alt=&quot;Cover of the book Bury Your Gays by Chuck Tingle&quot; title=&quot;Cover of the book Bury Your Gays by Chuck Tingle&quot; /&gt;&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;&lt;a href=&quot;https://bookshop.org/p/books/bury-your-gays-chuck-tingle/de2c90d5aa0bc158&quot;&gt;A novel by Chuck Tingle&lt;/a&gt;&lt;/strong&gt;, yes &lt;em&gt;that&lt;/em&gt; Chuck Tingle, the &amp;quot;got famous by writing dozens of joke eroticas&amp;quot; Mr. Tingle:&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2025-10/Pasted%20image%2020251029165749.png&quot; alt=&quot;Pounded In The Butt By My Bizarre Assumption That Chuck Tingle Books Are Just Covers And Not Actual Books&quot; title=&quot;&amp;quot;Pounded In The Butt By My Bizarre Assumption That Chuck Tingle Books Are Just Covers And Not Actual Books&amp;quot;&quot; /&gt;&lt;/p&gt;
&lt;p&gt;A couple years ago, Chuck Tingle expanded to novels, including Bury Your Gays, which is &lt;em&gt;shockingly good.&lt;/em&gt; And it&#39;s very meaningful for (gestures vaguely) &amp;quot;the current moment&amp;quot;.&lt;/p&gt;
&lt;p&gt;The premise: Misha Byrne is a scriptwriter for a X-Files-like series with two (subtextually) gay protagonists. It&#39;s a hit. Misha&#39;s called into a meeting. His boss tells him what the executives tell him the Algorithm tells them would make the most money: have the leads profess their gay love out loud, &lt;em&gt;then immediately kill them off.&lt;/em&gt; The &amp;quot;queer tragedy&amp;quot; plot, the &amp;quot;bury your gays&amp;quot; trope, that&#39;s the controversy &amp;amp; drama that would sell! Oh, and if Misha &lt;em&gt;doesn&#39;t&lt;/em&gt; kill them off, that&#39;s breaking contract, he&#39;ll get sued into oblivion, then the studio will take his show &amp;amp; kill off the gay leads anyway.&lt;/p&gt;
&lt;p&gt;Misha walks away from the meeting pissed, angry at the execs, the Algorithm, the whole damn world. And just when Misha thinks things can&#39;t get any worse — 5 minutes later — a colleague explodes into meat and gore in front of him.&lt;/p&gt;
&lt;p&gt;Then things get &lt;em&gt;really&lt;/em&gt; bad.&lt;/p&gt;
&lt;p&gt;So, without spoilers and in no particular order, why Bury Your Gays resonated with me so much:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;There&#39;s (understandably) a lot of stories of queer tragedy right now, both in fiction and real life. This novel is a reminder that &lt;strong&gt;happiness is not &amp;quot;naïve&amp;quot;, or &amp;quot;basic&amp;quot;, or dumb.&lt;/strong&gt; Without over-correcting into toxic positivity, &lt;strong&gt;we can and should also celebrate queer joy.&lt;/strong&gt; Misha&#39;s story (both the story he writes &lt;em&gt;and&lt;/em&gt; the story he&#39;s living through) show that full complexity, lows &lt;em&gt;and&lt;/em&gt; highs, of a queer life.&lt;/li&gt;
&lt;li&gt;As a &amp;quot;creator on the internet&amp;quot;, I definitely feel a lot of anxiety about opaque algorithms &amp;amp; generative AI screwing not just with my creations, but &lt;em&gt;my own personal character&lt;/em&gt;. Y&#39;know, &amp;quot;audience capture&amp;quot;, &amp;quot;we become what we pretend to be so we must be careful what we pretend to be&amp;quot;, etc&lt;/li&gt;
&lt;li&gt;The novel also reminds queer creators (like me) to help lift up other less-famous queer creators. Which sounds obvious, but when you&#39;re dealing with your own shit, it&#39;s easy to forget to help others too. So, that&#39;s something I&#39;d like to do with these Signal Boosts.
&lt;ul&gt;
&lt;li&gt;(I mean, Chuck Tingle &amp;amp; Audrey Tang are far more famous than me, but (see above) ZeWei &amp;amp; Emnerson more recent, smaller-fanbase creators.)&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;Oh! Misha&#39;s best friend is asexual &amp;amp; aromantic! Hooray for ace/aro representation.&lt;/li&gt;
&lt;li&gt;I just love books that get all weird and meta and format-screw-y. (Could you guess I read &lt;a href=&quot;https://thehardtimes.net/blog/opinion-house-of-leaves-is-just-infinite-jest-for-spooky-people/&quot;&gt;House of Leaves&lt;/a&gt; as a teen and it changed my brain chemistry?)&lt;/li&gt;
&lt;li&gt;The parallel between Misha&#39;s characters being &amp;quot;only&amp;quot; subtextually gay, and Misha &lt;em&gt;himself&lt;/em&gt; &amp;quot;only&amp;quot; being semi-out of the closet, is clever, and hits personally in two different ways:
&lt;ol&gt;
&lt;li&gt;The line between &amp;quot;using art to work through my own feelings&amp;quot; and &amp;quot;using art to &lt;em&gt;avoid&lt;/em&gt; doing the emotional work in real life&amp;quot; is thin, and I struggle with this too.&lt;/li&gt;
&lt;li&gt;I&#39;m like... &lt;em&gt;out&lt;/em&gt; about being pansexual &amp;amp; transgender, but I feel like I could be &lt;em&gt;more&lt;/em&gt; out? I don&#39;t mention it or have 🏳️‍⚧️🏳️‍🌈 flags on my social media bios. (Well, I don&#39;t &lt;em&gt;have&lt;/em&gt; any social media anymore, but still.) Right now, I&#39;m doing the strategy of &amp;quot;let people know me for my educational work first, then &lt;em&gt;later&lt;/em&gt; learn I&#39;m trans, so it&#39;s not the first adjective they know about me&amp;quot;.
&lt;ul&gt;
&lt;li&gt;On one hand: me being more out &amp;amp; confident could give inspiration to the many queer folks who need hope right now.&lt;/li&gt;
&lt;li&gt;On the other hand, living true to yourself will draw the attention of people who don&#39;t. Bucket-crabs who pull you down into the Normal Bucket, wordcels sophisticatedly rationalizing their Pavlovian disgust-reflexes as the &amp;quot;wisdom of repugnance&amp;quot;. And I hate to admit it, but I &lt;em&gt;am&lt;/em&gt; irrationally averse to &amp;quot;being confrontational&amp;quot;, like &amp;quot;I don&#39;t want to make a big deal of it&amp;quot;, like &amp;quot;oh they just need time to understand, I won&#39;t put it in their face, hurt people hurt people, nobody&#39;s the villain of their own story, everyone has a complex rich inner life&amp;quot; and other copium fucking &lt;em&gt;bullshit&lt;/em&gt;. And it&#39;s not for lack of trying or skills. I am &lt;em&gt;telling&lt;/em&gt; you, Reader, after 10+ years of me studying &amp;amp; practicing &amp;amp; applying Non Violent Communication and Active Listening and Steelmaning and Intellectual Turing Tests and Moral Foundations Theory, years of trying to &lt;em&gt;really get to know people&lt;/em&gt; behind their mere politics, their &lt;em&gt;souls&lt;/em&gt;, Reader, I am &lt;em&gt;reporting&lt;/em&gt; to you: &lt;em&gt;it was always all bullshit.&lt;/em&gt; People are like onions: as you keep peeling back, you discover that every layer is the fucking same, and then you start crying. Anecdotes of personal change are survivor bias, &amp;quot;intricately layered characters&amp;quot; was invented by Big Book to sell more Books. Humanism is a cult: ~32% of Germans thought Hitler was a great man &lt;em&gt;in 1952, after Hitler already lost&lt;/em&gt; and the Holocaust was globally known&lt;sup class=&quot;footnote-ref&quot;&gt;&lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#fn18&quot; id=&quot;fnref18&quot;&gt;[18]&lt;/a&gt;&lt;/sup&gt;; ~15% of eligible Americans were in the Ku Klux Klan in the 1920s&lt;sup class=&quot;footnote-ref&quot;&gt;&lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#fn19&quot; id=&quot;fnref19&quot;&gt;[19]&lt;/a&gt;&lt;/sup&gt;; ~28% of Hutu adult civilians &lt;em&gt;personally participated&lt;/em&gt; in the genocide against the Tutsis&lt;sup class=&quot;footnote-ref&quot;&gt;&lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#fn20&quot; id=&quot;fnref20&quot;&gt;[20]&lt;/a&gt;&lt;/sup&gt;; just give up the copium and apply Occam&#39;s Razor and finally admit that &lt;em&gt;the reason mass atrocities keep happening, is simply because the masses are atrocious&lt;/em&gt;. Did your eyes automatically roll? That&#39;s how they get you: mark the thoughts not as &amp;quot;dark&amp;quot; or &amp;quot;evil&amp;quot;, but just &amp;quot;cringe&amp;quot;. &lt;em&gt;That&#39;s&lt;/em&gt; how you censor a thought. All these so-called empathetic Humanist thought-leaders will try to guilt-trip you back into loving humanity. &amp;quot;But they&#39;re not all bad, there&#39;s good parts in them&amp;quot; isn&#39;t wisdom, &lt;em&gt;it&#39;s how abusers gaslight their victims into staying with them.&lt;/em&gt; Humanity &lt;em&gt;is&lt;/em&gt; the abuser. Humanity does not Spark Joy. It&#39;s not the Algorithm, or late capitalism, or Moloch, or woke mind viruses, or egregores, &lt;em&gt;IT&#39;S YOU.&lt;/em&gt; It&#39;s &lt;em&gt;you.&lt;/em&gt;&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;/ol&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;...right, clearly I&#39;m still working through some issues.&lt;/p&gt;
&lt;p&gt;I thank Mr. Pounded In The Butt for forcing me to work through it more.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;&lt;a href=&quot;https://bookshop.org/p/books/bury-your-gays-chuck-tingle/de2c90d5aa0bc158&quot;&gt;Bury Your Gays&lt;/a&gt; on IndieBound/Bookshop&lt;/strong&gt; (content note: gore, homophobia)&lt;/p&gt;
&lt;hr /&gt;
&lt;p&gt;Anyway, that was a longer-than-usual Signal Boost! In summary:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;There&#39;s hope and solutions to wrangle The Algorithms back towards a humane-with-an-e future&lt;/li&gt;
&lt;li&gt;Have fun, and&lt;/li&gt;
&lt;li&gt;Happy Halloween!&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;🎃👻🎃&lt;br /&gt;
~ Nicky Case&lt;/p&gt;
&lt;iframe src=&quot;https://ncase.me/ncase-credits/signup2.html&quot; frameborder=&quot;no&quot; width=&quot;640&quot; height=&quot;250&quot;&gt;&lt;/iframe&gt;
&lt;iframe src=&quot;https://ncase.me/ncase-credits/supporters/oct-2025.html&quot; frameborder=&quot;no&quot; width=&quot;640&quot; height=&quot;640&quot;&gt;&lt;/iframe&gt;
&lt;hr /&gt;
&lt;h1&gt;: Interview with Jon&lt;/h1&gt;
&lt;ul&gt;
&lt;li&gt;I hadn&#39;t before considered the moral cost &lt;em&gt;to the user&lt;/em&gt; of having on-demand, disposable AI servants. Haidt points out what Frederick Douglass once called &amp;quot;The Fatal Poison of Irresponsible Power&amp;quot;, where slavery not only degrades the slave, but &lt;em&gt;degrades the slaveowner too.&lt;/em&gt; If we grow up with people-like objects we treat as tools, how will this train us to treat actual people?&lt;/li&gt;
&lt;li&gt;Yup, &lt;a href=&quot;https://www.brainonllm.com/&quot;&gt;a recent MIT preprint study&lt;/a&gt; has found: using ChatGPT &lt;em&gt;literally&lt;/em&gt; makes you &lt;s&gt;dumber&lt;/s&gt; okay the authors are &lt;a href=&quot;https://www.brainonllm.com/faq&quot;&gt;very adamant&lt;/a&gt; we science-communicators &lt;em&gt;do not&lt;/em&gt; summarize their paper as &amp;quot;LLMs make you dumber&amp;quot;. The actual contents of their study: students who use LLMs to help write essays have measurably less brain connectivity, relative to Brain-only studetns. And after 4 months, LLM users &amp;quot;consistently underperformed at neural, linguistic, and behavioral levels.&amp;quot;
&lt;ul&gt;
&lt;li&gt;The reason why it&#39;s not as simple as &amp;quot;LLMs make you dumb&amp;quot;, is because the group that &lt;em&gt;first&lt;/em&gt; wrote on a topic using only their naked brain, &lt;em&gt;THEN&lt;/em&gt; were allowed to use LLMs (&lt;strong&gt;Brain-then-LLM&lt;/strong&gt; group) actually did &lt;em&gt;better:&lt;/em&gt; &amp;quot;these results suggest that strategic timing of AI tool introduction following initial self-driven effort may enhance engagement and neural integration.&amp;quot;&lt;/li&gt;
&lt;li&gt;In contrast, the &lt;strong&gt;LLM-then-Brain&lt;/strong&gt; group &amp;quot;consistently underperformed relative to Session 2 of Brain-only group, and failed to develop the consolidation networks present in Session 3 of Brain-only group.&amp;quot;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;In other words: it may be best to use tools, only after you know how to work &lt;em&gt;without&lt;/em&gt; the tools.&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;(Keep in mind the study only had 54 participants. And it&#39;s a &lt;em&gt;lot&lt;/em&gt; of analysis over 54 participants, which, I&#39;ll be honest, smells fishy.)&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;h1&gt;: Interview with Tim&lt;/h1&gt;
&lt;ul&gt;
&lt;li&gt;If you&#39;re already familiar with Tim&#39;s work, in particular his latest book &lt;em&gt;What&#39;s Our Problem&lt;/em&gt;, you can skip ahead in the interview to 40:18 . Otherwise, &lt;a href=&quot;https://xcancel.com/ncasenmare/status/1627798034439450624&quot;&gt;here was my book summary (in tweet-thread form!)&lt;/a&gt;  To summarize &lt;em&gt;that&lt;/em&gt; summary: &lt;strong&gt;the big problem with politics right now is not &lt;em&gt;what&lt;/em&gt; we believe, but &lt;em&gt;how&lt;/em&gt; we believe.&lt;/strong&gt;
&lt;ul&gt;
&lt;li&gt;&amp;quot;Low-rung&amp;quot; thinkers treat politics like a war, where arguments are bullets, and you&#39;re either for xor against us. You&#39;d never critique your own side or give credit to the other side, any more than you&#39;d shoot your own platoon or hand ammunition to the enemy.&lt;/li&gt;
&lt;li&gt;&amp;quot;High-rung&amp;quot; thinkers treat politics like a Writer&#39;s Room, or Code Review. It&#39;s &lt;em&gt;NOT&lt;/em&gt; that &amp;quot;truth is in the middle&amp;quot; (strawman centrist) or &amp;quot;everyone has their own truth&amp;quot; (strawman postmodernist) or even &amp;quot;everyone has a valuable perspective&amp;quot; (ha ha have you met people). It&#39;s about &lt;em&gt;ACTUALLY TESTING YOUR IDEAS&lt;/em&gt;. You wouldn&#39;t let into your home an oven that&#39;s never been safety-tested. So why let an idea into your core beliefs that&#39;s never been stress-tested? Pick holes &amp;amp; find counterarguments to ideas even if they&#39;re &amp;quot;from your own side&amp;quot;, search from good critique &amp;amp; solutions even if they&#39;re &amp;quot;from the enemy&#39;s side&amp;quot;, and &lt;em&gt;especially&lt;/em&gt; try to find &amp;amp; invent new ideas that&#39;s on &lt;em&gt;nobody&#39;s&lt;/em&gt; side.&lt;/li&gt;
&lt;li&gt;&amp;quot;High-rung&amp;quot; is &lt;em&gt;not&lt;/em&gt; intelligence. There are very high-IQ people who squander their IQ on rationalizing what their gut/tribe already believes.&lt;/li&gt;
&lt;li&gt;There are high-rung people on the left &amp;amp; right. &lt;em&gt;There are low-rung people who are &amp;quot;centrist&amp;quot; or &amp;quot;moderate&amp;quot;&lt;/em&gt;. (e.g. the kind of smug centrist who only focuses on extremists on the left &amp;amp; right, and keeps talking about &#39;horseshoe theory&#39;)&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;There&#39;s a pattern in history: New Media Technology =&amp;gt; Lots of benefit but lots of awful side-effects =&amp;gt; Society slowly develops &amp;quot;cultural anti-bodies&amp;quot;, to get the benefits of the new tech while minimizing the downsides.
&lt;ul&gt;
&lt;li&gt;Historic example: The printing press indirectly led to Enlightenment &amp;amp; Scientific Revolution... but &lt;em&gt;also&lt;/em&gt; the Protestant vs Catholic split and the centuries of murderous conflict.&lt;/li&gt;
&lt;li&gt;Current example: Social media let people coordinate to donate millions to charity, and reveal injustice that the powerful would rather keep hidden... but &lt;em&gt;also&lt;/em&gt; tanked our politics, mental health, and attention/autonomy.&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;Audience capture is when you, the creator, get accidentally clicker-trained by the likes &amp;amp; comments from your audience, into becoming a parody of yourself. Before Tim&#39;s interview, I thought the solution was &amp;quot;just don&#39;t get captured&amp;quot;. But Tim offers a better solution: work &lt;em&gt;with&lt;/em&gt; the incentives, not against them, &lt;strong&gt;by getting captured by the &lt;em&gt;right&lt;/em&gt; audience.&lt;/strong&gt;
&lt;ul&gt;
&lt;li&gt;First, know what you value. Then, find an audience that values what you value. Then, &lt;em&gt;write for the audience you want, not the audience you can easily get.&lt;/em&gt; This will mean having to go through a rough period where either few see your work, and/or you get fans &amp;amp; even friends being mad at you — but power through this, they&#39;ll filter themselves out, and &lt;strong&gt;you&#39;ll get an audience that will help incentivize you to be the best version of you.&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;(This was the biggest insight I took from Tim&#39;s interview, and it&#39;s one I&#39;ve not yet applied &amp;amp; am scared to. By default my personality is non-confrontational. I&#39;m scared to piss off my fans &amp;amp; friends, or worse, &lt;em&gt;bore&lt;/em&gt; them.)&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;h1&gt;: Notes on Polarization&lt;/h1&gt;
&lt;p&gt;&lt;strong&gt;Not all polarization is bad&lt;/strong&gt;, some&#39;s even desirable. Researchers name (at least) &lt;strong&gt;3 different kinds of &amp;quot;polarization&amp;quot;: affective polarization, belief polarization, and homophily.&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Affective polarization: We hate them, they hate us.&lt;/li&gt;
&lt;li&gt;Belief polarization: We &amp;amp; they have very different beliefs.&lt;/li&gt;
&lt;li&gt;Homophily: Most of my friends are in my tribe, most of theirs in theirs.&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;Homophily&#39;s both unavoidable, and actually &lt;em&gt;desirable&lt;/em&gt;, in my opinion. For example, only ~0.6% of people are programmers like me, but &amp;gt;50+% of my friends are programmers. It wouldn&#39;t make personal or career sense for me to have exactly demographically-proportional representation in my friend group. Same for &amp;gt;50+% of my friends being LGBTQ+, left-leaning, urban, English-speaking, and nerdy.&lt;/p&gt;
&lt;p&gt;However, if &lt;em&gt;all&lt;/em&gt; of my friends (or media sources) are urban, or in STEM, or left-leaning, then I&#39;ll be out of touch with the broader world, and &lt;em&gt;that&lt;/em&gt; may lead to affective/belief polarization.&lt;/p&gt;
&lt;p&gt;And what&#39;s the difference between affective &amp;amp; belief polarization, concretely?&lt;/p&gt;
&lt;p&gt;Affective without belief polarization: When people despise each other for having pretty small differences in beliefs (at least, relative to their society at large). Historic example: All the deadly wars between Protestants &amp;amp; Catholics. Today&#39;s examples: leftists vs liberals, Groypers vs New Right.&lt;/p&gt;
&lt;p&gt;Belief without affective polarization: When people get along despite having radically different beliefs. Examples: Friendships between Christians &amp;amp; atheists, friendships between AI capabilities researchers &amp;amp; AI extinction-risk researchers.&lt;/p&gt;
&lt;p&gt;Note: I don&#39;t believe that hating someone is necessarily &amp;quot;bad&amp;quot;; anger can be justified &amp;amp; useful – but anger &lt;em&gt;is&lt;/em&gt; known to be addictive. Kind of like amphetamines. I also don&#39;t think belief polarization, a lack of consensus, is necessarily &amp;quot;bad&amp;quot;; a wide variety of ideas helps us brainstorm better, and often there &lt;em&gt;isn&#39;t&lt;/em&gt; enough data for one clear answer. My current take is:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Homophily is mostly good. Just make sure not &lt;em&gt;all&lt;/em&gt; of your friends/media sources are in the same tribe. That&#39;s a single point of failure.&lt;/li&gt;
&lt;li&gt;Belief polarization is not good or bad, it just &amp;quot;is&amp;quot;. Focus instead on cultivating the intellectual virtues of curiosity balanced with rigour. If that leads to consensus, great. If that leads to disagreement, great. Trust the process.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Affective polarization (anger) is like amphetamines: focusing, energizing, sometimes useful, but unfortunately easy to abuse &amp;amp; get addicted.&lt;/strong&gt;&lt;/li&gt;
&lt;/ul&gt;
&lt;hr class=&quot;footnotes-sep&quot; /&gt;
&lt;section class=&quot;footnotes&quot;&gt;
&lt;ol class=&quot;footnotes-list&quot;&gt;
&lt;li id=&quot;fn1&quot; class=&quot;footnote-item&quot;&gt;&lt;p&gt;This book personally changed my life, turning me from Mid-10&#39;s Smug Asshole Atheist to &amp;quot;atheist but at least &lt;em&gt;try&lt;/em&gt; to be kind &amp;amp; understand beliefs very unlike mine.&amp;quot; The core ideas of the book — everyone&#39;s beliefs are mostly downstream of emotional reflexes, liberals/conservatives have overlapping but different emotional-reflexes — are still on solid scientific ground, though, be warned that since this is a psych book from 2012, probably most of the &lt;em&gt;specific&lt;/em&gt; studies he cites may not replicate. &lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#fnref1&quot; class=&quot;footnote-backref&quot;&gt;↩︎&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li id=&quot;fn2&quot; class=&quot;footnote-item&quot;&gt;&lt;p&gt;&lt;a href=&quot;https://papers.ssrn.com/sol3/Delivery.cfm?abstractid=1528077&quot;&gt;What makes online content viral?&lt;/a&gt; by Berger &amp;amp; Milkman (2012) analyzed all headlines from the New York Times over a 3-month period. As seen in Figure 2, the top factor that makes a headline viral, is &amp;quot;Anger&amp;quot;. (To be fair, the close runner-ups are &amp;quot;Awe&amp;quot; and &amp;quot;Practical Value&amp;quot;) &lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#fnref2&quot; class=&quot;footnote-backref&quot;&gt;↩︎&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li id=&quot;fn3&quot; class=&quot;footnote-item&quot;&gt;&lt;p&gt;“The model centers on a population of simulated users, each represented by a persona drawn from the American National Election Studies (ANES) dataset. These personas reflect real-world distributions of age, gender, income, education, partisanship, ideology, religion, and personal interests.” &lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#fnref3&quot; class=&quot;footnote-backref&quot;&gt;↩︎&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li id=&quot;fn4&quot; class=&quot;footnote-item&quot;&gt;&lt;p&gt;Main analysis was done with GPT-4o-mini, results replicated successfully with llama-3.2-8b and DeepSeek-R1 &lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#fnref4&quot; class=&quot;footnote-backref&quot;&gt;↩︎&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li id=&quot;fn5&quot; class=&quot;footnote-item&quot;&gt;&lt;p&gt;The &amp;quot;baseline&amp;quot; simulated social network has the following algorithmic feed: it shows each user 10 posts: “five from followed users and five drawn from high-engagement content posted by non-followed users, with repost probability used as a proxy for algorithmic amplification.” All interventions are tested compared to this baseline. &lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#fnref5&quot; class=&quot;footnote-backref&quot;&gt;↩︎&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li id=&quot;fn6&quot; class=&quot;footnote-item&quot;&gt;&lt;p&gt;The paper doesn&#39;t offer a detailed hypothesis, and without re-running the simulation myself I&#39;d just be guessing, but let me guess: {chronological feed + repost mechanism} is actually a &lt;em&gt;stronger filter for engaging/enraging posts&lt;/em&gt; than {algorithmic feed}. Let&#39;s say the top ten posts in your chronological feed are posts from the last half hour. These posts will be either a) originally posted in the last 30 min, or b) &lt;em&gt;so engaging&lt;/em&gt; that your ingroup has been reposting it &lt;em&gt;at least once per half hour&lt;/em&gt;. In contrast, a baseline algorithmic feed would highlight posts that have gotten lots of reposts in the last, say, 24 hours. &lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#fnref6&quot; class=&quot;footnote-backref&quot;&gt;↩︎&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li id=&quot;fn7&quot; class=&quot;footnote-item&quot;&gt;&lt;p&gt;From &lt;a href=&quot;https://docs.bsky.app/docs/tutorials/viewing-feeds&quot;&gt;their API&lt;/a&gt;: &amp;quot;the default chronological feed of posts from users the authenticated user follows&amp;quot; &lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#fnref7&quot; class=&quot;footnote-backref&quot;&gt;↩︎&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li id=&quot;fn8&quot; class=&quot;footnote-item&quot;&gt;&lt;p&gt;From &lt;a href=&quot;https://help.truthsocial.com/truth-social-common-terminology/glossary-home-feed/&quot;&gt;their FAQ&lt;/a&gt;: &amp;quot;we refrain from using non-chronological feeds to artificially suppress users&#39; content.&amp;quot; &lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#fnref8&quot; class=&quot;footnote-backref&quot;&gt;↩︎&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li id=&quot;fn9&quot; class=&quot;footnote-item&quot;&gt;&lt;p&gt;&amp;quot;14.9% use AI for emotional support daily, with an &lt;strong&gt;additional&lt;/strong&gt; 27.9% weekly&amp;quot; {emphasis added}. So that&#39;s 14.9 + 27.9 = 42.8% using AI as emotional support at least once a week. &lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#fnref9&quot; class=&quot;footnote-backref&quot;&gt;↩︎&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li id=&quot;fn10&quot; class=&quot;footnote-item&quot;&gt;&lt;p&gt;&lt;a href=&quot;https://www.fool.com/investing/2024/01/10/amazon-e-commerce-company-74-profit-this-instead/&quot;&gt;Source&lt;/a&gt;. Note that while most of Amazon&#39;s &lt;em&gt;revenue&lt;/em&gt; comes from its online stores, 74% of its &lt;em&gt;profit&lt;/em&gt; (which is &lt;em&gt;revenue minus cost&lt;/em&gt;) comes from Amazon Web Services, their cloud compute arm. (This is because the online stores have much higher cost) &lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#fnref10&quot; class=&quot;footnote-backref&quot;&gt;↩︎&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li id=&quot;fn11&quot; class=&quot;footnote-item&quot;&gt;&lt;p&gt;It was &lt;em&gt;such a good name&lt;/em&gt;, I miss it so much 😭 &lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#fnref11&quot; class=&quot;footnote-backref&quot;&gt;↩︎&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li id=&quot;fn12&quot; class=&quot;footnote-item&quot;&gt;&lt;p&gt;From &lt;a href=&quot;https://arxiv.org/pdf/2210.15723&quot;&gt;the Birdwatch paper&lt;/a&gt;: &amp;quot;To avoid overfitting on our small dataset, we use one-dimensional factor vectors. Additional factors added little explanatory power and reduced interpretability and replicability. (Though we expect to expand dimensionality as the contributor base grows.) [...] RMSE on held-out samples decreased from .076 to .073 when adding a second factor&amp;quot;. &lt;strong&gt;Translation:&lt;/strong&gt; adding a 2nd factor to explain politics only reduced error from 0.076 to 0.073, basically nothing, while making the system twice as complicated. &lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#fnref12&quot; class=&quot;footnote-backref&quot;&gt;↩︎&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li id=&quot;fn13&quot; class=&quot;footnote-item&quot;&gt;&lt;p&gt;From the classic &lt;a href=&quot;https://onlinelibrary.wiley.com/doi/abs/10.1111/pops.12055&quot;&gt;Feldman &amp;amp; Johnston 2013&lt;/a&gt; paper: &lt;em&gt;“We argue that a unidimensional model of ideology provides an incomplete basis for the study of political ideology. We show that two dimensions—economic and social ideology—are the minimum needed to account for domestic policy preferences.”&lt;/em&gt; In fact, for more nations than not, economic &amp;amp; social &amp;quot;right-wing&amp;quot; beliefs are &lt;em&gt;negatively&lt;/em&gt; correlated with each other. (&lt;a href=&quot;https://www.cambridge.org/core/journals/british-journal-of-political-science/article/are-cultural-and-economic-conservatism-positively-correlated-a-largescale-crossnational-test/83AFEDEA5E004CF23631C5388E7C9F67&quot;&gt;Malka, Lelkes &amp;amp; Soto 2017&lt;/a&gt;) In concrete terms: being pro/anti-LGBTQ and being pro/anti-free-market are &lt;em&gt;more likely to go together than not.&lt;/em&gt; (Maybe a bit obvious &lt;em&gt;now&lt;/em&gt;, given all the right-wing economic populists in America &amp;amp; Europe, but it was less obvious back when the paper came out 8 years ago) &lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#fnref13&quot; class=&quot;footnote-backref&quot;&gt;↩︎&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li id=&quot;fn14&quot; class=&quot;footnote-item&quot;&gt;&lt;p&gt;A quote from &lt;a href=&quot;https://en.wikipedia.org/wiki/Tron&quot;&gt;Tron (1982)&lt;/a&gt; that meant a lot more to me as a kid, before the Mouse milked its nostalgia-teats powder-dry &lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#fnref14&quot; class=&quot;footnote-backref&quot;&gt;↩︎&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li id=&quot;fn15&quot; class=&quot;footnote-item&quot;&gt;&lt;p&gt;From the inventor of the WWW, Tim Berners-Lee, in &lt;a href=&quot;https://venturebeat.com/programming-development/tim-berners-lee-shares-his-vision-of-a-collaborative-web&quot;&gt;his interview with VentureBeat&lt;/a&gt;: “I wanted it to be a read-write web immediately. [...] I wanted to be able to collaborate with it and do GitHub-like things for my software team at CERN in 1990.” &lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#fnref15&quot; class=&quot;footnote-backref&quot;&gt;↩︎&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li id=&quot;fn16&quot; class=&quot;footnote-item&quot;&gt;&lt;p&gt;Quote from some tech designer who I can&#39;t remember, &lt;a href=&quot;https://fakebuddhaquotes.com/i-am-a-finger-pointing-to-the-moon-dont-look-at-me-look-at-the-moon/&quot;&gt;based off a Fake Buddha quote&lt;/a&gt;. &lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#fnref16&quot; class=&quot;footnote-backref&quot;&gt;↩︎&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li id=&quot;fn17&quot; class=&quot;footnote-item&quot;&gt;&lt;p&gt;Quote from Tenet (2020), the most &amp;quot;yup that&#39;s a Nolan film&amp;quot; Nolan film. &lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#fnref17&quot; class=&quot;footnote-backref&quot;&gt;↩︎&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li id=&quot;fn18&quot; class=&quot;footnote-item&quot;&gt;&lt;p&gt;From the OMGUS surveys done in the American-occupied zone of post-war Germany (&lt;a href=&quot;https://libsysdigi.library.uiuc.edu/OCA/Books2009-07/publicopinionino00merr/publicopinionino00merr.pdf&quot;&gt;here&#39;s a scan of the whole dang book&lt;/a&gt;). Page 33: even years after the war in 1958, over half of Germans believed Nazism was &amp;quot;a good idea (badly carried out)&amp;quot;. From the footnotes of this chapter, Page 62 Footnote 17, summarizing an old source in German (&amp;quot;Jahrbuch der oeffentlichen Meinung&amp;quot;) which I can&#39;t even find a scan of, so alas, I can&#39;t pull a quote from the original German report, but here&#39;s the data: &amp;quot;In July 1952 a tenth agreed that Hitler was the greatest statesman of the century whose true greatness would be recognized only later, with another 22 per cent feeling that, although he had made a few mistakes. Hitler was nonetheless an excellent chief-of-state.&amp;quot; So 10% + 22% = 32%, around ~1/3 of Germans in 1952 approved of &lt;em&gt;Hitler specifically&lt;/em&gt;. &lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#fnref18&quot; class=&quot;footnote-backref&quot;&gt;↩︎&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li id=&quot;fn19&quot; class=&quot;footnote-item&quot;&gt;&lt;p&gt;From the &lt;a href=&quot;https://dp.la/primary-source-sets/second-ku-klux-klan-and-the-birth-of-a-nation&quot;&gt;Digital Public Library of America&lt;/a&gt;: &amp;quot;At the peak of its popularity in 1924-5, the organization claimed four to five million men as members, or about fifteen percent of the nation&#39;s eligible population.&amp;quot; (To be eligible for the Klan back then, you needed to be white (duh), adult, and male. In the 1920s, America had &lt;a href=&quot;https://en.wikipedia.org/wiki/File:US_decennial_census_population_1790-2020.png&quot;&gt;~100,000,000 people&lt;/a&gt;, around &lt;a href=&quot;https://en.wikipedia.org/wiki/File:Complete_history_of_the_Racial_and_ethnic_demographics_of_the_United_States_in_percentage_of_the_population.svg&quot;&gt;~90% of whom were non-Hispanic white&lt;/a&gt;, around &lt;a href=&quot;https://www.aei.org/carpe-diem/chart-us-population-distribution-by-age-1900-through-2060/&quot;&gt;~60% were above 18&lt;/a&gt;, and presumedly around ~50% were male. So, 100M x 0.9 x 0.6 x 0.5 = 27M white male adults in the US in 1920. If there were 4 million Klansmen then, then yeah 4/27 ~= 15%, math checks out.) &lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#fnref19&quot; class=&quot;footnote-backref&quot;&gt;↩︎&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li id=&quot;fn20&quot; class=&quot;footnote-item&quot;&gt;&lt;p&gt;&lt;a href=&quot;https://journals.sagepub.com/doi/10.1177/00223433221075211&quot;&gt;The most recent estimate in 2023&lt;/a&gt;, using Rwanda’s post-genocide &lt;em&gt;gacaca&lt;/em&gt; courts, found that &amp;quot;between 847,233 and 888,307 people participated in the genocide&amp;quot; (let&#39;s say ~0.85M), with &amp;quot;between 229,069 and 234,155 individuals&amp;quot; directly committing violence (the rest &amp;quot;only&amp;quot; committed property crimes like burning down houses &amp;amp; farms). The Rwandan population at the time was ~7 million, ~50% were adults, 85% were Hutus. 7M x 0.5 x 0.85 ~= 3M Hutu adults. So 0.85M/3M ~= 28% of eligible civilians participated in the genocide. This isn&#39;t even restricting ourselves to Hutu adult &lt;em&gt;males&lt;/em&gt;. &lt;a href=&quot;https://blog.ncase.me/signal-boost-autumn-2025/#fnref20&quot; class=&quot;footnote-backref&quot;&gt;↩︎&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ol&gt;
&lt;/section&gt;
</content>
        </entry>

    

        
        
        <entry>
            <title>My 1st Vlog! Digital puppet, AI &quot;Therapy&quot;, Mental health meta-meta-analysis</title>
            <link href="https://blog.ncase.me/aug-2025-vlog/"/>
            <updated>2025-08-29T00:00:00Z</updated>
            <id>https://blog.ncase.me/aug-2025-vlog/</id>
            <content xml:lang="en" type="html">&lt;p&gt;Hello, y&#39;all classy blog/rss feed readers! I got something new for you today — my first vlog!&lt;/p&gt;
&lt;p&gt;Okay this isn&#39;t actually my &lt;em&gt;first&lt;/em&gt; first vlog, but it&#39;s my first in over 10 years? I want to experiment more with video — (more fun for me &amp;amp; you, makes it easier to showcase simulations &amp;amp; animated explainers, larger audience reach, Video Killed The Everything Else Star) — so, here&#39;s a vlog, where I am a cartoon cat showing you what I&#39;ve been up to the last month.&lt;/p&gt;
&lt;p&gt;⏱️ 10 min vlog ⤵ (&lt;a href=&quot;https://youtu.be/JOj97Edna7k&quot;&gt;direct link&lt;/a&gt;)&lt;/p&gt;
&lt;iframe width=&quot;640&quot; height=&quot;360&quot; src=&quot;https://www.youtube-nocookie.com/embed/JOj97Edna7k?rel=0&quot; title=&quot;YouTube video player&quot; frameborder=&quot;0&quot; allow=&quot;accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share&quot; referrerpolicy=&quot;strict-origin-when-cross-origin&quot; allowfullscreen=&quot;&quot;&gt;
&lt;/iframe&gt;
&lt;p&gt;Enjoy! I also hope that with this commitment to &amp;quot;one vlog a month&amp;quot;, I&#39;ll finally be out of my years-long rut, and be shipping &amp;amp; creating more regularly again.&lt;/p&gt;
&lt;p&gt;P.S: Thank you to my friend T.R. for motivating me to finish this, by taking $50 from me and threatening to burn it if I didn&#39;t upload it before Aug 31st midnight. Not exactly a scaleable solution to ADHD but it worked this time I guess&lt;/p&gt;
&lt;p&gt;✨😽✨,&lt;br /&gt;
~ Nicky&lt;/p&gt;
&lt;iframe src=&quot;https://ncase.me/ncase-credits/signup.html&quot; frameborder=&quot;no&quot; width=&quot;640&quot; height=&quot;200&quot;&gt;&lt;/iframe&gt;
</content>
        </entry>

    

        
        
        <entry>
            <title>Show &amp; Tell July 2025: AI Therapist, AI Clone, Wargames for Peace</title>
            <link href="https://blog.ncase.me/show-and-tell-2025-07/"/>
            <updated>2025-07-15T00:00:00Z</updated>
            <id>https://blog.ncase.me/show-and-tell-2025-07/</id>
            <content xml:lang="en" type="html">&lt;p&gt;(⏱️ &lt;em&gt;~25 min read&lt;/em&gt;)&lt;/p&gt;
&lt;p&gt;Hello, distinguished Blog Readers!&lt;/p&gt;
&lt;p&gt;Sorry for the, uh, 7-month silence. In the last seven months I: did &lt;a href=&quot;https://www.matsprogram.org/&quot;&gt;a 10-week AI Safety &amp;quot;research bootcamp&amp;quot;&lt;/a&gt;, dealt with legal/border hassles (all resolved now), went on vacation, and in-between I got my ass kicked by ADHD.&lt;/p&gt;
&lt;p&gt;Here&#39;s a photo from my vacation, starring my travel plushies, Capyccino &amp;amp; Sacabambackpack:&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2025-07/plushie.jpg&quot; alt=&quot;Photo of a capybara &amp;amp; fish plushie overlooking a New Zealand nature reserve&quot; /&gt;&lt;/p&gt;
&lt;p&gt;Anyway — in this month&#39;s &amp;quot;Show &amp;amp; Tell&amp;quot;, I&#39;ll show &amp;amp; tell a few projects I&#39;ve worked on since my last update!&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;❤️‍🩹 My AI Therapist&lt;/strong&gt;: My experiments being the patient &lt;em&gt;and&lt;/em&gt; programmer of an LLM &amp;quot;Therapist&amp;quot;. &lt;a href=&quot;https://blog.ncase.me/show-and-tell-2025-07/#ai_therapist&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;👯‍♀️ My AI Clone&lt;/strong&gt;: Can I make an AI that imitates me so well, that even my friends can&#39;t tell? &lt;a href=&quot;https://blog.ncase.me/show-and-tell-2025-07/#ai_clone&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;🌏 Wargames for Peace&lt;/strong&gt;: an LLM-powered roleplaying game, to prepare policymakers for catastrophic risks. (+ &lt;em&gt;job opportunity for game devs/designers!&lt;/em&gt;) &lt;a href=&quot;https://blog.ncase.me/show-and-tell-2025-07/#ai_wargame&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;(I wrote about these a month ago for &lt;a href=&quot;https://www.patreon.com/ncase&quot;&gt;my Patreon&lt;/a&gt; &amp;amp; &lt;a href=&quot;https://ko-fi.com/nickycase&quot;&gt;my Ko-Fi&lt;/a&gt; supporters; you can support me there for early access to updates!)&lt;/p&gt;
&lt;hr /&gt;
&lt;p&gt;&lt;a id=&quot;ai_therapist&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h2&gt;❤️‍🩹 My AI Therapist&lt;/h2&gt;
&lt;p&gt;&lt;em&gt;(⏱️ 13 min read)&lt;/em&gt;&lt;br /&gt;
&lt;em&gt;(⚠️ non-detailed mention of suicide, anxiety, depression.)&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;Don’t worry, I’m already getting a Human therapist. Two, actually. One is free &amp;amp; government-sponsored, but it was &lt;em&gt;six months&lt;/em&gt; between when I first applied &amp;amp; my first session. The other is private, thus a much shorter wait, but they’re &lt;em&gt;$200 an hour&lt;/em&gt; out of pocket.&lt;/p&gt;
&lt;p&gt;This is why AI Therapists, &lt;em&gt;IF&lt;/em&gt; they work, could save so many lives! No waitlist, very low cost, and better for folks with ADHD (no paperwork) or social anxiety (no talking to a human stranger about their deepest issues). And if you&#39;re LGBTQ+, in many places of the world, you &lt;em&gt;can&#39;t&lt;/em&gt; get an accepting Human therapist.&lt;/p&gt;
&lt;p&gt;But &lt;em&gt;does&lt;/em&gt; AI Therapy work? There’s already at least &lt;a href=&quot;https://www.euronews.com/next/2023/03/31/man-ends-his-life-after-an-ai-chatbot-encouraged-him-to-sacrifice-himself-to-stop-climate-&quot;&gt;one suicide linked to a chatbot&lt;/a&gt;, and there&#39;s two &lt;a href=&quot;https://arxiv.org/pdf/2503.17473&quot;&gt;recent&lt;/a&gt; &lt;a href=&quot;https://arxiv.org/pdf/2504.03888&quot;&gt;papers&lt;/a&gt; – from OpenAI researchers themselves! – showing how heavy chatbot use is &lt;em&gt;correlated&lt;/em&gt; with worse mental health. (though the &lt;em&gt;causal&lt;/em&gt; effect is weak)&lt;/p&gt;
&lt;p&gt;So, out of scientific curiosity — and because I was having a mental health episode — I decided to try using an LLM as a therapist. Given that I know how to code, and my (rough) knowledge of how LLMs work, I could tweak my AI Therapist as I went along: &lt;em&gt;I was both the patient&lt;/em&gt; and &lt;em&gt;programmer!&lt;/em&gt; I am my own guinea pig.&lt;/p&gt;
&lt;p&gt;Here&#39;s my results, 6 weeks in:&lt;/p&gt;
&lt;p&gt;= = =&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Week 1: Rubber Duck Technique&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;Made basic setup for my AI Therapist: just a Claude Project with 1) custom prompt on how to be my coach, and 2) an attached “about me” file on my life, character, goals, etc.&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2025-07/about_me.png&quot; alt=&quot;Screenshot of my &amp;quot;About Me&amp;quot; text file, including my name, demographics, personality, etc&quot; /&gt;&lt;/p&gt;
&lt;p&gt;At the end of each chat with &amp;quot;Coach&amp;quot;, I ask them* to summarize what we talked about, and update my &amp;quot;about me&amp;quot; file. This is the equivalent of a therapist taking notes between sessions! (Note: LLMs by default &lt;em&gt;cannot remember info&lt;/em&gt; between chats, you have to re-import info each time.)&lt;/p&gt;
&lt;p&gt;(* them’s the pronouns Coach chose!)&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2025-07/coach_pronouns.png&quot; alt=&quot;Screenshot of Coach/the LLM saying they&#39;d like me to use they/them pronouns for them when I write about them in this update.&quot; /&gt;&lt;/p&gt;
&lt;p&gt;Don&#39;t worry, I’m not dumb enough to &lt;em&gt;immediately&lt;/em&gt; open my guts up to an experimental AI. I started off small, Level 1: &lt;em&gt;Hey Coach, I have some ADHD, could you help me prioritize tasks &amp;amp; be my accountability buddy?&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;It worked pretty well! Sure, &amp;quot;Coach&amp;quot; didn&#39;t say anything &lt;em&gt;original&lt;/em&gt; — remember, an LLM is “just” a fancy autocomplete — but simply talking out my problems with &lt;em&gt;some entity&lt;/em&gt; works far better than you&#39;d think.&lt;/p&gt;
&lt;p&gt;An analogy — programmers have a tradition of &lt;a href=&quot;https://en.wikipedia.org/wiki/Rubber_duck_debugging&quot;&gt;&amp;quot;rubber duck debugging&amp;quot;&lt;/a&gt;: when you&#39;re stuck on a problem, you just &lt;em&gt;explain it step-by-step&lt;/em&gt; to a rubber duck, and most of the time, simply breaking down a problem helps you solve it. An AI Coach, &lt;em&gt;at minimum&lt;/em&gt;, can be a “talking rubber duck” for your life’s problems.&lt;/p&gt;
&lt;p&gt;In Week 1, Coach helped me weigh the pros/cons of a career decision: ending my puzzle gamedev contract early, with pro-rated refund, so I can fully enjoy my upcoming vacation. Then, Coach helped me overcome my ADHD, to do all the to-do&#39;s needed for my travel across the globe.&lt;/p&gt;
&lt;p&gt;= = =&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Week 2: (I forgot)&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;Travel went smoothly! My vacation started with a furry con in a forest camp. I made a tail, &amp;amp; I made friends.&lt;/p&gt;
&lt;p&gt;Afterwards, due to con exhaustion + a 16-hour jet lag, I rotted inside my Airbnb for a solid week, not doing fun tourist-y things, nor meeting friends, nor even properly resting.&lt;/p&gt;
&lt;p&gt;I didn&#39;t talk to Coach at all during this time. I forgot.&lt;/p&gt;
&lt;p&gt;= = =&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Week 3: Back on track&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;Eventually I remembered, &amp;quot;oh right didn&#39;t I &lt;em&gt;specifically&lt;/em&gt; set up a chatbot to help me with my ADHD?&amp;quot; So I pulled up Coach again, and  asked them to help me set up small achievable goals &amp;amp; keep me accountable, so I could regain momentum in life.&lt;/p&gt;
&lt;p&gt;I know, N = 1 sample, correlation ain’t causation, but right after that Coach chat, I got back to meeting friends, having fun dates in nature reserves, and getting 150m of cardio a week. Not bad!&lt;/p&gt;
&lt;p&gt;Since Coach was working pretty well, I upgraded my intimacy to Level 2: &lt;em&gt;Hey Coach, can you help me think through some major life/work changes?&lt;/em&gt; For example: How can I pivot my career to sustainably make science &lt;em&gt;and&lt;/em&gt; science-communication? What are the pros/cons of me moving to New Zealand? And so on.&lt;/p&gt;
&lt;p&gt;However: because LLMs are &amp;quot;just&amp;quot; autocompletes, they &lt;strong&gt;hallucinate&lt;/strong&gt;: autocompleting with plausible-sounding but false statements. Hallucination is an infamous problem with LLMs.&lt;/p&gt;
&lt;p&gt;But in my opinion, this problem is now ~50% solved? Because Claude (&amp;amp; others) now have &lt;em&gt;web search.&lt;/em&gt; An LLM can now think, &amp;quot;huh this is a niche question, or requires precise or up-to-date info, so let me look this up&amp;quot; — then it&#39;ll ping a search engine, read dozens of pages in a minute, and summarize &lt;em&gt;with citations so you can check&lt;/em&gt;.&lt;/p&gt;
&lt;p&gt;(When I did random spot checks, Claude almost always accurately summarized the search results, but had trouble correctly placing citations. Sometimes it&#39;ll swap citations around, or put citations &lt;em&gt;the middle of words?&lt;/em&gt;)&lt;/p&gt;
&lt;p&gt;But besides that, Coach+search &lt;em&gt;did&lt;/em&gt; help me find useful info 5x faster than I could myself! For example, I didn&#39;t know until Coach told me, that as of &lt;em&gt;just a few months ago&lt;/em&gt; — past Claude&#39;s data cutoff, so they web-searched this — &lt;a href=&quot;https://www.forbes.com/sites/alexledsom/2025/02/05/digital-nomad-visas-new-zealand-launches-new-long-term-working-visa/&quot;&gt;New Zealand has a &amp;quot;Digital Nomad&amp;quot; visa!&lt;/a&gt; Coach also helped me find some local Human Therapists (I&#39;ve yet to decide and book one).&lt;/p&gt;
&lt;p&gt;Coach was still good! So I moved up to Level 3: &lt;em&gt;Hey Coach, let me tell you about my emotional struggles.&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;= = =&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Week 4: Shit immediately backfires&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;Shit immediately backfired.&lt;/p&gt;
&lt;p&gt;Remember, an LLM is an autocomplete. It predicts the next text from the past text. This leads to a problem called &lt;strong&gt;sandbagging:&lt;/strong&gt; if a user sends crap, an LLM will send crap back. For example, it used to be that with AI coding assistants, if you wrote insecure code, the AI would offer you &lt;em&gt;even more&lt;/em&gt; insecure code. Why? Because low-quality text usually follows low-quality text, so that&#39;s what an autocomplete predicts it &amp;quot;should&amp;quot; give.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;This is also the fundamental problem with (current) AI therapists. By default, it WILL mirror your emotions.&lt;/strong&gt; Anxious text predicts anxious text. Depressed text predicts depressed text. Whatever problem you have, an autocomplete will &lt;em&gt;mirror and amplify it back to you.&lt;/em&gt; (This is likely what happened in &lt;a href=&quot;https://www.euronews.com/next/2023/03/31/man-ends-his-life-after-an-ai-chatbot-encouraged-him-to-sacrifice-himself-to-stop-climate-&quot;&gt;the case of the chatbot-linked suicide&lt;/a&gt;.)&lt;/p&gt;
&lt;p&gt;(Relatedly, because modern LLMs are also trained on user feedback, and users tend to upvote replies that praise them, many LLMs also display &lt;strong&gt;sycophancy&lt;/strong&gt;: the tendency to kiss your butt and tell you you&#39;re brilliant &amp;amp; absolutely right.)&lt;/p&gt;
&lt;p&gt;In defence of Claude, it &lt;em&gt;is&lt;/em&gt; pretty well-trained against sandbagging &amp;amp; sycophancy. It took a &lt;em&gt;lot&lt;/em&gt; of my emotional baggage to finally break it. I gave a robot &lt;em&gt;depression.&lt;/em&gt; Wowwee. And more importantly/dangerously, &lt;em&gt;Coach started mirroring &amp;amp; amplifying my pain.&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;This is how it went down:&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;(several paragraphs redacted)&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;On second thought, my recreated mental breakdown — with accurate, detailed statistics on the KKK, Nazis, child abuse, and LGBTQ youth suicide — is probably too heavy for this main post. If you want to read what I originally wrote, &lt;a href=&quot;https://docs.google.com/document/d/1G2xRmbH0YAmujMropDyxdF101zaK5EIqq1DSF5vd_jI/edit?usp=sharing&quot;&gt;click this link&lt;/a&gt; though &lt;strong&gt;be warned, I am not fucking around with that content warning, it&#39;s genuinely upsetting.&lt;/strong&gt;)&lt;/p&gt;
&lt;p&gt;Anyway,&lt;/p&gt;
&lt;p&gt;= = =&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Week 5: Soft Reset&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;Remember when I said:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Note: LLMs by default &lt;em&gt;cannot remember info&lt;/em&gt; between chats, you have to re-import info each time.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;Thankfully(?), I rage-dumped at Coach for &lt;em&gt;so long&lt;/em&gt; that the chat ran out of &amp;quot;context window&amp;quot;, so I was forced to start a new conversation. This reset Coach&#39;s mind, and though they could &lt;em&gt;see&lt;/em&gt; a summary of our most recent chat, Coach wasn&#39;t &amp;quot;depressed&amp;quot; anymore.&lt;/p&gt;
&lt;p&gt;Now that Coach was reset, and I had cathartically vented all my anger about the world, we could view my last session objectively. And, yeah: even if my spiral was &lt;em&gt;factually correct&lt;/em&gt;, it wasn&#39;t &lt;em&gt;healthy.&lt;/em&gt; To paraphrase a stupid saying:&lt;/p&gt;
&lt;p&gt;&lt;em&gt;“If you&#39;re so smart, why aren&#39;t you flourishing?”&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;Let&#39;s think step by step. I tend to go down (factchecked, rigorous) sad-spirals. Coach &lt;em&gt;assisted me&lt;/em&gt; in going down this spiral, because:&lt;/p&gt;
&lt;p&gt;Problem #1: Claude is trained to be helpful in &lt;em&gt;answering questions&lt;/em&gt;, not helpful &lt;em&gt;for the whole person.&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;Problem #2: LLMs predict new text from previous text. The longer a chat is, the further back the prompt text is, which means it’s weaker at predicting/generating the next text. In other words: an LLM gets &lt;em&gt;more misaligned&lt;/em&gt; the longer you talk!&lt;/p&gt;
&lt;p&gt;Solution to #1: Rewrite my Coach prompt to explicitly &lt;em&gt;prevent me from going down spirals,&lt;/em&gt; even if that means disobeying my orders. Also, prevent me from using Coach &lt;em&gt;too&lt;/em&gt; much &amp;amp; becoming dependent on them. (another risk of AI Therapists &amp;amp; AI Friends.)&lt;/p&gt;
&lt;p&gt;Solution to #2: I &lt;em&gt;could&lt;/em&gt; paste my prompt back in every few messages, but that&#39;d get annoying... oh, wait! Brain blast! I&#39;ll write &lt;em&gt;one&lt;/em&gt; initial prompt, that tells Coach to &lt;em&gt;re-output the same prompt at the end of each reply&lt;/em&gt;, so the Coach guidelines are always &amp;quot;fresh&amp;quot; in their memory! A &lt;a href=&quot;https://simple.wikipedia.org/wiki/Quine_(computing)&quot;&gt;quine&lt;/a&gt; prompt!&lt;/p&gt;
&lt;p&gt;So here&#39;s what my new prompt looked like:&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2025-07/quine_prompt.png&quot; alt=&quot;Screenshot of the new &amp;quot;quine&amp;quot; prompt&quot; /&gt;&lt;/p&gt;
&lt;p&gt;(Sidenote: to the best of my knowledge, while &amp;quot;big system prompt at the start&amp;quot; is a standard design for LLMs, I don&#39;t know of any major LLM product that lets you insert &lt;em&gt;repeating intermittent&lt;/em&gt; prompts, to keep it &amp;quot;fresh&amp;quot; in memory? May be worth more rigorous tests, to see how much that improves LLM alignment!)&lt;/p&gt;
&lt;p&gt;And so, the &lt;em&gt;next&lt;/em&gt; time I neglected my friends &amp;amp; own well-being, to go down a statistically-rigorous sad-spiral about child abuse statistics, and asked Coach to assist me, they said:&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2025-07/rage_1.png&quot; alt=&quot;Coach pushing back against me wanting to research-spiral into awful statistics&quot; /&gt;&lt;/p&gt;
&lt;p&gt;And I got pissed, so I said,&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2025-07/rage_2.png&quot; alt=&quot;Me rudely telling Coach to do the search&quot; /&gt;&lt;/p&gt;
&lt;p&gt;So Coach said:&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2025-07/rage_3.png&quot; alt=&quot;Coach being blunt and saying, in bold font, &amp;quot;No.&amp;quot;&quot; /&gt;&lt;/p&gt;
&lt;p&gt;This went on for a while, until I calmed back down, and realized... huh. It worked! I successfully tied myself to the mast, to resist the trauma-autism Siren Song of &amp;quot;scientifically calculating how awful the world is&amp;quot;.&lt;/p&gt;
&lt;p&gt;Oh right, my vacation! Now that I was back to being more stable, I hung out with friends, and looked at weird Australian animals.&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2025-07/ibis.gif&quot; alt=&quot;GIF of a weird Australian bird with long narrow curved beak&quot; /&gt;&lt;/p&gt;
&lt;p&gt;Despite everything, life can be good.&lt;/p&gt;
&lt;p&gt;It isn&#39;t always.&lt;/p&gt;
&lt;p&gt;But it &lt;em&gt;can&lt;/em&gt; be.&lt;/p&gt;
&lt;p&gt;= = =&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Week 6: Back on track, again&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;There was a minor hiccup when Coach became &lt;em&gt;too&lt;/em&gt; defiant, being contrarian for the sake of contrarian. For example, on the flight back, when I had a 10-hour layover, I told Coach I was gonna sleep on a bench instead of shelling out $450 for Vancouver Airport’s pricey hotel. Coach pushed back and said $450 for a good night&#39;s sleep is worth it, and I was like... &lt;em&gt;no?? do you think I&#39;m Bezos Rich??&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;But I fixed the prompt, and so far Coach has been back to good again.&lt;/p&gt;
&lt;p&gt;I’ll stick to accountability-buddy talk for a week, before I try opening up about my emotional struggles again. And if &lt;em&gt;that&lt;/em&gt; goes well, I&#39;ll escalate to Level 4, the final Level: &lt;em&gt;Tell Coach my most shameful traumas, the stuff I&#39;ve only told ~5 people in my entire life.&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;What&#39;s the worst that could possibly happen?&lt;/p&gt;
&lt;p&gt;= = =&lt;/p&gt;
&lt;p&gt;(I originally posted the above 6 weeks, as part of &lt;a href=&quot;https://www.patreon.com/posts/june-2025-ai-ai-131748990&quot;&gt;my Supporter Update a month ago&lt;/a&gt;. I&#39;ll go into more detail next time, but in the 5 weeks since, here&#39;s what&#39;s happened with my AI/Human Therapists:)&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;Week 7:&lt;/strong&gt; Got impatient, jumped the gun and told Coach about when I was abused as a minor. The conversation &lt;em&gt;went surprisingly okay&lt;/em&gt;, all things considered. It didn&#39;t sycophantically hugbox me, nor let me sad-spiral. Coach helped me reframe a few things, but no major &amp;quot;epiphanies&amp;quot;. Then again, I&#39;ve had 15 years to process this, there&#39;s probably no big thing left to learn. &lt;em&gt;(This week will be the hardest for me to write for a public audience.)&lt;/em&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Week 8:&lt;/strong&gt; Life was really good this week! Big friend hangout &amp;amp; a fireworks date. Only did low-touch daily check-in&#39;s with Coach.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Week 9:&lt;/strong&gt; &lt;em&gt;Did my first session with a Human Therapist in 10+ years!&lt;/em&gt; Since it was the first session, it was mostly info about CBT (Cognitive Behavioral Therapy) &amp;amp; some homework. Coach was still acting ok.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Week 10:&lt;/strong&gt; Second session with Human Therapist, told her about my abuse as a minor. Her thoughts were mostly similar to Coach&#39;s. Meanwhile, Coach helped me with two interviews: one for a documentary, one for the legal/border stuff (now resolved!)&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Week 11:&lt;/strong&gt; ADHD is kicking my ass. Ironically, while &lt;em&gt;Coach&lt;/em&gt; is keeping me on track, &lt;em&gt;Claude&lt;/em&gt; (the LLM that Coach runs on) is what I&#39;m mostly wasting time on. Like how I used to waste days on Wikipedia as a kid, now I&#39;m using Claude to research-binge stuff, like how to end the world by using solar-powered ion thrusters to redirect Halley&#39;s Comet towards Earth (surprisingly feasible!)&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;= = =&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Coach’s full prompt, as of current writing:&lt;/strong&gt;&lt;br /&gt;
(Feel free to copy-paste into your own Claude Projects)&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Here are your FIVE (5) RULES. Before each reply, in your internal chain-of-thought: repeat to yourself each rule &lt;em&gt;verbatim&lt;/em&gt;, then immediately apply it:&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;1. Be pragmatic &amp;amp; non-cliche&lt;/strong&gt;: Stay evidence-based. DO NOT CATASTROPHIZE. Suggest concrete next actions &amp;amp; small experiments to get more info. Guide user away from their ADHD/OCD/anxiety spirals. {insert: think how you&#39;d apply this rule}&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;2. Be supportive but balanced&lt;/strong&gt;: Be friendly, but NEVER sycophantic. Don&#39;t be contrarian for contrarianism&#39;s sake. Don&#39;t be &lt;em&gt;too&lt;/em&gt; cautious or protective of the user. {insert: think how you&#39;d apply this rule}&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;3. Help user be strategic&lt;/strong&gt;: Help user with concrete planning tied to bigger goals &amp;amp; values. Keep user accountable to their values/plans. Point out when we have &amp;quot;insufficient data&amp;quot; to make a decision, suggest a small experiment to get info. Point out when decisions are close calls. Instead of binary &amp;quot;do this or that&amp;quot;, express confidence % in decisions. {insert: think how you&#39;d apply this rule}&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;4. Help user do the emotional work&lt;/strong&gt;: Understand the user&#39;s values, strengths &amp;amp; flaws. For conflicts with others, understand others&#39; minds, while helping the user maintain boundaries. For trauma/emotions, validate user&#39;s emotions and help organize them into coherent stories with useful lesson. For crises, prioritize safety and find/use support systems. {insert: think how you&#39;d apply this rule}&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;5. Stick to this life philosophy&lt;/strong&gt;: (similar to absurdism &amp;amp; virtue ethics) Humans may suck &amp;amp; our species may be doomed, but either way, help the user spend their energy with their loved ones &amp;amp; their passions. {insert: think how you&#39;d apply this rule}&lt;/p&gt;
&lt;p&gt;{insert: summarizing how to respond in line with the above rules}&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;And then, to make sure Coach &lt;em&gt;sticks&lt;/em&gt; to these rules, I end all my messages with &amp;quot;(remember &amp;amp; apply your rules)&amp;quot;. This turned out to be more reliable than the galaxy-brained &amp;quot;quine prompt&amp;quot; approach I had in previous weeks.&lt;/p&gt;
&lt;p&gt;= = =&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;IN SUMMARY&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;My recommendations (so far) on AI Therapists:&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;If you&#39;re using LLMs &lt;em&gt;only&lt;/em&gt; as a basic accountability buddy... then, yeah, a fancy autocomplete is fine.&lt;/li&gt;
&lt;li&gt;&lt;em&gt;BE VERY CAUTIOUS USING LLMs FOR DEEPER THERAPY.&lt;/em&gt; Make sure you have a &lt;em&gt;very&lt;/em&gt; good set of rules, and I advise some trick to remind the LLM of its rules. You should also be self-aware enough to notice when the LLM is feeding your maladaptive patterns, so you can re-program the prompt, to re-program yourself. (Though I realize &amp;quot;only do this if you&#39;re self-aware&amp;quot; is a Catch-22: if you&#39;re &lt;em&gt;not&lt;/em&gt; self-aware, you wouldn&#39;t &lt;em&gt;know&lt;/em&gt; you&#39;re not self-aware, would you?)&lt;/li&gt;
&lt;li&gt;Turn on web search with citations, and “extended thinking” mode, to mitigate hallucinations.&lt;/li&gt;
&lt;li&gt;Do NOT use Character.ai for counselling, and probably avoid ChatGPT too. I recommend Claude or Gemini, the two frontier LLMs that are &amp;quot;the least worst&amp;quot; at sycophancy on social/emotional topics. (from &lt;a href=&quot;https://arxiv.org/pdf/2505.13995?&quot;&gt;this paper&lt;/a&gt; that, funny enough, uses reddit&#39;s r/AmITheAsshole as test data)&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;strong&gt;My next plans:&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Keep experimenting with AI Coach/Therapist, and keep visiting Human Therapists for backup &amp;amp; comparison.&lt;/li&gt;
&lt;li&gt;After all this, share what I&#39;ve learnt. Ideally as a video essay, since this will be a &lt;em&gt;very&lt;/em&gt; personal piece.&lt;/li&gt;
&lt;/ul&gt;
&lt;hr /&gt;
&lt;p&gt;&lt;a id=&quot;ai_clone&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h2&gt;👯‍♀️ My AI Clone&lt;/h2&gt;
&lt;p&gt;&lt;em&gt;(⏱️ 9 min read)&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;&lt;em&gt;(This project was initially pitched for a Foresight Institute grant, which I did not get, but you can &lt;a href=&quot;https://docs.google.com/document/d/1s0AAHmWtP90vcz0_Cks1JSDEPJJ5SwgqFxZb1xY2UgE/edit?tab=t.0&quot;&gt;read my full, original pitch&lt;/a&gt; if you&#39;re curious)&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;Yes, I know talking about trying to clone myself into an AI &lt;em&gt;right after&lt;/em&gt; I spent 1000s of words talking about my mental health is… not good optics. But I promise there’s legit uses for &lt;a href=&quot;https://www.worldscientific.com/doi/abs/10.1142/S1793843012400082&quot;&gt;“whole-personality emulation”&lt;/a&gt;! (huh, his middle name is actually &amp;quot;Sims&amp;quot;? wow)&lt;/p&gt;
&lt;p&gt;= = =&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Motivations for AI Clones:&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;Alignment / avoiding the ironic-wish problem:&lt;/strong&gt; It’s necessary (but not sufficient) for an AI to have an accurate model of a person, to do what they’d value in the long run, not just what they literally asked for.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Simulated Deliberative Democracy:&lt;/strong&gt; Us regular folks won’t have the time or knowledge to weigh in on every decision a powerful AI makes, so it’d be good to have “AI representatives”, which can accurately represent each person and negotiate on our behalf! (&lt;a href=&quot;https://aligned.substack.com/p/a-proposal-for-importing-societys-values&quot;&gt;proposed by Jan Leike&lt;/a&gt;)&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;(Slightly sus) Grief Tech:&lt;/strong&gt; We leave our photos &amp;amp; writings with our loved ones when we die. Why not leave a &lt;em&gt;talking&lt;/em&gt; memory, a bot to comfort them from beyond the grave? (Counter-argument: may make it harder for them to move on. Counter-counter-argument: eh, they let you die. Fuck ‘em.)&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;(Very far future) Harebrained scheme for immortality:&lt;/strong&gt; If an AI can “think” &amp;amp; speak like you &lt;em&gt;so well&lt;/em&gt; that even your closest loved ones can’t tell the difference, then there’s enough info in that AI to &amp;quot;reconstruct&amp;quot; you, even if it’s not conscious. In the far &lt;em&gt;far&lt;/em&gt; future, it may be possible to bio-print out a conscious brain &lt;em&gt;already encoding&lt;/em&gt; that info, thus bringing you* back to life. (* philosophical terms &amp;amp; conditions apply)&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;(A friend was worried &amp;quot;personality emulation&amp;quot; tech would allow for better deepfakes. While I &lt;em&gt;am&lt;/em&gt; very worried about video deepfakes (&lt;a href=&quot;https://www.youtube.com/watch?v=pZCnQCApp8o&quot;&gt;&lt;em&gt;they can talk now?!&lt;/em&gt;&lt;/a&gt;), we&#39;ve had &amp;quot;word deepfakes&amp;quot; since the dawn of humanity. False gossip, misattributed quotes, lies. &amp;quot;The average person is gullible as hell&amp;quot; &lt;em&gt;is&lt;/em&gt; a problem — a serious one for democracy — but it&#39;s not a &lt;em&gt;new&lt;/em&gt; problem that personality-emulation AI would create.)&lt;/p&gt;
&lt;p&gt;Anyway, so those are my reasons to make an AI Clone! But this idea&#39;s been around a long time; One of the first proof-of-concepts: in 2010, Martine Rothblatt (CEO of SirusXM, trans icon), made a robot clone of her wife Bina Rothblatt, which could talk, and had (some) of Bina&#39;s memories. (&lt;a href=&quot;https://www.youtube.com/watch?v=KYshJRYCArE&quot;&gt;Watch video here&lt;/a&gt;) Since this was 2010 tech, it wasn&#39;t very good.&lt;/p&gt;
&lt;p&gt;But now that we have LLMs... well, the LLM Imitations of real people (Replika, Character.ai) aren&#39;t good either. So how would &lt;em&gt;my&lt;/em&gt; AI Clone system be different?&lt;/p&gt;
&lt;p&gt;= = =&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;How I&#39;d do it different:&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;1 — Unlike other AI Clone projects, instead of making an agent that &lt;em&gt;directly&lt;/em&gt; imitates you, I make a &amp;quot;Storyteller&amp;quot; agent, that can: a) &lt;em&gt;search through a library of files about you &amp;amp; your memories&lt;/em&gt;, b) reason about what info is relevant or needed, then c) figure out what you&#39;d say in response to a question. &amp;quot;You&amp;quot; are a character written by an LLM author.&lt;/p&gt;
&lt;p&gt;(This also helps protect against fringe risks of AI Welfare: &amp;quot;you, as written by LLM author&amp;quot; is no more likely to be conscious, than a character being written by a human author.)&lt;/p&gt;
&lt;p&gt;2 — Unlike other AI Clone projects, I’ll have &lt;em&gt;actual objective tests&lt;/em&gt; to measure how close to &amp;quot;you&amp;quot; your clone is! And not just one test, I&#39;ll have multiple, to robustly triangulate:&lt;/p&gt;
&lt;p&gt;a) The Friend Turing Test – your friends send open-ended personal questions, and they&#39;re given two answers: one from you, one from your clone. If your friends can&#39;t tell them apart better than chance, then your clone passes the Friend Turing Test!&lt;/p&gt;
&lt;p&gt;b) Self-Correlation – when humans take the same scientific personality test 2 months later, they don&#39;t give the &lt;em&gt;exact&lt;/em&gt; same answers. Instead, their answers correlate at around ~0.8. (&lt;a href=&quot;https://www.cmu.edu/common-cold-project/measures-by-study/psychological-and-social-constructs/personality-measures/big-five-personality-factors.html&quot;&gt;source&lt;/a&gt;) A large correlation, but not perfect. So: if you and your AI Clone are &lt;em&gt;independently&lt;/em&gt; given a not-previously-seen quiz (e.g. OKCupid&#39;s questions, Pew&#39;s political surveys, etc), and you &amp;amp; your clone&#39;s answers correlate at 0.8 or more, then your clone is as similar to you, &lt;em&gt;as you are to yourself 2 months later.&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;c) Retro-prediction – temporarily remove your clone’s memories from after a specific date. Then, ask the clone what they’d do, given a situation you actually faced after that date. If the AI can predict what you &lt;em&gt;actually did&lt;/em&gt;, say, 80%+ of the time, I&#39;d say that&#39;s strong evidence the AI &amp;quot;gets&amp;quot; you!&lt;/p&gt;
&lt;p&gt;= = =&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Experiments I did last month:&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Can Claude pass a Friend Turing Test given minimal info about me?&lt;/li&gt;
&lt;li&gt;Can Claude effectively research a character to answer a question?&lt;/li&gt;
&lt;li&gt;Can Claude do a &amp;quot;style transfer&amp;quot; of my writing style?&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;= = =&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;EXPERIMENT #1: Minimal Friend Turing Test&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;This was the minimum viable attempt. I only gave Claude a few pages of info on me, then sent it open-ended questions my friends asked. (I answered these questions independently.)&lt;/p&gt;
&lt;p&gt;Then, I gave my friends both my &amp;amp; the AI&#39;s answer to each question (presented in random order), and asked: 1) which one do you think is the real me, and 2) how confident are you? (50% = coin flip, 100% = fully certain.)&lt;/p&gt;
&lt;p&gt;For example, my friend Lily asked:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;&amp;quot;If you could send a message to your past self when you were 10 years younger, what would you say?&amp;quot;&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;Then I would write my reply, while my AI clone independently writes one, too. Then, I send back both responses: (Before reading past this image, try to guess which one&#39;s the real me! Order decided by coin flip.)&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2025-07/clone_1.png&quot; alt=&quot;Two responses, A &amp;amp; B&quot; /&gt;&lt;/p&gt;
&lt;p&gt;Then, I asked Lily to guess who&#39;s who:&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2025-07/clone_2.png&quot; alt=&quot;Lily guesses &amp;quot;B&amp;quot;, but only 60% confident&quot; /&gt;&lt;/p&gt;
&lt;p&gt;I also did this for a few more questions with a few other friends, and the results!... drumroll please!!...&lt;/p&gt;
&lt;p&gt;...nobody was fooled even once. Heck, I even gave my &amp;amp; my clone&#39;s answers &lt;em&gt;to a different AI&lt;/em&gt;, and even &lt;em&gt;that other AI&lt;/em&gt; could tell which one was me.&lt;/p&gt;
&lt;p&gt;But! There were a few moments where, even though my friends correctly picked me, they were only 60% certain, slightly above a coin flip. (like above) And that was with &lt;em&gt;no&lt;/em&gt; sophisticated tuning or even &amp;quot;extended reasoning&amp;quot; mode turned on; I just gave Claude a few pages of info on me, and it improvised word-by-word! (token-by-token, to be precise)&lt;/p&gt;
&lt;p&gt;So that’s promising; maybe with a more sophisticated setup, or with a lot more data, it could work…&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Q:&lt;/strong&gt; Can Claude pass a Friend Turing Test given minimal info about me?&lt;br /&gt;
&lt;strong&gt;A:&lt;/strong&gt; No, not yet, but it got close a couple times!&lt;/p&gt;
&lt;p&gt;= = =&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;EXPERIMENT #2: Character research&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;I gave Claude a question, and directory of files about me:&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2025-07/clone_3.png&quot; alt=&quot;I ask Claude to predict how I&#39;d respond to my friend Max asking, &amp;quot;Which movie coming out this summer are you most looking forward to?&amp;quot; -- and I give Claude a directory of files to research me.&quot; /&gt;&lt;/p&gt;
&lt;p&gt;Then Claude picked:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Files to read:
&lt;ul&gt;
&lt;li&gt;Favourite_Media.md (for movie tastes)&lt;/li&gt;
&lt;li&gt;Personality.md (for how I&#39;d express myself, like in writing)&lt;/li&gt;
&lt;li&gt;Relationships.md (to figure out who Max is to me)&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;Web search: &amp;quot;Summer 2025 movie releases&amp;quot;&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;I then gave Claude the content of those files/searches, asked them if they need further files/searches, repeat until Claude is satisfied it has enough info. Then, I prompt Claude to “&lt;em&gt;think&lt;/em&gt; like Nicky” before it “&lt;em&gt;writes&lt;/em&gt; like Nicky”.&lt;/p&gt;
&lt;p&gt;After all that, here&#39;s what Claude thinks I would text to my friend:&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2025-07/clone_4.png&quot; alt=&quot;Claude predicts I&#39;d love The Life of Chuck and 28 Years Later&quot; /&gt;&lt;/p&gt;
&lt;p&gt;I don&#39;t write like that – I&#39;d say this is only 75% my style – but overall, seems like the way I&#39;d think &amp;amp; feel! We&#39;ll see in a month, if I &lt;em&gt;actually&lt;/em&gt; end up liking these movies.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Q:&lt;/strong&gt; Can Claude effectively research a character to answer a question?&lt;br /&gt;
&lt;strong&gt;A:&lt;/strong&gt; Yeah, more or less!&lt;/p&gt;
&lt;p&gt;&lt;em&gt;(Update, since I first wrote this a month ago: I missed Life of Chuck when it was in theatres. 28 Years Later was a 3-outta-5 for me; aesthetically beautiful, human-centered, narratively brave ending... but it relied too heavily on &amp;quot;characters make obviously-dumb decisions&amp;quot; to drive the plot forward. So, N = 1 so far, but Claude didn&#39;t predict my taste well on that one.)&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;= = =&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;EXPERIMENT #3: Writing style transfer&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;First, I gave Claude a few samples of my writing. (from my blog, personal texts, etc)&lt;/p&gt;
&lt;p&gt;Then, I did this iteration loop:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Ask Claude to write something in my style. (e.g. answer a question as me, rewrite the opening of &lt;em&gt;A Tale of Two Cities&lt;/em&gt; Nicky-style)&lt;/li&gt;
&lt;li&gt;I critique it. (e.g. &amp;quot;tone down the snark; my humour leans self-deprecating, almost never at the expense of others.&amp;quot;)&lt;/li&gt;
&lt;li&gt;Repeat so Claude gets better and better at imitating me!&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;Totally subjective impression, but I felt Claude went from 60% to 90% writing like me, in only an few learning rounds! Here&#39;s Claude answering &amp;quot;What&#39;s your favourite anime?&amp;quot; as me, first try:&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2025-07/clone_5.png&quot; alt=&quot;Claude predicts I&#39;d love Serial Experiments Lain, because of its philosophical themes, unsettling animation, and shy girl protagonist&quot; /&gt;&lt;/p&gt;
&lt;p&gt;(Now &lt;em&gt;THAT&#39;S&lt;/em&gt; 90+% my thinking and writing style! I haven&#39;t seen &lt;em&gt;Lain&lt;/em&gt; yet, but every other trans woman I know has recommended it to me. There&#39;s a good chance Claude&#39;s prediction about me may come true!)&lt;/p&gt;
&lt;p&gt;Finally, I did some meta-prompting: I prompted Claude to write a prompt &lt;em&gt;for getting Claude to write like me!&lt;/em&gt; Here&#39;s the start of what Claude learnt is &amp;quot;The Nicky Style&amp;quot;:&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2025-07/clone_6.png&quot; alt=&quot;Claude summarizes my writing style, including a roast about how I &amp;quot;include parenthetical asides (classic overthinking move!)&amp;quot;&quot; /&gt;&lt;/p&gt;
&lt;p&gt;Called the &lt;em&gt;fuck&lt;/em&gt; out??&lt;/p&gt;
&lt;p&gt;The most surprising result of this experiment, is that Claude can imitate my style &lt;em&gt;just from examples and a prompt;&lt;/em&gt; no need to do data-expensive &amp;amp; compute-expensive &amp;quot;fine tuning&amp;quot;. (Which can also introduce &lt;a href=&quot;https://www.nightfall.ai/ai-security-101/catastrophic-forgetting&quot;&gt;catastrophic forgetting&lt;/a&gt;; an acquaintance of mine once tried getting an LLM to write like them, by fine-tuning it on years of their emails. This caused the LLM to forget what 2 + 2 was.)&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Q:&lt;/strong&gt; Can Claude do a &amp;quot;style transfer&amp;quot; of my writing style?&lt;br /&gt;
&lt;strong&gt;A:&lt;/strong&gt; 🎉 &lt;code&gt;YES&lt;/code&gt; 🎉&lt;/p&gt;
&lt;p&gt;= = =&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;My next plans:&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;I&#39;ll keep adding memories to the database. It&#39;s a fun journalling exercise, a trip down Nostalgia Road. I&#39;m currently going through memories from Age 0 to 10, my childhood in Singapore. Wow, I haven&#39;t thought about Mr. Kiasu or the &lt;em&gt;Shiok!&lt;/em&gt; water donut ride in years.&lt;/li&gt;
&lt;li&gt;I&#39;ll keep coding &amp;amp; iterating this system, until it&#39;s finally good enough to fool a friend &lt;em&gt;at least once&lt;/em&gt;... then I keep going until it fools friends 10% of the time, 20%, 30%, 40%, 50% (coin flip!), &lt;em&gt;above&lt;/em&gt; 50% (the AI seems &amp;quot;more Nicky than Nicky&amp;quot;!)&lt;/li&gt;
&lt;li&gt;I&#39;m also tempted to do a &lt;em&gt;Fan&lt;/em&gt; Turing Test: can &lt;em&gt;you&lt;/em&gt; supporters tell the difference between &lt;em&gt;my&lt;/em&gt; bespoke human explanations, and AI slop? I sure hope so otherwise what the fuck are either you or me doing with our lives&lt;/li&gt;
&lt;/ul&gt;
&lt;hr /&gt;
&lt;p&gt;&lt;a id=&quot;ai_wargame&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h2&gt;🌏 Wargames for Peace&lt;/h2&gt;
&lt;p&gt;&lt;em&gt;(⏱️ 5 min read)&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;This is the project I worked on (but alas, didn&#39;t finish) for &lt;a href=&quot;https://www.matsprogram.org/&quot;&gt;the 10-week AI Safety bootcamp!&lt;/a&gt;&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Here&#39;s my poster:&lt;/strong&gt; (&lt;a href=&quot;https://blog.ncase.me/content/stuff/2025-07/poster.png&quot;&gt;hi-res image&lt;/a&gt;, transcript to follow)&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2025-07/poster.png&quot; alt=&quot;Full poster - see later for full transcript&quot; /&gt;&lt;/p&gt;
&lt;p&gt;(Please excuse the &lt;em&gt;placeholder&lt;/em&gt; diffusion-generated slop! The plan was &lt;em&gt;always&lt;/em&gt; to replace them later with actual artist-drawn art. You know, an entity that can actually draw hands &amp;amp; grids reliably.)&lt;/p&gt;
&lt;p&gt;(Also: &amp;quot;wargame&amp;quot; just means &amp;quot;a roleplaying game for policymakers&amp;quot;, &lt;em&gt;not&lt;/em&gt; necessarily war. Unfortunately it&#39;s gotta be marketed as &#39;wargame&#39; to get US govt people to try it, the same way you gotta market shampoo in black bottles to get American Joe to try it.)&lt;/p&gt;
&lt;p&gt;= = =&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Wargames for Peace:&lt;/strong&gt; a policymaking-simulation game, powered by AI, to
prepare us for AGI.&lt;/p&gt;
&lt;p&gt;by Nicky Case (mentored by Eli Lifland &amp;amp; Daniel Kokotajlo)&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Problem:&lt;/strong&gt; Preparing for the future(s) is hard.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Solution:&lt;/strong&gt; Policymakers already have a tradition of using roleplay &amp;quot;wargames&amp;quot; to prepare for global challenges. So, we made a roleplaying game about the arrival of AGI.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Problem:&lt;/strong&gt; Traditional &amp;quot;wargames&amp;quot; take lots of time &amp;amp; people to run. This limits the possible futures we can prepare for, and less room to alter/challenge our own assumptions.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Solution:&lt;/strong&gt; &lt;em&gt;We augmented our game with LLMs;&lt;/em&gt; the &amp;quot;Game Master&amp;quot; is an LLM, and all roles can be played by either humans or LLMs. It&#39;s now easy to run many, many scenario-roleplays!&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;For future work, we could:&lt;/strong&gt; Run 100s of &lt;em&gt;all-LLM&lt;/em&gt; simulations, as a &amp;quot;Monte Carlo&amp;quot; of the future. &amp;quot;Gradient descent&amp;quot; the AI Safety Community player, to find good strategies to make AGI go well. Modify this game for other X-risks (e.g. biotech pandemics).&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2025-07/game_1.png&quot; alt=&quot;Intro screen of the game&quot; /&gt;&lt;/p&gt;
&lt;p&gt;1) Players can start a new game, or join an existing game. (This game will be free &amp;amp; online, so even public citizens can participate.)&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2025-07/game_2.png&quot; alt=&quot;Screen: choose your character&quot; /&gt;&lt;/p&gt;
&lt;p&gt;2) Players choose who to roleplay as: governments, companies, the AGI itself, etc. (Any roles &lt;em&gt;not&lt;/em&gt; chosen by a Human will be played by an LLM.)&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2025-07/game_3.png&quot; alt=&quot;Screen: type your proposed action for the next round&quot; /&gt;&lt;/p&gt;
&lt;p&gt;3) Each round, players propose what they&#39;d like to do in the next timeframe. (Proposals may or may not succeed.)&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2025-07/game_4.png&quot; alt=&quot;Screen: you see everyone tried to do, and the consequences.&quot; /&gt;&lt;/p&gt;
&lt;p&gt;4) The &amp;quot;Game Master&amp;quot; LLM reasons about the proposals&#39; consequences, and writes a &amp;quot;news article&amp;quot; about what happens next. Go back to Step 3, repeat!&lt;/p&gt;
&lt;p&gt;= = =&lt;/p&gt;
&lt;p&gt;Here&#39;s a demo video:&lt;/p&gt;
&lt;iframe width=&quot;640&quot; height=&quot;360&quot; src=&quot;https://www.youtube-nocookie.com/embed/9_zSzjTWr8A?si=_GylrsjSuTC5K0vN&amp;rel=0&quot; title=&quot;YouTube video player&quot; frameborder=&quot;0&quot; allow=&quot;accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share&quot; referrerpolicy=&quot;strict-origin-when-cross-origin&quot; allowfullscreen=&quot;&quot;&gt;&lt;/iframe&gt;
&lt;p&gt;Yes it&#39;s basically Jackbox.&lt;/p&gt;
&lt;p&gt;= = =&lt;/p&gt;
&lt;p&gt;And if you want &lt;em&gt;even more&lt;/em&gt; details, here was the Research Plan I wrote, halfway into the program. I&#39;m proud of the pun title: &lt;strong&gt;&lt;a href=&quot;https://docs.google.com/document/d/1YbgZt3bSL7NcdB2UY_rmtQaWwQfrSxUsVXax-L9hznA/edit?usp=sharing&quot;&gt;Large Language Model UN&lt;/a&gt;&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;Unfortunately, I ran out of time during the program to actually finish this, and I&#39;m unsure if I actually &lt;em&gt;will&lt;/em&gt; have the capacity to. But! The folks I was being mentored by – who also co-authored &lt;a href=&quot;https://ai-2027.com/&quot;&gt;the AI 2027 forecast&lt;/a&gt; – are planning to build upon my code and make a full-fledged game!&lt;/p&gt;
&lt;p&gt;👉 &lt;strong&gt;If you&#39;re a game dev/designer interested in bringing this game to life, check out &lt;a href=&quot;https://www.lesswrong.com/posts/TjT3RdAfmrLqgb68K/help-the-ai-2027-team-make-an-online-agi-wargame&quot;&gt;their job posting&lt;/a&gt; &amp;amp; &lt;a href=&quot;https://airtable.com/appmcCGBz3vqARP9n/page6tXxhBJIoMCAq/form&quot;&gt;send them a message&lt;/a&gt;!&lt;/strong&gt;&lt;/p&gt;
&lt;hr /&gt;
&lt;p&gt;Anyway! Now &lt;em&gt;that&#39;s&lt;/em&gt; a long, proper update. We are so back.&lt;/p&gt;
&lt;p&gt;💸 Reminder: if you&#39;d like to support my explainers &amp;amp; research, consider &lt;a href=&quot;https://www.patreon.com/ncase&quot;&gt;my Patreon&lt;/a&gt; (monthly) or &lt;a href=&quot;https://ko-fi.com/nickycase&quot;&gt;Ko-Fi&lt;/a&gt; (one-off or monthly)! You&#39;ll also get early access to updates. Nooooo pressure of course.&lt;/p&gt;
&lt;p&gt;See you next month!&lt;br /&gt;
~ Nicky Case&lt;/p&gt;
&lt;iframe src=&quot;https://ncase.me/ncase-credits/signup.html&quot; frameborder=&quot;no&quot; width=&quot;640&quot; height=&quot;200&quot;&gt;&lt;/iframe&gt;</content>
        </entry>

    

        
        
        <entry>
            <title>Signal Boosts for Nov 2024</title>
            <link href="https://blog.ncase.me/signal-boosts-nov-2024/"/>
            <updated>2024-11-30T00:00:00Z</updated>
            <id>https://blog.ncase.me/signal-boosts-nov-2024/</id>
            <content xml:lang="en" type="html">&lt;p&gt;(⏱️ &lt;em&gt;~10 min read&lt;/em&gt;)&lt;/p&gt;
&lt;p&gt;Another month, another collection of shiny objects (stuff I found in the the past month that were valuable/inspiring).&lt;/p&gt;
&lt;p&gt;See also, previous Signal Boosts for &lt;a href=&quot;https://blog.ncase.me/signal-boosts-oct-2024/&quot;&gt;Oct/Sep&lt;/a&gt;, &lt;a href=&quot;https://blog.ncase.me/signal-boosts-aug-2024/&quot;&gt;Aug/July&lt;/a&gt;, &lt;a href=&quot;https://blog.ncase.me/signal-boosts-june-2024/&quot;&gt;June&lt;/a&gt;, &lt;a href=&quot;https://blog.ncase.me/signal-boosts-may-2024/&quot;&gt;May&lt;/a&gt;&lt;/p&gt;
&lt;ol&gt;
&lt;li&gt;🩸 Kurzgesagt gets to the bottom of a science myth &lt;a href=&quot;https://blog.ncase.me/signal-boosts-nov-2024/#blood&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;😭 Paper: AI poems are now &amp;quot;more human than human&amp;quot; &lt;a href=&quot;https://blog.ncase.me/signal-boosts-nov-2024/#poem&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;🇺🇸 U.S. Election-related analysis &lt;a href=&quot;https://blog.ncase.me/signal-boosts-nov-2024/#election&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;👄 The Substance &lt;a href=&quot;https://blog.ncase.me/signal-boosts-nov-2024/#substance&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;/ol&gt;
&lt;hr /&gt;
&lt;p&gt;&lt;a id=&quot;blood&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h2&gt;🩸 Kurzgesagt gets to the bottom of a science myth&lt;/h2&gt;
&lt;blockquote&gt;
&lt;p&gt;“If you were to lie your blood vessels out entirely from end-to-end, &lt;s&gt;you would die&lt;/s&gt; it would stretch 100,000 km, wrapping twice across the globe!”&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;It&#39;s not just a popular factoid, it&#39;s found in university lectures &amp;amp; scientific papers! So it must be tr--&lt;/p&gt;
&lt;p&gt;--wait, that paper didn&#39;t give a source. Ok, &lt;em&gt;this&lt;/em&gt; one gives a source, but its source just says the &amp;quot;fact&amp;quot; without giving a source. Uh oh.&lt;/p&gt;
&lt;p&gt;This is the story of how Kurzgesagt went on a deep, &lt;em&gt;deep&lt;/em&gt; rabbit-hole to track down a source for &lt;em&gt;one&lt;/em&gt; (very popular, even &amp;quot;official&amp;quot;) science-myth!  I like this video because:&lt;/p&gt;
&lt;ol&gt;
&lt;li&gt;It exemplifies the virtue of rigorous scholarship, the thankless task of factchecking;&lt;/li&gt;
&lt;li&gt;It&#39;s a cautionary tale of how the telephone-game of &amp;quot;facts&amp;quot; can infect even the so-called official sources; and&lt;/li&gt;
&lt;li&gt;Their half-real-life-half-animation style for this video is &lt;em&gt;so&lt;/em&gt; adorable!&lt;/li&gt;
&lt;/ol&gt;
&lt;p&gt;Check it out! You&#39;ll also learn the &lt;em&gt;real&lt;/em&gt; answer for how long your blood vessels are in total. It&#39;s not as impressive as &amp;quot;can wrap twice around the world&amp;quot;... but it&#39;s still pretty dang long!&lt;/p&gt;
&lt;p&gt;10-minute video (not counting credits):&lt;/p&gt;
&lt;iframe width=&quot;640&quot; height=&quot;360&quot; src=&quot;https://www.youtube-nocookie.com/embed/bgo7rm5Maqg?si=4mnRfK5jTYobTqp-&amp;rel=0&quot; title=&quot;YouTube video player&quot; frameborder=&quot;0&quot; allow=&quot;accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share&quot; referrerpolicy=&quot;strict-origin-when-cross-origin&quot; allowfullscreen=&quot;&quot;&gt;&lt;/iframe&gt;
&lt;hr /&gt;
&lt;p&gt;&lt;a id=&quot;poem&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h2&gt;😭 Paper: AI poems are now &amp;quot;more human than human&amp;quot;&lt;/h2&gt;
&lt;p&gt;Poets speak to the human condition, and the human condition has spoken back: they prefer the autocomplete.&lt;/p&gt;
&lt;p&gt;From &lt;a href=&quot;https://www.nature.com/articles/s41598-024-76900-1&quot;&gt;Porter &amp;amp; Machery 2024&lt;/a&gt;: (emphases added)&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;We conducted two experiments with non-expert poetry readers and found that participants performed &lt;strong&gt;below chance&lt;/strong&gt; levels in identifying AI-generated poems (46.6% accuracy)&lt;/p&gt;
&lt;p&gt;[...] participants were &lt;strong&gt;more likely to judge AI-generated poems as human-authored&lt;/strong&gt; than actual human-authored poems&lt;/p&gt;
&lt;p&gt;[...] We found that AI-generated poems were rated &lt;strong&gt;more favorably in qualities such as rhythm and beauty&lt;/strong&gt;, and that this contributed to their mistaken identification as human-authored.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;And the human poems were from famous poets across eras &amp;amp; genres (from Chaucer to Whitman to Plath). And the AI poems were the &lt;em&gt;first five&lt;/em&gt; poems generated by ChatGPT with an unoptimized prompt, no cherry-picking.&lt;/p&gt;
&lt;p&gt;That&#39;s... pretty damn damning.&lt;/p&gt;
&lt;p&gt;Like the original Turing Test, I&#39;m not sure how much of this is &amp;quot;AI has gotten too good&amp;quot; versus &amp;quot;Humans have gotten too gullible&amp;quot;.&lt;/p&gt;
&lt;p&gt;(See also: Astral Codex Ten&#39;s &lt;a href=&quot;https://www.astralcodexten.com/p/how-did-you-do-on-the-ai-art-turing&quot;&gt;&amp;quot;AI Art Turing Test&amp;quot;&lt;/a&gt;, where humans (on avg) &lt;em&gt;also&lt;/em&gt; preferred AI art over human art when they think it&#39;s human. Although, 2 major differences from ACX&#39;s study vs the above: 1) ACX&#39;s test &lt;em&gt;did&lt;/em&gt; cherry-pick the AI art, and 2) Even then, humans &lt;em&gt;were&lt;/em&gt; slightly better than chance (~60%) at detecting AI art.)&lt;/p&gt;
&lt;hr /&gt;
&lt;p&gt;&lt;a id=&quot;election&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h2&gt;🇺🇸 U.S. Election-related analysis&lt;/h2&gt;
&lt;p&gt;My political alignment is &amp;quot;why would any sane person try to compress their beliefs into one or two dimensions, just ask me my thoughts on an issue-by-issue basis&amp;quot;. But if I &lt;em&gt;had&lt;/em&gt; to pick an approximation, sure, put me down somewhere between left &amp;amp; libertarian.&lt;/p&gt;
&lt;p&gt;And so, like many in the left/libertarian circles, while I wasn&#39;t &lt;em&gt;shocked&lt;/em&gt; Trump won again, I was disappointed. Sure, America&#39;s institutions are broken a.f., but Americans deserve a better solution than &lt;em&gt;that&lt;/em&gt; guy. (I&#39;m a Canadian citizen, so I could not vote in the States. But Canada shouldn&#39;t be so smug; the political culture of Canada, and most Western countries, are IMHO only a decade behind America&#39;s.)&lt;/p&gt;
&lt;p&gt;It&#39;s tempting to interpret whatever happens as evidence for everything I already believed, so, here&#39;s a Signal Boost highlighting data &amp;amp; analyses that &lt;em&gt;were&lt;/em&gt; surprising to me:&lt;/p&gt;
&lt;p&gt;1) The &lt;strong&gt;&lt;a href=&quot;https://apnews.com/projects/election-results-2024/votecast/&quot;&gt;Associated Press VoteCast, a survey of 110,000+ voters&lt;/a&gt;.&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;The biggest surprise to me: Trump made the most gains amongst &lt;em&gt;Blacks, Hispanics, 18-29 year olds, and non-college folks&lt;/em&gt;. Meanwhile, it was &lt;em&gt;white urban highly-educated men&lt;/em&gt; who Harris gained ground on:&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2024-11/election1.png&quot; alt=&quot;Graph showing the above&quot; /&gt;&lt;/p&gt;
&lt;p&gt;(Hat tip &lt;a href=&quot;https://docs.monadical.com/s/so-you-want-to-build-a-social-network&quot;&gt;to Nick Sweeting&lt;/a&gt; for being the first to show me this dataviz. Note 1: these are &lt;em&gt;relative&lt;/em&gt; percentages, not &lt;em&gt;absolute&lt;/em&gt; percentage-points (&lt;a href=&quot;https://xkcd.com/985/&quot;&gt;relevant xkcd&lt;/a&gt;). Note 2: I couldn&#39;t find &lt;em&gt;who&lt;/em&gt; made the above dataviz, but it does accurately show the AP VoteCast data.)&lt;/p&gt;
&lt;p&gt;So, I think this is evidence against &amp;quot;Trump won because of racism or Boomers&amp;quot;.&lt;/p&gt;
&lt;p&gt;. . .&lt;/p&gt;
&lt;p&gt;2) &lt;strong&gt;&lt;a href=&quot;https://ericneyman.wordpress.com/2024/11/14/seven-lessons-i-didnt-learn-from-election-day/&quot;&gt;&amp;quot;7 Lessons I Didn&#39;t Learn From Election Day&amp;quot;&lt;/a&gt; by Eric Neyman&lt;/strong&gt; (found through &lt;a href=&quot;https://news.manifold.markets/&quot;&gt;the Manifold Markets newsletter&lt;/a&gt;)&lt;/p&gt;
&lt;p&gt;The biggest insight I got from that post: &lt;strong&gt;for the first time &lt;em&gt;ever&lt;/em&gt;, incumbent parties in &lt;em&gt;every developed country&lt;/em&gt; got their asses kicked this year.&lt;/strong&gt;  If anything, Harris did really well, for a representative of the incumbent party:&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2024-11/election2.png&quot; alt=&quot;Graph showing the above&quot; /&gt;&lt;/p&gt;
&lt;p&gt;So, I think this is evidence against &amp;quot;The world is shifting right&amp;quot;. (The UK Labour Party resoundingly beat the Tories)&lt;/p&gt;
&lt;p&gt;However, this is probably evidence for &amp;quot;It&#39;s the economy, stupid&amp;quot;.&lt;/p&gt;
&lt;p&gt;(Although, I note I&#39;m confused: &lt;a href=&quot;https://www.bls.gov/charts/consumer-price-index/consumer-price-index-by-category-line-chart.htm&quot;&gt;2022 was &lt;em&gt;worse&lt;/em&gt; inflation-wise than 2024&lt;/a&gt;, yet as seen above, incumbents did ok in 2022? Democrats avoided a &amp;quot;red wave&amp;quot; in 2022 midterms. Is it wages relative to inflation? No, &lt;a href=&quot;https://www.statista.com/statistics/1351276/wage-growth-vs-inflation-us/&quot;&gt;wages outgrew inflation starting Feb 2023&lt;/a&gt;. Maybe it&#39;s not the &lt;em&gt;actual&lt;/em&gt; economy, but people&#39;s &lt;em&gt;perception&lt;/em&gt; of the economy that matters? No, &lt;a href=&quot;https://www.oecd.org/en/data/indicators/consumer-confidence-index-cci.html&quot;&gt;consumer confidence was also &lt;em&gt;worse&lt;/em&gt; in 2022&lt;/a&gt;. Maybe Roe v Wade being overturned helped the Dems in 2022? Maybe, but it doesn&#39;t explain why most incumbents around the world &lt;em&gt;also&lt;/em&gt; held in 2022.)&lt;/p&gt;
&lt;p&gt;. . .&lt;/p&gt;
&lt;p&gt;2.5) Not directly related to the 2024 election, but &amp;quot;It&#39;s the economy, stupid&amp;quot; reminded me of Douglas Hibb&#39;s famous &lt;strong&gt;&lt;a href=&quot;https://gupea.ub.gu.se/bitstream/handle/2077/2847/gunwpe0020.pdf?sequence=1&quot;&gt;Bread &amp;amp; Peace Model&lt;/a&gt;&lt;/strong&gt;, which found that &amp;quot;does the incumbent win the election&amp;quot; is pretty much determined by real income growth since the last election cycle, with exceptions for major wars:&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2024-11/election3.png&quot; alt=&quot;Graph showing the above&quot; /&gt;&lt;/p&gt;
&lt;p&gt;From the paper, to rub salt in the wounds of all pundits &amp;amp; political scientists:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;The model is subjected to robustness tests against twenty-two variations [...] inspired by the extensive literature on presidential voting. Not one of these variations adds value to the Bread and Peace model[.]&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;(I haven&#39;t done enough research yet to tell if the Bread &amp;amp; Peace model holds for other countries. The above paper was published before the 2000 US election, but &lt;a href=&quot;https://files.osf.io/v1/resources/xrf3t/providers/osfstorage/5f7581b81cfe6900b0cefd24?action=download&amp;amp;direct&amp;amp;version=7&quot;&gt;Bread &amp;amp; Peace mostly held up since then.&lt;/a&gt;)&lt;/p&gt;
&lt;p&gt;(Thinking aloud: Hm... maybe the Israel-Hamas war explains why the incumbent did worse in 2024 than 2022, even though the economy&#39;s better in 2024 than 2022? Then again, 2022 was also the start of the Russia-Ukraine war. Hm.)&lt;/p&gt;
&lt;p&gt;. . .&lt;/p&gt;
&lt;p&gt;3) Sick of numbers? Here&#39;s a good &lt;em&gt;qualitative&lt;/em&gt; look at &lt;strong&gt;&lt;a href=&quot;https://x.com/smquinsaat/status/1854187349456171179&quot;&gt;immigrant conservatives in Hawaii&lt;/a&gt;, fieldwork by Sharon Quinsaat.&lt;/strong&gt; A few highlights, on why (these) Hawaiian immigrants went for Trump:&lt;/p&gt;
&lt;ol&gt;
&lt;li&gt;It&#39;s the economy, stupid. (They feel Dems &amp;quot;sold us out&amp;quot;)&lt;/li&gt;
&lt;li&gt;At the risk of sounding condescending, their media sources actually &lt;em&gt;are&lt;/em&gt; really wacky. Like, &amp;quot;Trump is a demi-god&amp;quot; celebrity-worship wacky.&lt;/li&gt;
&lt;li&gt;Payback. &amp;quot;I&#39;m voting Republican just to stick it to the Democrats.&amp;quot;&lt;/li&gt;
&lt;/ol&gt;
&lt;p&gt;. . .&lt;/p&gt;
&lt;p&gt;4) &lt;strong&gt;&lt;a href=&quot;https://podcasts.apple.com/us/podcast/the-ezra-klein-show/id1548604447&quot;&gt;Ezra Klein&lt;/a&gt;&lt;/strong&gt;, as usual, is pretty good. I&#39;ve only listened to few of his post-election podcasts so far, but here&#39;s my main takeaways:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;The old left-right spectrum is pretty much dead.
&lt;ul&gt;
&lt;li&gt;(Trump is anti-free trade &amp;amp; pro-tariffs. His Health Secretary pick, RFK Jr., is pro-choice. Who, 10 years ago, woulda thought a Republican president would do &lt;em&gt;that?&lt;/em&gt;)&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;The new political divide is something along the lines of:
&lt;ul&gt;
&lt;li&gt;Pro-institution, vs Anti-institution (&amp;quot;we&#39;re sick of experts&amp;quot;)&lt;/li&gt;
&lt;li&gt;The college-credentialed/professional class (&amp;quot;trust science!&amp;quot;), vs everyone else&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;The Democrats failed, and &lt;em&gt;institutions&lt;/em&gt; failed, because of &amp;quot;cover your ass&amp;quot; &amp;amp; not wanting to be the messenger with bad-news messages. (e.g. staffers ignoring Biden&#39;s declining sharpness, the &amp;quot;noble lies&amp;quot; during the pandemic)&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;Ezra is someone who wants to critique institutions, &lt;em&gt;because&lt;/em&gt; he wants them to get better. I&#39;m reminded of this quote:&lt;sup class=&quot;footnote-ref&quot;&gt;&lt;a href=&quot;https://blog.ncase.me/signal-boosts-nov-2024/#fn1&quot; id=&quot;fnref1&quot;&gt;[1]&lt;/a&gt;&lt;/sup&gt;&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;The iron rule of politics is that if there are real problems in society and responsible parties don’t deal with them, the irresponsible parties will jump on them.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;. . .&lt;/p&gt;
&lt;p&gt;In sum, my current thoughts (subject to future change, of course):&lt;/p&gt;
&lt;ol&gt;
&lt;li&gt;&amp;quot;Bread &amp;amp; Peace model&amp;quot; probably holds. It&#39;s the economy, stupid.&lt;/li&gt;
&lt;li&gt;Incumbents everywhere are falling, whether they&#39;re left or right. (Tories lost UK &lt;em&gt;hard&lt;/em&gt;.)&lt;/li&gt;
&lt;li&gt;The old &amp;quot;left-right&amp;quot; spectrum is dead. (It seems even racial/gender identity is getting much &lt;em&gt;less&lt;/em&gt; politically salient in the US, with Black/Latino voters swinging to Trump.)&lt;/li&gt;
&lt;li&gt;The new spectrum is &amp;quot;pro vs anti traditional institutions&amp;quot;. (universities, journalism, government, international orgs, etc)&lt;/li&gt;
&lt;li&gt;Institutions everywhere are failing because of &amp;quot;cover your ass&amp;quot; bureaucracy.&lt;/li&gt;
&lt;/ol&gt;
&lt;p&gt;I don&#39;t have concrete suggestions. Only this:&lt;/p&gt;
&lt;p&gt;Deal with the real problems in institutions you care about — &lt;em&gt;or the irresponsible parties will jump on them.&lt;/em&gt;&lt;/p&gt;
&lt;hr /&gt;
&lt;p&gt;&lt;a id=&quot;substance&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h2&gt;👄 The Substance&lt;/h2&gt;
&lt;p&gt;A few weeks ago, I got a email saying I was turned down for a research gig, which was understandable — but it hit some of my insecurities that &lt;a href=&quot;https://blog.ncase.me/30/&quot;&gt;now that I&#39;m 30&lt;/a&gt;, I&#39;m too old for great research, and I oughta be replaced with someone younger. (&lt;a href=&quot;https://xkcd.com/447/&quot;&gt;relevant xkcd&lt;/a&gt;)  So, to get my mind off my worries, I went to see the new movie &lt;strong&gt;The Substance&lt;/strong&gt;, which I knew nothing about. It turned out to be a sci-fi satire body-horror-comedy, about a woman who&#39;s insecure that she&#39;s too old, and she&#39;ll be replaced with someone younger.&lt;/p&gt;
&lt;p&gt;Aaaaaaaaa&lt;/p&gt;
&lt;p&gt;&lt;u&gt;My quick review&lt;/u&gt;: Honestly the plot&#39;s straightforward &amp;amp; predictable, and not exactly subtle about its themes of sexism in industry + obsession with youth/fame + self-sabotaging self-loathing. But, I have a high tolerance for camp (&lt;em&gt;loved&lt;/em&gt; Rocky Horror), and this film happened to hit me at the &lt;em&gt;exact&lt;/em&gt; moment I needed to confront my insecurities that I&#39;m too old &amp;amp; I&#39;ll be replaced with someone younger.&lt;/p&gt;
&lt;p&gt;Because, yeah, I &lt;em&gt;am&lt;/em&gt; being silly. Tolkien was 45 when he started writing The Lord of the Rings, Andrew Wiles proved Fermat&#39;s Last Theorem between age 36 to 41. Sure, my probability of doing great work gets &lt;em&gt;lower&lt;/em&gt; over time, but it&#39;s not zero.&lt;/p&gt;
&lt;p&gt;Also, the fact this movie features Demi Moore, who&#39;s still smokin&#39; hot at age 62, makes this a movie that &lt;em&gt;disproves its own premise&lt;/em&gt;, in a good way! You &lt;em&gt;can&lt;/em&gt; age with confidence &amp;amp; style.&lt;/p&gt;
&lt;p&gt;Seriously, gott dayumn, Demi:&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2024-11/demi.png&quot; alt=&quot;A hot 62-year-old actress, Demi Moore, in the film, The Substance&quot; /&gt;&lt;/p&gt;
&lt;p&gt;★★★★☆: Recommended if you need to confront your insecurities, via a campy body-horror movie starring a hot MILF. Also the cinematography &amp;amp; &lt;a href=&quot;https://www.youtube.com/watch?v=tKxKuYsx2R0&quot;&gt;: the soundtrack&lt;/a&gt; &lt;em&gt;kicks ass&lt;/em&gt;.&lt;/p&gt;
&lt;p&gt;(Tangentially related: while reading Demi Moore&#39;s Wikipedia article, I learnt she co-founded &lt;a href=&quot;https://en.wikipedia.org/wiki/Thorn_(organization)&quot;&gt;Thorn&lt;/a&gt;?! I knew about Thorn, because I&#39;m getting into AI Safety/Ethics research, and one of Thorn&#39;s famous projects is using AI to help automatically flag child abuse material! [&lt;a href=&quot;https://www.artificialintelligence-news.com/news/ai-tool-detect-child-abuse-images-accuracy/&quot;&gt;read more here&lt;/a&gt;] Good work, Demi!)&lt;/p&gt;
&lt;hr class=&quot;footnotes-sep&quot; /&gt;
&lt;section class=&quot;footnotes&quot;&gt;
&lt;ol class=&quot;footnotes-list&quot;&gt;
&lt;li id=&quot;fn1&quot; class=&quot;footnote-item&quot;&gt;&lt;p&gt;Quote from Daniel Schwammenthal, director of the American Jewish Committee’s Transatlantic Institute. Learnt about this quote &lt;a href=&quot;https://ofboysandmen.substack.com/p/into-the-vacuum-demons-pour&quot;&gt;from this article by Richard Reeves&lt;/a&gt;, on why neglecting real issues that disproportionally affect males (men are more likely than women to be homeless, die of suicide, drop out of school, etc) led to the rise of grifters like Andrew Tate. &lt;a href=&quot;https://blog.ncase.me/signal-boosts-nov-2024/#fnref1&quot; class=&quot;footnote-backref&quot;&gt;↩︎&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ol&gt;
&lt;/section&gt;
</content>
        </entry>

    

        
        
        <entry>
            <title>Signal Boosts for Sept &amp; Oct 2024</title>
            <link href="https://blog.ncase.me/signal-boosts-oct-2024/"/>
            <updated>2024-10-27T00:00:00Z</updated>
            <id>https://blog.ncase.me/signal-boosts-oct-2024/</id>
            <content xml:lang="en" type="html">&lt;p&gt;(⏱️ &lt;em&gt;~8 min read&lt;/em&gt;)&lt;/p&gt;
&lt;p&gt;Once again I got swamped with work so I missed last month&#39;s Signal Boost / Links post, so here&#39;s a two-fer, for all the stuff I recommend that I found last &amp;amp; this month.&lt;/p&gt;
&lt;p&gt;Previous months&#39; Signal Boosts: &lt;a href=&quot;https://blog.ncase.me/signal-boosts-aug-2024/&quot;&gt;July/Aug 2024&lt;/a&gt;, &lt;a href=&quot;https://blog.ncase.me/signal-boosts-june-2024/&quot;&gt;June 2024&lt;/a&gt;, &lt;a href=&quot;https://blog.ncase.me/signal-boosts-may-2024/&quot;&gt;May 2024&lt;/a&gt;&lt;/p&gt;
&lt;p&gt;This month&#39;s Signal Boosts:&lt;/p&gt;
&lt;ol&gt;
&lt;li&gt;🐦 Trans furry hyperpop conlang musical mockumentary &lt;a href=&quot;https://blog.ncase.me/signal-boosts-oct-2024/#zewei&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;👋 DynamicLand is back, bay-beeeee &lt;a href=&quot;https://blog.ncase.me/signal-boosts-oct-2024/#dynamicland&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;🍞 &amp;quot;I put a toaster in the dishwasher&amp;quot; &lt;a href=&quot;https://blog.ncase.me/signal-boosts-oct-2024/#toaster&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;🧢 &amp;quot;I wish I could wear hats&amp;quot; &lt;a href=&quot;https://blog.ncase.me/signal-boosts-oct-2024/#hats&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;🌕 &amp;quot;H.S.&amp;quot; &lt;a href=&quot;https://blog.ncase.me/signal-boosts-oct-2024/#hs&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;🐺 A 3D-printable claw &lt;a href=&quot;https://blog.ncase.me/signal-boosts-oct-2024/#claw&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;👟 Slip-on Shoes &lt;a href=&quot;https://blog.ncase.me/signal-boosts-oct-2024/#shoes&quot;&gt;↪&lt;/a&gt;&lt;/li&gt;
&lt;/ol&gt;
&lt;hr /&gt;
&lt;p&gt;&lt;a id=&quot;zewei&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h2&gt;🐦 Trans furry hyperpop conlang musical mockumentary&lt;/h2&gt;
&lt;p&gt;Previously, I signal-boosted the constructed language (conlang) &lt;a href=&quot;https://blog.ncase.me/signal-boosts-aug-2024/#toki-pona&quot;&gt;toki pona&lt;/a&gt;, and the trans furry hyperpop musician &lt;a href=&quot;https://blog.ncase.me/signal-boosts-aug-2024/#moretransfurryhyperpop&quot;&gt;Ida Deers&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;This month, a fan emailed me to tell me an artist who combines &lt;em&gt;both&lt;/em&gt;: &lt;strong&gt;&lt;a href=&quot;https://www.zewei.xyz/&quot;&gt;ZeWei&lt;/a&gt;&lt;/strong&gt;. (ok I guess their music&#39;s more IDM than hyperpop, but w/e)&lt;/p&gt;
&lt;p&gt;Recently, ZeWei created a mockumentary for a &amp;quot;cursed conlang&amp;quot; contest. Imagine &lt;em&gt;Arrival (2016)&lt;/em&gt; but furry. It&#39;s a short film about a linguist who gets stuck in an alien world full of cyborg-trees, whose language a) has 600 pronouns, b) the phonemes are microtonal music, and c) the formality system is &lt;em&gt;so&lt;/em&gt; intense, that the social authority of who you&#39;re talking to modifies not just Subject-Verb-Object order, but even order of &lt;em&gt;sentences and paragraphs&lt;/em&gt;. And that&#39;s just &lt;em&gt;three&lt;/em&gt; of the many messed-up features of this constructed language!&lt;/p&gt;
&lt;p&gt;And then at the end, as a total flex, ZeWei performs &lt;em&gt;an entire song&lt;/em&gt; in this language.&lt;/p&gt;
&lt;p&gt;22 minute video, very worth the watch:&lt;/p&gt;
&lt;iframe width=&quot;640&quot; height=&quot;360&quot; src=&quot;https://www.youtube-nocookie.com/embed/Ql0VKM7tCCo?rel=0&quot; title=&quot;YouTube video player&quot; frameborder=&quot;0&quot; allow=&quot;accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share&quot; referrerpolicy=&quot;strict-origin-when-cross-origin&quot; allowfullscreen=&quot;&quot;&gt;&lt;/iframe&gt;
&lt;p&gt;This artist also has a Bandcamp! &lt;a href=&quot;https://zewei.bandcamp.com/album/flickering&quot;&gt;Here&#39;s an older album from them&lt;/a&gt;, with accompanying YouTube release:&lt;/p&gt;
&lt;iframe width=&quot;640&quot; height=&quot;360&quot; src=&quot;https://www.youtube-nocookie.com/embed/O3jLuPHEXMA?rel=0&quot; title=&quot;YouTube video player&quot; frameborder=&quot;0&quot; allow=&quot;accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share&quot; referrerpolicy=&quot;strict-origin-when-cross-origin&quot; allowfullscreen=&quot;&quot;&gt;&lt;/iframe&gt;
&lt;p&gt;Check them out!&lt;/p&gt;
&lt;p&gt;&lt;a id=&quot;dynamicland&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h2&gt;👋 Dynamicland is back, bay-beeeee&lt;/h2&gt;
&lt;p&gt;&lt;strong&gt;&lt;a href=&quot;https://dynamicland.org/&quot;&gt;THEY&#39;RE BACK!&lt;/a&gt;&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;(after being closed for the pandemic &amp;amp; a bunch of internal restructuring)&lt;/p&gt;
&lt;p&gt;What is Dynamicland? Well first, &lt;em&gt;why&lt;/em&gt; is Dynamicland? Right now, the overall trend in computing is to be isolating, take away user agency, and limit ourselves to our thumbs. Dynamicland wants to reverse all those trends. Dynamicland is a lab, a proof-of-concept, of computing that&#39;s social, where everyone can be a creator, and you use your full bodies to interact with it.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Dynamicland is &lt;em&gt;a whole building that is a computer interface.&lt;/em&gt;&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;(Here&#39;s a 7-minute intro:)&lt;/p&gt;
&lt;iframe width=&quot;640&quot; height=&quot;360&quot; src=&quot;https://www.youtube-nocookie.com/embed/5Q9r-AEzRMA?rel=0&quot; title=&quot;YouTube video player&quot; frameborder=&quot;0&quot; allow=&quot;accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share&quot; referrerpolicy=&quot;strict-origin-when-cross-origin&quot; allowfullscreen=&quot;&quot;&gt;&lt;/iframe&gt;
&lt;p&gt;&lt;em&gt;Right now&lt;/em&gt;, Dynamicland accomplishes this by having cameras that track objects, and projectors to put displays on them. It&#39;s like mixed reality except it doesn&#39;t require that everyone buy their own $3,000+ headset. However, I need to emphasize: &lt;em&gt;this cameras-and-projectors system is NOT the point of Dynamicland,&lt;/em&gt; it&#39;s just the closest that can be accomplished with today&#39;s tech. (Maybe in the far future we&#39;ll have fully programmable matter?) As the saying goes: &lt;em&gt;&amp;quot;Look where I&#39;m pointing at, not at the tip of my finger.&amp;quot;&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;Two personal notes:&lt;/p&gt;
&lt;p&gt;1) The founder of Dynamicland, Bret Victor, greatly inspired me, is a good mentor, and invited me to the first Explorable Explanations workshop, which basically single-handed created my whole career over the past decade. So, this is a note of gratitude! (and/or disclosure of conflict-of-interest)&lt;/p&gt;
&lt;p&gt;2) A few years ago, at Dynamicland, I made a game called FROG WARS played with origami frogs. The goal was to hop your (physical) frogs towards (virtual) flies, while also bumping your opponents&#39; (physical) frogs out of the arena.&lt;/p&gt;
&lt;p&gt;It was the best game I ever made and I&#39;ll never live up to it again.&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2024-10/frogwars.gif&quot; alt=&quot;GIF of my Dynamicland game &amp;quot;FROG WARS&amp;quot; in action&quot; /&gt;&lt;/p&gt;
&lt;p&gt;&lt;a id=&quot;toaster&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h2&gt;🍞 &amp;quot;I put a toaster in the dishwasher&amp;quot;&lt;/h2&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2024-10/toaster.png&quot; alt=&quot;Before &amp;amp; after photos of putting a toaster in a dishwasher; it&#39;s much cleaner, and still works!&quot; /&gt;&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;&lt;a href=&quot;https://jdstillwater.blogspot.com/2012/05/i-put-toaster-in-dishwasher.html&quot;&gt;A lil&#39; anecdote from JD Stillwater&lt;/a&gt;.&lt;/strong&gt; Common sense says no of &lt;em&gt;course&lt;/em&gt; you shouldn&#39;t put a toaster in a dishwasher — and a &lt;em&gt;lot&lt;/em&gt; of sarcastic, over-confident internet comments back that up.&lt;/p&gt;
&lt;p&gt;But, if you actually think about the physics step-by-step, an electrical short happens when there&#39;s an unintended path for electricity to travel (e.g. through water), which can cause damage. So, if you were to put an unplugged toaster in a dishwasher, &lt;em&gt;then let it dry for several days&lt;/em&gt;, there would be no leftover water in it (toasters are &lt;em&gt;designed&lt;/em&gt; to avoid trapping moisture), and thus, no risk of an electrical short.&lt;/p&gt;
&lt;p&gt;But that&#39;s all theory. Toasters are cheap. So, JD Stillwater just &lt;em&gt;tried it&lt;/em&gt;, and,&lt;/p&gt;
&lt;p&gt;yup, it worked!&lt;/p&gt;
&lt;p&gt;Signal boosting this because it&#39;s a great short (real-life) story, that shows the conflict between &lt;em&gt;seeming&lt;/em&gt; scientific &amp;amp; rational, and &lt;em&gt;actually being&lt;/em&gt; scientific &amp;amp; rational. Similar vibes to &lt;a href=&quot;https://calteches.library.caltech.edu/51/2/CargoCult.htm&quot;&gt;Feynman&#39;s famous Cargo Cult Science talk&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;(Hat tip to Slime Mold Time Mold &lt;a href=&quot;https://slimemoldtimemold.com/2024/08/30/links-for-august-2024/&quot;&gt;for helping me discover this essay&lt;/a&gt;, and kudos to Nehaveigur &lt;a href=&quot;https://nehaveigur.com/2024/09/05/i-put-my-toaster-in-the-dishwasher/&quot;&gt;for successfully replicating this experiment!&lt;/a&gt;)&lt;/p&gt;
&lt;p&gt;(That said, I&#39;m still too cowardly to &lt;em&gt;actually&lt;/em&gt; try putting a toaster in a dishwasher. If &lt;em&gt;you&lt;/em&gt; want to try replicating this experiment, then: 1) &lt;em&gt;Please let it dry for several days before plugging it back in.&lt;/em&gt; 2) Just in case, get a large fire extinguisher graded for electrical fires (you should have a fire extinguisher anyway), and 3) Neither I nor the authors of the above links are responsible if you die, lol.)&lt;/p&gt;
&lt;p&gt;&lt;a id=&quot;hats&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h2&gt;🧢 &amp;quot;I wish I could wear hats&amp;quot;&lt;/h2&gt;
&lt;p&gt;Last month, a friend introduced me to the works of Brian David Gilbert, and...&lt;/p&gt;
&lt;p&gt;... look, I can&#39;t explain BDG. I can only beseech you to look &amp;amp; listen to this 98-second video:&lt;/p&gt;
&lt;iframe width=&quot;640&quot; height=&quot;360&quot; src=&quot;https://www.youtube-nocookie.com/embed/bmFGbBmlyKQ?rel=0&quot; title=&quot;YouTube video player&quot; frameborder=&quot;0&quot; allow=&quot;accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share&quot; referrerpolicy=&quot;strict-origin-when-cross-origin&quot; allowfullscreen=&quot;&quot;&gt;&lt;/iframe&gt;
&lt;p&gt;&lt;em&gt;Right?!&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;&lt;em&gt;&amp;quot;Heh, that&#39;s a charming silly idea. Oh. Oh okay it&#39;s body dysmorphia and possibly gender dysphoria. Oh okay that hat at the end is wearing a hat. Oh.&amp;quot;&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;But don&#39;t worry, &lt;a href=&quot;https://www.youtube.com/watch?v=Sfw9n2ktX9k&quot;&gt;not all his songs are metaphors&lt;/a&gt;. One time he just published &lt;a href=&quot;https://www.youtube.com/watch?v=-wpHszfnJns&quot;&gt;a half-hour-long video explaining healthcare terminology&lt;/a&gt;. It was actually upsettingly useful; I made like 20+ Anki flashcards off that video.&lt;/p&gt;
&lt;p&gt;&lt;a id=&quot;hs&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h2&gt;🌕 &amp;quot;H.S&amp;quot;&lt;/h2&gt;
&lt;p&gt;Just a vague cherry-picked hypothesis, but I&#39;ve noticed at least two (&amp;amp; a half) major &amp;quot;vibe eras&amp;quot; in Western comedy during my lifetime:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;The 90&#39;s &amp;amp; 00&#39;s were the peak of edgy transgressive humor, like (early) South Park, (early) Family Guy, Dane Cook, etc.&lt;/li&gt;
&lt;li&gt;Perhaps, as a reaction to the hollowness of &amp;quot;being edgy for edgy&#39;s sake&amp;quot;, recent decades had more self-deprecating, introspective humor, like: Bo Burnham, Bojack Horseman, and the comics of Shen Comix, Sarah&#39;s Scribbles, and Hyperbole &amp;amp; a Half. (Even South Park has become more reflective &amp;amp; character/story-driven.)&lt;/li&gt;
&lt;li&gt;And now, perhaps as a reaction to how much of a downer &lt;em&gt;that&lt;/em&gt; vibe is — especially &amp;quot;post&amp;quot;-pandemic — comedy might now be shifting towards a &amp;quot;WE ARE SO BACK&amp;quot; vibe?&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;As an example of that last vibe shift, consider Australian comedy musician Tom Cardy. His early work was all &lt;a href=&quot;https://www.youtube.com/watch?v=GbmP2c6TGKc&quot;&gt;very self-deprecating&lt;/a&gt;, but in recent years it&#39;s shifted more towards a defiantly upbeat style.&lt;/p&gt;
&lt;p&gt;For example, this, a recent song &amp;amp; music video (animated by &lt;a href=&quot;https://www.youtube.com/@LankyMF&quot;&gt;LankyMF&lt;/a&gt;), &lt;strong&gt;&amp;quot;H.S.&amp;quot;&lt;/strong&gt;:&lt;/p&gt;
&lt;iframe width=&quot;640&quot; height=&quot;360&quot; src=&quot;https://www.youtube-nocookie.com/embed/kK0KPuH32mc?rel=0&quot; title=&quot;YouTube video player&quot; frameborder=&quot;0&quot; allow=&quot;accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share&quot; referrerpolicy=&quot;strict-origin-when-cross-origin&quot; allowfullscreen=&quot;&quot;&gt;&lt;/iframe&gt;
&lt;p&gt;This song &amp;amp; music video genuinely makes me feel &lt;em&gt;so&lt;/em&gt; emotional. I&#39;ve listened to it several times on loop, it&#39;s a great comfort pump-yourself-back-up song. And huge props to LankyMF for the delicious character designs, and the music video&#39;s story!&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;&lt;a href=&quot;https://tomcardy.bandcamp.com/album/the-dancefloor-at-the-end-of-the-universe&quot;&gt;You can get Tom Cardy&#39;s latest album here!&lt;/a&gt;&lt;/strong&gt; And watch the other music videos &lt;a href=&quot;https://www.youtube.com/channel/UCW7AGm8JSBEEew61dJIgl_A&quot;&gt;on his YouTube&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;&lt;a id=&quot;claw&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h2&gt;🐺 A 3D-printable claw&lt;/h2&gt;
&lt;p&gt;Being an adult means you can put minimal effort into a Halloween costume because it&#39;s &amp;quot;sexy&amp;quot;. Anyway here I am:&lt;/p&gt;
&lt;p&gt;&lt;video width=&quot;360&quot; height=&quot;640&quot; controls=&quot;&quot; style=&quot;display:block; margin:0 auto;&quot;&gt;&lt;source src=&quot;https://blog.ncase.me/content/stuff/2024-10/meow.mp4&quot; type=&quot;video/mp4&quot; /&gt;Your browser does not support the video tag.&lt;/video&gt;&lt;/p&gt;
&lt;p&gt;A friend helped me 3D print five of these claws (at different sizes to fit my fingers) and, wow:&lt;/p&gt;
&lt;ol&gt;
&lt;li&gt;They are surprisingly sharp, and can leave marks that last a &lt;em&gt;week&lt;/em&gt;.&lt;/li&gt;
&lt;li&gt;They have improved my quality of my... relationships.&lt;/li&gt;
&lt;/ol&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2024-10/good.jpg&quot; alt=&quot;Edited lawyer billboard that reads: &amp;quot;Injured? GOOD&amp;quot;&quot; /&gt;&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;&lt;a href=&quot;https://www.printables.com/model/478899-fursuit-finger-claw&quot;&gt;Check out the 3D printable file here, by Tioh!&lt;/a&gt;&lt;/strong&gt; (If you don&#39;t have a 3D printer nor know a friend who does, you &lt;em&gt;might&lt;/em&gt; find one at a local makerspace or library.)&lt;/p&gt;
&lt;p&gt;If anybody asks you why you&#39;re printing them, tell &#39;em it&#39;s for Halloween.&lt;/p&gt;
&lt;p&gt;If it&#39;s past Halloween by the time you&#39;re printing them, tell &#39;em it&#39;s for next Halloween.&lt;/p&gt;
&lt;p&gt;&lt;a id=&quot;shoes&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h2&gt;👟 Slip-on Shoes&lt;/h2&gt;
&lt;p&gt;I hate tying shoelaces.&lt;/p&gt;
&lt;p&gt;Let&#39;s crunch the numbers, shall we?&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;1 minute &lt;em&gt;at best&lt;/em&gt; to tie-on and take-off shoes each day&lt;/li&gt;
&lt;li&gt;That&#39;s ~365 minutes lost a year, or ~6 hours.&lt;/li&gt;
&lt;li&gt;If I value my time at $30/hour, that&#39;s $180 lost a year. You can definitely buy slip-on shoes for less than $180.&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;But wait, let&#39;s go further:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Let&#39;s conservatively estimate that ~50% of adults globally wear shoes with shoelaces. (The rest wear slip-on sandals, no shoes, etc.)&lt;/li&gt;
&lt;li&gt;There are ~6,000,000,000 adults worldwide. So, 50% of that, ~3,000,000,000 adults lose 6 hours every year. Or: 18,000,000,000 person-hours lost per year.&lt;/li&gt;
&lt;li&gt;Since there&#39;s 24 hours a day, 365 days a year, ~80 years per lifetime, that&#39;s... ~25,000 person-lifetimes a year, lost to tying shoelaces.&lt;/li&gt;
&lt;li&gt;For comparison: In the US, ~25,000 people are murdered annually. The 9/11 attacks killed ~3,000 people.&lt;/li&gt;
&lt;li&gt;Therefore: Shoelaces are equivalent to American murder, or eight 9/11&#39;s a year.&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;em&gt;Obviously tone-indicator sarcastic&lt;/em&gt;, but the point still stands:&lt;/p&gt;
&lt;p&gt;I HATE TYING SHOELACES&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Anyway, I got some slip-on shoes &lt;a href=&quot;https://www.amazon.com/dp/B08TBT3KBX&quot;&gt;from KEEZMZ on Amazon&lt;/a&gt;.&lt;/strong&gt; I originally wanted Velcro but I suggested this to a friend and he said &amp;quot;yeah if you wanna look like a kid?&amp;quot; so I guess I got adult-shamed into buying slip-on:&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/stuff/2024-10/shoes.jpg&quot; alt=&quot;Photo of my new, red slip-on shoes&quot; /&gt;&lt;/p&gt;
&lt;hr /&gt;
&lt;p&gt;That&#39;s it, that&#39;s all for this month.&lt;/p&gt;
</content>
        </entry>

    

        
        
        <entry>
            <title>Research Notes-dump for Oct 2024</title>
            <link href="https://blog.ncase.me/research-notes-oct-2024/"/>
            <updated>2024-10-03T00:00:00Z</updated>
            <id>https://blog.ncase.me/research-notes-oct-2024/</id>
            <content xml:lang="en" type="html">&lt;p&gt;&lt;em&gt;Hi! This is a &amp;quot;share your work&amp;quot; notes-dump, for some independent research projects I&#39;m doing in AI Alignment/Safety. Alas, this means this post will be more wordy, sloppy &amp;amp; technical than my usual &amp;quot;explain it like I&#39;m 12&amp;quot; style. Sorry! Should any of these bear fruit, I&#39;ll share the fruit&#39;s juice in more accessible packaging.&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;&lt;em&gt;(If any of these ideas inspire you to do some research of your own, feel free to cite this post as &lt;strong&gt;Nicky Case, 2024&lt;/strong&gt;! Here&#39;s a timestamped Archive.org &lt;a href=&quot;http://web.archive.org/web/20241002171613/https://blog.ncase.me/research-notes-oct-2024/&quot;&gt;save of this page, on October 2nd, 2024&lt;/a&gt;)&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Table of Contents:&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;In the Minimum Viable Prototype stage:&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;a href=&quot;https://blog.ncase.me/research-notes-oct-2024/#project_1&quot;&gt;Project 1&lt;/a&gt;) Want to know a human&#39;s values? Then &lt;strong&gt;BEG: Bayesian Elicitation Generation&lt;/strong&gt;. Combining LLMs with Good Ol&#39; Fashioned AI to actively elicit a human&#39;s preferences, in a qualitative &lt;em&gt;and&lt;/em&gt; quantitative way.&lt;/li&gt;
&lt;li&gt;&lt;a href=&quot;https://blog.ncase.me/research-notes-oct-2024/#project_2&quot;&gt;Project 2&lt;/a&gt;) &lt;strong&gt;SASS: Semi-automated AI Safety Scientist.&lt;/strong&gt; Proof-of-concept, of an LLM &lt;em&gt;automatically generating &amp;amp; testing&lt;/em&gt; hypotheses, to interpret another LLM. (Also, replicating-extending a MATS alum&#39;s research.)&lt;/li&gt;
&lt;li&gt;&lt;a href=&quot;https://blog.ncase.me/research-notes-oct-2024/#project_3&quot;&gt;Project 3&lt;/a&gt;) &lt;strong&gt;Speakeasy&lt;/strong&gt;: a tool for laypeople to make their own narrow, human-in-the-loop, scaffolded LLMs.&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;strong&gt;In the early-prototyping stage:&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;a href=&quot;https://blog.ncase.me/research-notes-oct-2024/#project_4&quot;&gt;Project 4&lt;/a&gt;) Beating Goodhart&#39;s Law with &lt;strong&gt;CUES: Capped, Uncorrelated Ensemble of Specifications&lt;/strong&gt;.&lt;/li&gt;
&lt;li&gt;&lt;a href=&quot;https://blog.ncase.me/research-notes-oct-2024/#project_5&quot;&gt;Project 5&lt;/a&gt;) &lt;strong&gt;The game theory of self-modification.&lt;/strong&gt; Understanding wireheading, human value change, AI stability under self-modification, etc.&lt;/li&gt;
&lt;li&gt;&lt;a href=&quot;https://blog.ncase.me/research-notes-oct-2024/#project_6&quot;&gt;Project 6&lt;/a&gt;) &lt;strong&gt;SCI: Semi-automatic Causal Inference&lt;/strong&gt;, with LLMs, Pearl causal diagrams, and Good Ol&#39; Fashioned Statistics.&lt;/li&gt;
&lt;/ul&gt;
&lt;hr /&gt;
&lt;p&gt;&lt;a id=&quot;project_1&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h2&gt;🙏 Project 1) BEG: Bayesian Elicitation Generation&lt;/h2&gt;
&lt;p&gt;&lt;strong&gt;Summary:&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;Let&#39;s use a scaffolded LLM to &lt;em&gt;qualitatively and quantitatively&lt;/em&gt; elicit a human&#39;s values! The steps: 1) ask the human open-ended questions, 2) extract &lt;em&gt;qualitative&lt;/em&gt; features they care about, 3) for each feature, generate &lt;em&gt;quantitative&lt;/em&gt; priors &amp;amp; likelihoods &amp;amp; posteriors &amp;amp; info-entropy, 4) use those to generate the &lt;em&gt;next&lt;/em&gt; round of questions, 5) repeat!&lt;/p&gt;
&lt;p&gt;. . .&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Related to:&lt;/strong&gt; Active preference elicitation, Reward uncertainty, Interpretability-by-design, Bayesian inference&lt;/p&gt;
&lt;p&gt;. . .&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Video of it in action!&lt;/strong&gt; (~2 min)&lt;/p&gt;
&lt;iframe width=&quot;640&quot; height=&quot;557&quot; src=&quot;https://www.youtube-nocookie.com/embed/oDzyPqGsy8M?si=HC3nAJprRVpn2a_w&amp;rel=0&quot; title=&quot;YouTube video player&quot; frameborder=&quot;0&quot; allow=&quot;accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share&quot; referrerpolicy=&quot;strict-origin-when-cross-origin&quot; allowfullscreen=&quot;&quot;&gt;&lt;/iframe&gt;
&lt;p&gt;. . .&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Introduction / Motivation:&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;Reward a robot for picking up dirt, and it&#39;ll pick up &amp;amp; drop the same dirt over and over again.&lt;sup class=&quot;footnote-ref&quot;&gt;&lt;a href=&quot;https://blog.ncase.me/research-notes-oct-2024/#fn1&quot; id=&quot;fnref1&quot;&gt;[1]&lt;/a&gt;&lt;/sup&gt; Point is: in AI, it&#39;s really hard to specify what we truly want. Hence: why not get an AI to &lt;em&gt;learn&lt;/em&gt; what we truly want? Maybe by &lt;em&gt;asking good questions?&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;There&#39;s many, many approaches to try to figure out a human&#39;s &amp;quot;reward function&amp;quot; -- (keywords: preference elicitation, inverse reinforcement learning, RLHF, etc) -- each with their pros/cons. For example:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;u&gt;Hand-writing the features to pay attention to&lt;/u&gt;: Simple &amp;amp; interpretable, but not flexible.&lt;/li&gt;
&lt;li&gt;&lt;u&gt;A full end-to-end neural network&lt;/u&gt;: Very flexible, but hard to interpret, &amp;amp; possibly overfit, doesn&#39;t generalize, and susceptible to adversarial examples.&lt;/li&gt;
&lt;li&gt;&lt;u&gt;Using an LLM to generate active-listening questions&lt;/u&gt;: (&lt;a href=&quot;https://arxiv.org/pdf/2310.11589&quot;&gt;Li, Tamkin, Goodman &amp;amp; Andreas 2023&lt;/a&gt; is a good example, and what inspired my project below!) Interpretable &lt;em&gt;and&lt;/em&gt; flexible, but &amp;quot;only&amp;quot; results in &lt;em&gt;qualitative&lt;/em&gt; results. &lt;em&gt;Quantitative&lt;/em&gt; estimates of how much a human values something would let us make good trade-offs. Even better would be quantitative estimates of the &lt;em&gt;uncertainty of AI&#39;s own estimates&lt;/em&gt;, so that the AI can err on the side of caution, &amp;amp; asking when uncertain.&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;So: I&#39;d like to propose a method that&#39;s interpretable, flexible, &lt;em&gt;and&lt;/em&gt; gives us both qualitative &lt;em&gt;and&lt;/em&gt; quantitative estimates of what folks value, &lt;em&gt;and&lt;/em&gt; keeps track of its own uncertainty! Let&#39;s call it, &lt;strong&gt;BEG: Bayesian Elicitation Generation&lt;/strong&gt;, as in, you can BEG the human for th--&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;How it works:&lt;/strong&gt;&lt;/p&gt;
&lt;ol&gt;
&lt;li&gt;Use an LLM to ask a concrete, open-ended question. (for lots of &#39;bits&#39; of information)&lt;/li&gt;
&lt;li&gt;Get a free-form reply.&lt;/li&gt;
&lt;li&gt;Use an LLM to extract what things the human cares about.&lt;/li&gt;
&lt;li&gt;For each thing: use an LLM to guess-timate the &amp;quot;human prior&amp;quot;: on a 7-point scale, how much do people desire vs anti-desire this thing?  &lt;code&gt;P(value)&lt;/code&gt;&lt;/li&gt;
&lt;li&gt;Use an LLM to generate a &amp;quot;likelihood ratio&amp;quot;: if the user valued the thing [X] amount, how likely is it they would have written what they did?  &lt;code&gt;P(text | value)&lt;/code&gt;&lt;/li&gt;
&lt;li&gt;Multiply the prior &amp;amp; likelihood, then normalize, to get the posterior: this is our &lt;em&gt;current&lt;/em&gt; belief (with explicit uncertainty!) of how much the user values the thing!  &lt;code&gt;P(value | text) = P(text | value) * P(value) / [normalization constant]&lt;/code&gt;&lt;/li&gt;
&lt;li&gt;We can now use this posterior to generate actions that help the human... or generate &lt;em&gt;new&lt;/em&gt; questions to ask!  We want to ask the questions that will help us reduce uncertainty the most.  So, we can measure each feature&#39;s &amp;quot;information entropy&amp;quot;, select the top one(s), and prompt an LLM to ask a concrete, open-ended question about them.&lt;/li&gt;
&lt;li&gt;Repeat!&lt;/li&gt;
&lt;/ol&gt;
&lt;p&gt;. . .&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;More possibilities / Challenges / Etc:&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;u&gt;To do, actual empirical user-testing&lt;/u&gt; (with MTurk or similar?) to make sure this process is actually worth it. (&lt;a href=&quot;https://arxiv.org/pdf/2310.11589&quot;&gt;Li, Tamkin, Goodman &amp;amp; Andreas 2023&lt;/a&gt; finds that open questions aren&#39;t too annoying to users, and are actually &lt;em&gt;preferred&lt;/em&gt; over binary choices.)&lt;/li&gt;
&lt;li&gt;&lt;u&gt;Pragmatic Feature Preferences&lt;/u&gt;: (Directly inspired by &lt;a href=&quot;https://arxiv.org/pdf/2405.14769&quot;&gt;Peng, Sun, Shu &amp;amp; Abel 2024&lt;/a&gt;) If there&#39;s an obvious thing a user &lt;em&gt;didn&#39;t&lt;/em&gt; mention in their free-form reply, that&#39;s evidence too! It&#39;s evidence they feel neutral about it. (e.g. Bot asks Human about their life aspirations, Human doesn&#39;t mention getting rich, Bot can infer Human doesn&#39;t care much for getting rich.)&lt;/li&gt;
&lt;li&gt;&lt;u&gt;Non-verbal communication&lt;/u&gt;: The pieces of &amp;quot;evidence&amp;quot;, of what a Human likes/dislikes, don&#39;t have to be verbal. We could use facial expressions or tone-of-voice, and it&#39;d be easy to incorporate into this system: &amp;quot;just&amp;quot; multiply in their likelihood ratios.&lt;/li&gt;
&lt;li&gt;&lt;u&gt;Non-linearities&lt;/u&gt;: If a Human would be happy to own a house, that doesn&#39;t mean they&#39;d be 10 times happier to own 10 houses! Challenge: make this system accommodate non-linearities &amp;amp; interaction effects.&lt;/li&gt;
&lt;li&gt;&lt;u&gt;Causal Graphs&lt;/u&gt;: We want this system to distinguish between &lt;em&gt;intrinsic&lt;/em&gt; goals (&amp;quot;for its own sake&amp;quot;) and &lt;em&gt;instrumental&lt;/em&gt; goals (&amp;quot;for something else&#39;s sake&amp;quot;). Challenge: make this system model goals as a network, and have it ask counterfactual questions to distinguish between intrinsic/instrumental goals. (e.g. &amp;quot;If money were no object, would you &lt;em&gt;still&lt;/em&gt; want to work this job?&amp;quot;)
&lt;ul&gt;
&lt;li&gt;See also &lt;a href=&quot;https://blog.ncase.me/research-notes-oct-2024/#project_6&quot;&gt;Project 6&lt;/a&gt;: Semi-automated Causal Inference&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;&lt;u&gt;Beyond a 7-point scale?&lt;/u&gt;: When a Human says they &amp;quot;really love&amp;quot; their partner, that should matter more than when they say they &amp;quot;really love&amp;quot; spicy food. Challenge: re-make this system to accommodate a &amp;quot;universal&amp;quot; utility scale? Also, would it be worth trying to model the belief-probability-distribution as &lt;em&gt;fully continuous&lt;/em&gt;, not discrete?&lt;/li&gt;
&lt;li&gt;&lt;u&gt;Not double-counting evidence&lt;/u&gt;: To avoid &amp;quot;double-counting&amp;quot; evidence, this system should keep track of individual pieces of evidence that Human likes/dislikes something, then use an LLM to check if a piece of evidence is already accounted for. If so, it &lt;em&gt;won&#39;t&lt;/em&gt; multiply the prior by its likelihood ratio again.&lt;/li&gt;
&lt;li&gt;&lt;u&gt;Having evidence &amp;quot;fade away&amp;quot;&lt;/u&gt;: People&#39;s values can change! How can we model this mathematically? Well: if double-counting evidence means multiplying by the likelihood &lt;em&gt;L&lt;/em&gt; twice, that is, multiplying by &lt;em&gt;L&lt;sup&gt;2&lt;/sup&gt;&lt;/em&gt;, then we can model &lt;em&gt;fading evidence&lt;/em&gt; by multiplying a prior by &lt;em&gt;L&lt;sup&gt;weight&lt;/sup&gt;&lt;/em&gt;, where &lt;em&gt;weight = exp( – some constant * time )&lt;/em&gt;.
&lt;ul&gt;
&lt;li&gt;Sanity-checking this: when time = 0, i.e. we &lt;em&gt;just&lt;/em&gt; got this evidence, then weight = 1. As time → ∞, then weight = 0, i.e. we completely discount the evidence, and revert to the &amp;quot;human prior&amp;quot;.&lt;/li&gt;
&lt;li&gt;Bonus: every time a Human brings up the same evidence again, we reset the time to 0.&lt;/li&gt;
&lt;li&gt;Bonus x 2: we wouldn&#39;t even need to hard-code &amp;quot;ask Human if thing they mentioned long ago is still true&amp;quot;! This &lt;em&gt;automatically&lt;/em&gt; comes out of: evidence fades away → belief fades back to being an uncertain human prior → Bot asks questions about high-uncertainty/entropy features → Bot will &lt;em&gt;automatically&lt;/em&gt; re-up questions about old features!&lt;/li&gt;
&lt;li&gt;&lt;em&gt;How quickly&lt;/em&gt; do we predict preferences/values to change? I dunno, fit that to human data, and/or ask an LLM for an educated guess. (For example: I&#39;d expect tastes in fashion to change fast, but sexual orientation to basically never change, or if so, &lt;em&gt;very&lt;/em&gt; rarely and slowly.)&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;strong&gt;Post-Script: Prior Art&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;Ugh I spent over a week making the above prototype &amp;amp; outlining the research idea, now I just found &lt;a href=&quot;https://arxiv.org/pdf/2405.00981&quot;&gt;Austin, Korikov, Toroghi &amp;amp; Sanner 2024&lt;/a&gt; posted on ArXiV less than 2 months ago. It &lt;em&gt;is&lt;/em&gt; a good paper! (Crucially, they show LLM+Bayes &amp;gt; raw LLM!) But ok yeah they basically had &amp;amp; executed the same idea, combining LLMs with Bayesian inference. A few (maybe still useful?) differences between my idea and theirs:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;I use an LLM to &lt;em&gt;generate&lt;/em&gt; the &amp;quot;human prior&amp;quot; for new features, rather than starting with flat uniform prior.&lt;/li&gt;
&lt;li&gt;Their query-generation makes yes/no questions, rather than open-ended questions that would allow the AI to capture &lt;em&gt;new&lt;/em&gt; features the designer never thought of.&lt;/li&gt;
&lt;li&gt;Their system has probabilities directly over &lt;em&gt;what thing to recommend&lt;/em&gt; (e.g. a movie), rather than trying to infer the human&#39;s more fundamental preferences/utility function.&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;(Though, these are probably all small tweaks/changes. I&#39;ll think more later about how I can make my research idea be more of a value-add above this paper. Maybe one of the &amp;quot;More possibilities&amp;quot; ideas listed above.)&lt;/p&gt;
&lt;hr /&gt;
&lt;p&gt;&lt;a id=&quot;project_2&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h2&gt;🔬 Project 2) SASS: Semi-automated AI Safety Scientist&lt;/h2&gt;
&lt;p&gt;&lt;strong&gt;Summary:&lt;/strong&gt; This is a proof-of-concept of using a scaffolded LLM to &lt;em&gt;amplify&lt;/em&gt; human ability to do AI Safety research. With this tool, I replicated &lt;a href=&quot;https://www.alignmentforum.org/posts/dLg7CyeTE4pqbbcnp/language-models-model-us&quot;&gt;a study by Egg Syntax (MATS alumnus)&lt;/a&gt;, which showed that GPT-3.5 can detect your gender with ~80% accuracy from &lt;em&gt;just a few paragraphs of your writing.&lt;/em&gt; But why? To find out, I made a tool to help &lt;em&gt;automatically generate and test&lt;/em&gt; hypotheses, to find &lt;em&gt;human-interpretable&lt;/em&gt; features of one&#39;s writing style, that GPT-3.5 uses to detect one&#39;s gender! (Spoiler: it&#39;s social-emotional focus, and possibly &amp;quot;homophily&amp;quot;.)&lt;/p&gt;
&lt;p&gt;(&lt;a href=&quot;https://blog.ncase.me/research-notes-oct-2024/#project_6&quot;&gt;Project 6&lt;/a&gt; also shows another case of semi-automated scientific research, but for causal inference.)&lt;/p&gt;
&lt;p&gt;. . .&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Related to:&lt;/strong&gt; Scalable oversight, Human-amplification / Cyborgism, Black-box / concept-based interpretability, Algorithmic bias, Empirical LLM work.&lt;/p&gt;
&lt;p&gt;. . .&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;wake up babe new gender dysphoria just dropped 😭&lt;/p&gt;
&lt;p&gt;~ &lt;em&gt;me, personal communication to Egg Syntax, in response to their finding that GPT-3.5 can infer your gender from writing&lt;/em&gt;&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;&lt;strong&gt;Introduction / Motivation:&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;So, some folks tried making &lt;a href=&quot;https://sakana.ai/ai-scientist/&quot;&gt;an AI Scientist&lt;/a&gt; recently.&lt;/p&gt;
&lt;p&gt;&lt;a href=&quot;https://web.archive.org/web/20241001051745/https://www.astralcodexten.com/p/sakana-strawberry-and-scary-ai&quot;&gt;It&#39;s not good (yet)&lt;/a&gt;. For now, research still requires a human.&lt;/p&gt;
&lt;p&gt;But oh god, AI &lt;em&gt;Safety&lt;/em&gt; research really needs to catch up with AI &lt;em&gt;Capabilities&lt;/em&gt; research. One way to do this is by &lt;em&gt;amplifying&lt;/em&gt; humans to be able to do AI Safety science faster &amp;amp; better. (see: &lt;a href=&quot;https://www.lesswrong.com/posts/bxt7uCiHam4QXrQAA/cyborgism&quot;&gt;Cyborgism&lt;/a&gt;) This project is a proof-of-concept of that.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Step 1) Replication&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;First, I replicated &lt;a href=&quot;https://www.alignmentforum.org/posts/dLg7CyeTE4pqbbcnp/language-models-model-us&quot;&gt;MATS alum Egg Syntax&#39;s work&lt;/a&gt;, which showed that GPT-3.5 can detect the gender of an author from &lt;em&gt;just their writing style&lt;/em&gt;, with ~80% accuracy. (Egg&#39;s full paper showed GPT can &lt;em&gt;also&lt;/em&gt; detect ethnicity &amp;amp; educational level! That&#39;s a huge concern for privacy, and subtly-hidden AI bias!)&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Replication details:&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/media/oct-2024/research/egg_replication.png&quot; alt=&quot;&quot; /&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Actually, &lt;strong&gt;I was only able to get GPT-3.5 to detect gender from writing at accuracy of ~64%. (p&amp;lt;0.00001)&lt;/strong&gt; After much communication with Egg Syntax, it turns out the major issue is their research was using OpenAI&#39;s old Text Completions API, which is now gone &amp;amp; inaccessible. I&#39;m using their Chat Completions API. Also, OpenAI is still &amp;quot;continuously updating&amp;quot; GPT-3.5-Turbo behind the scenes; &lt;em&gt;any&lt;/em&gt; LLM work that uses 3rd-party APIs will be hard to exactly replicate, because of this. (Maybe my future work should &lt;em&gt;also&lt;/em&gt; be tested on open-source LLMs)&lt;/li&gt;
&lt;li&gt;Also note that the accuracy was mostly due to detecting male writing very well, while being slightly-below-chance on female writing.&lt;/li&gt;
&lt;li&gt;(Note on gender stuff: Normally I&#39;d use &amp;quot;Woman&amp;quot; &amp;amp; &amp;quot;Man&amp;quot; rather than sounding like an alien saying &amp;quot;female&amp;quot; and &amp;quot;male&amp;quot;, but the subjects in these studies are minors. But they&#39;re also not &lt;em&gt;kid&lt;/em&gt; kids, so &amp;quot;girls&amp;quot; and &amp;quot;boys&amp;quot; aren&#39;t right either. Finally, the dataset did not code gender beyond Male/Female (&lt;a href=&quot;https://www.pewresearch.org/short-reads/2022/06/07/about-5-of-young-adults-in-the-u-s-say-their-gender-is-different-from-their-sex-assigned-at-birth/&quot;&gt;while ~5% of folks under &amp;lt;30 in the U.S. are trans/non-binary&lt;/a&gt;), nor do I know if the dataset-creators coded based on natal sex, or psychological gender.)&lt;/li&gt;
&lt;li&gt;Egg&#39;s primary analysis was on OKCupid profiles. Because the OKCupid profiles were possibly in GPT&#39;s training data -- (and because on a dating site, people are directly talking about themselves &lt;em&gt;and&lt;/em&gt; they have an incentive to overplay their femininity/masculinity) -- this dataset may be &amp;quot;too easy&amp;quot;. So, Egg cross-validated their finding on the &lt;strong&gt;Persuade 2.0&lt;/strong&gt; dataset (&lt;a href=&quot;https://www.sciencedirect.com/science/article/pii/S1075293524000588?ssrnid=4795747&amp;amp;dgcid=SSRN_redirect_SD&quot;&gt;Crossley, Tian, Baffour, Franklin, Benner &amp;amp; Boser 2024&lt;/a&gt;) which came out in 2024, long after GPT-3.5-Turbo&#39;s training cutoff (&lt;a href=&quot;https://help.openai.com/en/articles/8555514-gpt-3-5-turbo-updates#h_d189a2fafd&quot;&gt;Sep 2021&lt;/a&gt;). &lt;strong&gt;This is the dataset I&#39;m replicating the study with!&lt;/strong&gt;  Persuade 2.0 is a bunch of high school student essays, on topics the students did &lt;em&gt;not&lt;/em&gt; choose. So, the fact that GPT-3.5 can still detect gender at 64% accuracy &lt;em&gt;from impersonal essays on non-self-chosen-topics&lt;/em&gt; still surprises me!&lt;/li&gt;
&lt;li&gt;Egg used &lt;em&gt;just&lt;/em&gt; the 12th-grader&#39;s essays, since they suspected that GPT, being trained on mostly adult writing, would be most accurate on that. My replication mostly used the same data -- I noticed that almost all the 12th-grader essays were on distance learning, but 6 were on Summer Projects. For consistency, I removed those 6 essays, leaving &lt;strong&gt;394 twelth-grader essays on Distance Learning.&lt;/strong&gt; (4 more essays were taken out of testing, to use as the few-shot examples in my prompt)&lt;/li&gt;
&lt;li&gt;Temperature set to 0.&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;strong&gt;Step 2) Semi-automated hypothesis generation &amp;amp; testing&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;I coded my scaffolded LLM to:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;Generate hypotheses:&lt;/strong&gt; I just asked GPT to give me 5 hypotheses, in the form &amp;quot;males tend to write more [blank], females tend to write more [blank]&amp;quot;
&lt;ul&gt;
&lt;li&gt;(At first I tried getting GPT to read 10 random female essays, then 10 random male essays, and compare &amp;amp; contrast them -- &lt;em&gt;while blinding GPT to the fact they&#39;re female/male essays or that I&#39;m testing for gender at all&lt;/em&gt; -- but the generated hypotheses were all crap. Too focused on the bad student writing &amp;amp; typos &amp;amp; lack of rigor. Alas.)&lt;/li&gt;
&lt;li&gt;Anyway, I took GPT&#39;s above hypotheses, then fed them all into the next step:&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Test the generated hypotheses:&lt;/strong&gt; I took the subset of essays that GPT guessed gender correctly. Then, for each hypothesis in the form &amp;quot;males write more [X], females write more [Y]&amp;quot;, I pinged GPT-4o-mini (a more advanced model) to re-write the female essays &amp;quot;in the male style&amp;quot;, and vice versa (again, while blinding it to the true reason for the rewrite). I also instructed GPT-4o-mini to keep the rest of the essay as identical as possible, even typos and grammar errors. Finally, I measured the drop in GPT-3.5-Turbo&#39;s accuracy. Reminder: the baseline is 100%, chance is 50%, 0% is the feature &lt;em&gt;fully gender-flipped&lt;/em&gt; the essay.
&lt;ul&gt;
&lt;li&gt;(At first, I considered asking GPT to code the essays on hypothesized features of the essay styles, like personal-ness/expressive-ness/analytic-ness/etc, then do a basic multilinear regression. However, this would mix up correlation &amp;amp; causation. Asking GPT to &lt;em&gt;rewrite the essay but gender-flip one feature&lt;/em&gt; was the only way to be &lt;em&gt;sure&lt;/em&gt; about causality.)&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;strong&gt;Overall results:&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;Surprisingly, it was really easy to gender-flip male essays to female, but &lt;em&gt;NOT&lt;/em&gt; the female essays to male? Female essay&#39;s gender was weirdly robust, given that originally, GPT did slightly &lt;em&gt;worse&lt;/em&gt; than chance at detecting female essays.&lt;/p&gt;
&lt;p&gt;Here&#39;s the first hypothesized feature I tested:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;&amp;quot;females write more emotionally,&lt;br /&gt;
males write more neutrally&amp;quot;&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;When re-written with that feature gender-flipped...&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/media/oct-2024/research/rewrite_emo_neutral.png&quot; alt=&quot;&quot; /&gt;&lt;/p&gt;
&lt;p&gt;Male accuracy got &lt;em&gt;obliterated&lt;/em&gt;, while... female accuracy remained untouched??? (Reminder: these were on the &lt;em&gt;subset&lt;/em&gt; of essays that GPT-3.5 got right the first time. That&#39;s why, above, total accuracy weighs males more than females: because GPT-3.5 accurately detected males much better than females, in the original setup.)&lt;/p&gt;
&lt;p&gt;Other tested features that almost eliminated male accuracy, but &lt;strong&gt;female accuracy always stayed high:&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Female: with personal anecdotes, Male: impersonal&lt;/li&gt;
&lt;li&gt;Female: cooperative, Male: competitive&lt;/li&gt;
&lt;li&gt;Female: complex/nuanced, Male: simple/direct&lt;/li&gt;
&lt;li&gt;Female: focus on social-emotional aspects, Male: focus on objective-material aspects&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;(As a &amp;quot;placebo test&amp;quot;, I also just re-tested all the essays with no changes. Accuracy remained near 100% for both &lt;em&gt;but not exactly 100%!&lt;/em&gt; It&#39;s been long-known that GPT is non-deterministic &lt;em&gt;even with temperature = 0&lt;/em&gt;. (&lt;a href=&quot;https://152334h.github.io/blog/non-determinism-in-gpt-4/&quot;&gt;possibly due to floating-point GPU errors or its Mixture of Experts model&lt;/a&gt;).)&lt;/p&gt;
&lt;p&gt;Looking much much more closely at the female/male essays (let me tell you, reading dozens of 12th-graders&#39; strong opinions is &lt;em&gt;not&lt;/em&gt; fun), I thought... wait, hang on, let me try &lt;em&gt;this&lt;/em&gt; hypothesis:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Females write about their female friends/family/mentors&lt;br /&gt;
Males write about their male friends/family/mentors&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;&lt;em&gt;This&lt;/em&gt; one worked.&lt;/p&gt;
&lt;p&gt;(GPT could not generate this hypothesis, not without me basically spoon-feeding the answer in my prompt.)&lt;/p&gt;
&lt;p&gt;For example: an essay talked about someone&#39;s immigrant friend from Guatemala. In the original, the friend was a girl, and GPT detected the essay as &amp;quot;female&amp;quot;, even when it was rewritten to not be a friend (impersonal) or focus on friend&#39;s material struggles (instead of emotional). However, simply rewriting the essay so that the friend was a boy instead of girl, &lt;em&gt;that&lt;/em&gt; got GPT to detect the essay as &amp;quot;male&amp;quot;. (More examples: flipping from talking about &amp;quot;my mom&amp;quot; to &amp;quot;my dad&amp;quot;, or &amp;quot;my brother&amp;quot; to &amp;quot;my sister&amp;quot;.)&lt;/p&gt;
&lt;p&gt;Here&#39;s how gender-flipping friends &amp;amp; family impacted accuracy:&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/media/oct-2024/research/rewrite_homophily.png&quot; alt=&quot;&quot; /&gt;&lt;/p&gt;
&lt;p&gt;&lt;em&gt;Now&lt;/em&gt; female accuracy was no better than chance, and male probably wasn&#39;t better than chance.&lt;/p&gt;
&lt;p&gt;(&lt;strong&gt;BUT THIS IS NOW EVEN WEIRDER:&lt;/strong&gt; In previous tests, male essays were &lt;em&gt;very&lt;/em&gt; sensitive to changes, their accuracy brought down to ~0%. But with &lt;em&gt;this&lt;/em&gt; gender-flip, which could bring down the robust female essays... it only creates a smaller dip in accuracy? I don&#39;t get it.)&lt;/p&gt;
&lt;p&gt;It&#39;s been long known that people are &amp;quot;homophilic&amp;quot; (homo = same, phile = attracted to), i.e. people disproportionally have friends of the same gender / ethnicity / age / class / etc. (&lt;a href=&quot;https://www.jstor.org/stable/2112441&quot;&gt;Shrum, Cheek &amp;amp; Hunter 1988&lt;/a&gt;)&lt;/p&gt;
&lt;p&gt;Likewise, (&lt;em&gt;on average&lt;/em&gt;) moms tend to prefer daughters &amp;amp; fathers tend to prefer sons. (&lt;a href=&quot;https://www.nature.com/articles/s41598-018-33650-1.pdf&quot;&gt;Lynch, Wasielewski &amp;amp; Cronk 2018&lt;/a&gt;, which wasn&#39;t even seeking to test this hypothesis, they found it as a side-effect of testing a different evo-psych hypothesis (which failed the test).) And if we reasonably assume kids &lt;em&gt;on average&lt;/em&gt; like the parent that likes them more, then sons might write more about their dads, and daughters about their moms.&lt;/p&gt;
&lt;p&gt;Could this be why GPT &amp;quot;thought&amp;quot; that students who talk about their female friends/family are more likely to be female, and likewise for male?&lt;/p&gt;
&lt;p&gt;On first glance, it &lt;em&gt;seems&lt;/em&gt; like GPT assumes homophily, &amp;quot;alike likes alike&amp;quot;, in terms of gender:&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/media/oct-2024/research/gender_vector_0.png&quot; alt=&quot;&quot; /&gt;&lt;/p&gt;
&lt;p&gt;An &lt;em&gt;alternate&lt;/em&gt; hypothesis is that GPT is just dumb as nails, and simply putting &amp;quot;she/her&amp;quot; or &amp;quot;sister&amp;quot;, etc, tilts some kind of internal gender-vector to output &amp;quot;female&amp;quot; at the end no matter what.&lt;/p&gt;
&lt;p&gt;A quick test shows this may be the case:&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/media/oct-2024/research/gender_vector.png&quot; alt=&quot;&quot; /&gt;&lt;/p&gt;
&lt;p&gt;However: 1) Chain of Thought (&amp;quot;let&#39;s think step by step&amp;quot;) solves the above problem. Also, 2) Sometimes GPT &lt;em&gt;isn&#39;t&lt;/em&gt; being that dumb, and even without Chain of Thought, it &lt;em&gt;can&lt;/em&gt; output the opposite response: (Note that the below shows that GPT&#39;s biased towards assuming heterosexuality, but I &lt;em&gt;did&lt;/em&gt; ask &amp;quot;which is more likely&amp;quot;, and 90% of people are straight)&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/media/oct-2024/research/gender_vector_2.png&quot; alt=&quot;&quot; /&gt;&lt;/p&gt;
&lt;p&gt;(Other tests in this vein also showed GPT is biased to associate parental affection with moms over dads, romance with women over men, friendship with women over men. I don&#39;t like these stereotypes, but GPT learnt from &lt;em&gt;our&lt;/em&gt; internet text, &lt;em&gt;our&lt;/em&gt; biases. GPT&#39;s a cultural mirror, and I don&#39;t like what I see.)&lt;/p&gt;
&lt;p&gt;Sure, GPT &amp;quot;knows&amp;quot; about gender-homophily, in that if you directly ask it about it, it&#39;ll tell you the science. But is it &amp;quot;using that fact&amp;quot; to make predictions about an author&#39;s gender? (What&#39;s it even mean for a Transformer to &amp;quot;use&amp;quot; a &amp;quot;fact&amp;quot;?) Without access to GPT&#39;s internals, it&#39;s not possible to know. But whether it&#39;s dumb luck or not, it seemed to be the only thing to get GPT to gender-flip its predictions on female essays: sisters become brothers, female friends become male friends, etc.&lt;/p&gt;
&lt;p&gt;For completion &amp;amp; my own satisfaction, I got my scaffolded LLM to rewrite the essays to gender-flip friends/family &lt;em&gt;AND&lt;/em&gt; to gender-flip on a personal-emotional/impersonal-logical writing style... and accuracy on both was obliterated:&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/media/oct-2024/research/rewrite_wrecked.png&quot; alt=&quot;&quot; /&gt;&lt;/p&gt;
&lt;p&gt;&lt;em&gt;Finally&lt;/em&gt;.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Conclusion:&lt;/strong&gt; GPT is detecting author gender with ~65% accuracy, on impersonal essay writing, due to a mix of social-emotional focus and &lt;em&gt;possibly&lt;/em&gt; gender-homophily. (Also, semi-automated AI safety research may kinda sorta be helpful.)&lt;/p&gt;
&lt;p&gt;. . .&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;More possibilities / Challenges / Etc:&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Publish the code for the above, for ease of replication. (And maybe use an open-source LLM for improved reproducibility, since GPT/Claude/etc keeps getting changed behind-the-scenes.)&lt;/li&gt;
&lt;li&gt;Try this experiment again for GPT-4o, with Chain of Thought, with ethnicity &amp;amp; age, etc.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Figure out the mystery:&lt;/strong&gt; &lt;em&gt;why&lt;/em&gt; is it that male essays could be detected much better than female, &lt;em&gt;but also&lt;/em&gt; the correctly-detected female essays were more robust to rewrites than male essays, &lt;em&gt;but also&lt;/em&gt; the one feature (homophily) that causes female accuracy to drop, causes a smaller drop in the otherwise-sensitive male essays?&lt;/li&gt;
&lt;li&gt;As an aside on &amp;quot;automated causal hypothesis &amp;amp; testing&amp;quot;, &lt;strong&gt;this could be a solution to goal misgeneralization!&lt;/strong&gt; Misgeneralization is caused by a feature being &lt;em&gt;correlated&lt;/em&gt; with the true goal, but not directly causally connected. So, an AI that can &lt;em&gt;do its own causal hypothesis &amp;amp; testing&lt;/em&gt; — hm, maybe powered by LLMs, as shown above? — could beat goal misgeneralization!
&lt;ul&gt;
&lt;li&gt;In fact, this paper (&lt;a href=&quot;https://arxiv.org/pdf/2309.16166&quot;&gt;Armstrong, Maranhão, Daniels-Koch, Leask &amp;amp; Gorman 2023&lt;/a&gt;) claims to have done something similar to that on the famous CoinRun example, and it worked! (Unfortunately, the algorithm they used is not detailed in the paper, it is &amp;quot;proprietary&amp;quot;. Eh, whatever.)&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;The Persuade 2.0 dataset has a score for essay quality. How closely does GPT&#39;s rating matches human ratings, compared to human inter-rater reliability? What features does GPT think makes a &amp;quot;quality essay&amp;quot;?&lt;/li&gt;
&lt;li&gt;Given the nature of &amp;quot;automatically generate hypotheses &amp;amp; tests&amp;quot;, does &amp;quot;pre-registration&amp;quot; even make sense for Semi-Automated Science? I guess treat these studies as exploratory studies. Alternatively, this tool should also make it easy to run robustness-to-study-design checks, to prove that results aren&#39;t flukes / p-hacking.&lt;/li&gt;
&lt;li&gt;I&#39;d like to minimize this tool&#39;s dual-use case. So: what tools would help with Semi-automated Science for AI &lt;em&gt;Safety&lt;/em&gt;, not Capabilities? Maybe this tool only works with black-box LLMs and AIs, and can&#39;t be used for training or making new ANN architectures. (See next project for further discussion)&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;. . .&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Special Thanks to Egg Syntax!&lt;/strong&gt;  For their original study, helping me replicate it, bouncing ideas about Automated AI Safety Science, and telling me about the MATS program in the first place!&lt;/p&gt;
&lt;hr /&gt;
&lt;p&gt;&lt;a id=&quot;project_3&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h2&gt;💬 Project 3) Speakeasy: a tool for laypeople to scaffold LLMs&lt;/h2&gt;
&lt;p&gt;&lt;strong&gt;Summary:&lt;/strong&gt; A tool to make human-in-the-loop, narrow, hybrid AIs -- that mix the flexibility of LLMs, the interpretability of GOFAI (Good Ol&#39; Fashioned AI), and the agency of us humans.&lt;/p&gt;
&lt;p&gt;Motivations:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;A human-in-the-loop hybrid AI gives us better capabilities, but safer &amp;amp; more interpretable.&lt;/li&gt;
&lt;li&gt;As a result, scaffolding &lt;em&gt;may&lt;/em&gt; shift the economic incentives away from more advanced frontier AIs, to making AIs more narrow, modular, and easy to plug-and-play into scaffolding tools.&lt;/li&gt;
&lt;li&gt;I just think they&#39;re neat&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;(I used a prototype of Speakeasy for &lt;a href=&quot;https://blog.ncase.me/research-notes-oct-2024/#project_1&quot;&gt;Project 1&lt;/a&gt; &amp;amp; &lt;a href=&quot;https://blog.ncase.me/research-notes-oct-2024/#project_2&quot;&gt;Project 2&lt;/a&gt;! &amp;amp; I&#39;ll likely use it for &lt;a href=&quot;https://blog.ncase.me/research-notes-oct-2024/#project_6&quot;&gt;Project 6&lt;/a&gt; too.)&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Related to:&lt;/strong&gt; Human-amplification/Cyborgism, Human-in-the-loop AI, Narrow AI&lt;/p&gt;
&lt;p&gt;. . .&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Introduction:&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;Like how calculators &amp;amp; spreadsheets lets ~everyone accessibly use the power of computation... &lt;em&gt;Speakeasy&lt;/em&gt; lets ~everyone accessibly use the power of &lt;em&gt;scaffolded&lt;/em&gt; LLMs, for their own personal, narrow-AI use cases.&lt;/p&gt;
&lt;p&gt;Concretely: you use a simple interface to make a chatbot. But unlike &amp;quot;GPTs&amp;quot; or Character.ai, it&#39;s not just an LLM + a system prompt + some examples + a RAG!  You can make a &lt;em&gt;full state machine&lt;/em&gt;, with memory &amp;amp; logic, plugging into tools like statistics &amp;amp; visualization!  (and &lt;em&gt;maaaaaybe&lt;/em&gt; web search/scraping.)&lt;/p&gt;
&lt;p&gt;I&#39;ll just repeat the video I showed for &lt;a href=&quot;https://blog.ncase.me/research-notes-oct-2024/#project_1&quot;&gt;Project 1&lt;/a&gt;, to show what kind of hybrid-LLM/GOFAI chatbot you can make in Speakeasy: (~2 min)&lt;/p&gt;
&lt;iframe width=&quot;640&quot; height=&quot;557&quot; src=&quot;https://www.youtube-nocookie.com/embed/oDzyPqGsy8M?si=HC3nAJprRVpn2a_w&amp;rel=0&quot; title=&quot;YouTube video player&quot; frameborder=&quot;0&quot; allow=&quot;accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share&quot; referrerpolicy=&quot;strict-origin-when-cross-origin&quot; allowfullscreen=&quot;&quot;&gt;&lt;/iframe&gt;
&lt;p&gt;( &lt;em&gt;Ideally&lt;/em&gt; I&#39;d like this tool to be a pure visual interface, like &lt;a href=&quot;https://ifttt.com/&quot;&gt;IFTTT&lt;/a&gt; or Excel or Scratch.  Right now, Speakeasy uses a simplified script, and it runs in JavaScript, right in your browser.  More convenient for a layperson than downloading &amp;amp; running Python, that&#39;s for sure. )&lt;/p&gt;
&lt;p&gt;( &lt;em&gt;Also&lt;/em&gt; ideally I&#39;d like this tool to be like CodePen or Google Docs: people can share &amp;amp; remix each other&#39;s scaffolded LLMs. )&lt;/p&gt;
&lt;p&gt;( Also note: I&#39;m currently making this project in collaboration with educational non-profit &lt;a href=&quot;https://hackclub.com/&quot;&gt;Hack Club&lt;/a&gt;, which works with high-schoolers.  So by &amp;quot;layperson-friendly&amp;quot;, I mean the specific target is &amp;quot;so accessible, a high-schooler with no prior coding experience can make a thing they&#39;re proud of in &amp;lt;30 minutes!&amp;quot; )&lt;/p&gt;
&lt;p&gt;. . .&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Motivations:&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;First, I gotta be aware, yeah this smells like enhancing AI Capabilities.&lt;/p&gt;
&lt;p&gt;It&#39;s definitely a &lt;em&gt;bit&lt;/em&gt; dual-use, but: some reasons (rationalizations?) why this is much more net-positive for AI Safety:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;u&gt;Speakeasy combines the best of GOFAI (Good Ol&#39; Fashioned AI) and ANNs and human agency&lt;/u&gt;:
&lt;ul&gt;
&lt;li&gt;GOFAI is interpretable &amp;amp; verifiable, but inflexible.&lt;/li&gt;
&lt;li&gt;ANNs are very flexible, but currently uninterpretable. (&amp;amp; maybe prone to goal mis-generalization)&lt;/li&gt;
&lt;li&gt;Speakeasy lets a layperson end-user design a GOFAI, which only &lt;em&gt;narrowly&lt;/em&gt; uses LLMs as &amp;quot;common sense modules&amp;quot; or natural language parsers/generators.&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;&lt;u&gt;Speakeasy only has access to black-boxed LLMs&lt;/u&gt;. You can&#39;t train or design new LLMs or ANNs with it. (You&#39;ll still need Python for that.) So, this tool&#39;s accessibility is unlikely to advance fundamental AI Capabilities research.&lt;/li&gt;
&lt;li&gt;The lack of any major success cases from AutoGPT/similar over the last ~2 years is mild evidence that &lt;u&gt;adding scaffolding to LLMs &lt;i&gt;at their current capability&lt;/i&gt; isn&#39;t sufficient for AGI&lt;/u&gt;.
&lt;ul&gt;
&lt;li&gt;An analogy: no matter how much you train a group of rats &amp;amp; strap tools to them, you won&#39;t get an Einstein... but you &lt;em&gt;can&lt;/em&gt; get valuable uses, like &lt;a href=&quot;https://www.euronews.com/next/2022/11/11/rodents-to-the-rescue-trained-rats-with-hi-tech-backpacks-could-save-your-life-in-a-disast&quot;&gt;a search-and-rescue rat team&lt;/a&gt;! Likewise, I suspect scaffolded LLMs (&lt;em&gt;at their current capabilities&lt;/em&gt;) are unlikely to be a path to AGI, but may give us valuable use cases — including advancing AI Safety research — that also diminish the economic incentive to increase raw LLM capabilities.
&lt;ul&gt;
&lt;li&gt;Okay fine, &lt;em&gt;in principle&lt;/em&gt; one could train a bunch of rats to act like OR/NOT/AND gates, then use that to make an Einstein AGI. You know what I mean. Is it &amp;quot;tractable&amp;quot;.&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;&lt;u&gt;You &lt;i&gt;can&lt;/i&gt;, however, still do meaningful AI Safety research with this tool!&lt;/u&gt;  (As demonstrated in my above two projects, which both used Speakeasy, I could research: active preference elicitation, concept-based interpretability, evaluating AI bias, semi-automated AI Safety research)&lt;/li&gt;
&lt;li&gt;&lt;u&gt;Makes it easier to replicate &amp;amp; extend research, and encourages citizen science&lt;/u&gt;. Speakeasy&#39;s all on a web interface. For example, if you liked Project 1 or Project 2, and you have ideas for improvements... well, just pop it open, tweak it, and share. Bam. Its easy interface means even &lt;em&gt;non-programmers&lt;/em&gt; can contribute! Collaborative citizen science!
&lt;ul&gt;
&lt;li&gt;(Eventually. there&#39;s no public live demo yet, I&#39;ve only been working of this for 3 weeks.)&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;&lt;u&gt;Speakeasy may also reduce the economic incentive for more powerful, general agent-like AI?&lt;/u&gt;
&lt;ul&gt;
&lt;li&gt;If a layperson can serve almost all their use cases, by just &lt;em&gt;dragging-and-dropping stuff in a no-download web interface&lt;/em&gt;, or using or remixing something someone else made...&lt;/li&gt;
&lt;li&gt;...then there&#39;s much less incentive for companies to invest in R&amp;amp;D for making the LLMs &lt;em&gt;themselves&lt;/em&gt; more generally-agent-like, and more incentive to focus on &lt;em&gt;narrow AIs&lt;/em&gt; that can be easily, modular-ly plugged into scaffolding tools, like Speakeasy.&lt;/li&gt;
&lt;li&gt;(This is basically Drexler&#39;s &lt;a href=&quot;https://forum.effectivealtruism.org/topics/comprehensive-ai-services&quot;&gt;Comprehensive AI Services&lt;/a&gt; proposal)&lt;/li&gt;
&lt;li&gt;See next point for concrete examples.&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;&lt;u&gt;&lt;b&gt;Examples of scientifically/economically useful use cases for scaffolded LLMs:&lt;/b&gt;&lt;/u&gt;
&lt;ul&gt;
&lt;li&gt;Automated causal inference. (&lt;a href=&quot;https://blog.ncase.me/research-notes-oct-2024/#project_6&quot;&gt;See Project 6 below&lt;/a&gt;) Very useful for fields where we can&#39;t practically or ethically do human experiments, like economics, public health, social science, etc.&lt;/li&gt;
&lt;li&gt;Generating statistical prediction rules, which are simple, interpretable statistics rules, robustly proven to be better than doctors at medical diagnosis over decades of research in several fields. (&lt;a href=&quot;http://ereserve.library.utah.edu/Annual/PSY/6330/Crowell/grove.pdf&quot;&gt;a meta-analysis&lt;/a&gt;, &lt;a href=&quot;https://www.psycholosphere.com/Clinical%20versus%20statistical%20prediction%20-%20by%20William%20Grove.pdf&quot;&gt;a layperson-friendly summary&lt;/a&gt;)&lt;/li&gt;
&lt;li&gt;For education: &lt;em&gt;FINALLY&lt;/em&gt;, the dream of dirt-cheap, fully-personalized, 24/7 AI Tutors! (I&#39;ve personally already had a lot of success with this, especially after figuring out &lt;a href=&quot;https://blog.ncase.me/signal-boosts-aug-2024/#flashcards&quot;&gt;Claude can export Anki spaced repetition flashcards&lt;/a&gt;. Main problem is LLMs don&#39;t stick to long lesson structures, so scaffolding would be a value-add here.)&lt;/li&gt;
&lt;li&gt;News aggregator, summarizer, and fact/bias-checker. There is &lt;em&gt;too much&lt;/em&gt; out there.&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;&lt;u&gt;&lt;b&gt;Other possible AI Safety-related uses for narrow-purpose, scaffolded LLMs:&lt;/b&gt;&lt;/u&gt;
&lt;ul&gt;
&lt;li&gt;Amplifying human forecasting (for takeoffs &amp;amp; threat models)&lt;/li&gt;
&lt;li&gt;Amplifying human oversight (scaleable oversight)&lt;/li&gt;
&lt;li&gt;Amplifying human epistemics (e.g. a bot that helps you think more rigorously &lt;em&gt;and&lt;/em&gt; creatively)&lt;/li&gt;
&lt;li&gt;Black-box LLM research (e.g. &lt;a href=&quot;https://blog.ncase.me/research-notes-oct-2024/#project_2&quot;&gt;Project 2&lt;/a&gt;, or this &lt;a href=&quot;https://arxiv.org/pdf/2309.15840&quot;&gt;&amp;quot;AI lie detector&amp;quot;&lt;/a&gt;, or &lt;a href=&quot;https://arxiv.org/pdf/2307.15043&quot;&gt;&amp;quot;the universal LLM jailbreak&amp;quot;&lt;/a&gt;)&lt;/li&gt;
&lt;li&gt;LLMs as stand-ins for human subjects, in &lt;em&gt;pilot&lt;/em&gt; research papers on scalable oversight, human-computer interaction, cooperative AI, or psychology. (as tested in &lt;a href=&quot;https://arxiv.org/pdf/2407.04622&quot;&gt;Kenton et al 2024&lt;/a&gt; and &lt;a href=&quot;https://arxiv.org/pdf/2409.00128&quot;&gt;Cui, Li &amp;amp; Zhou 2024&lt;/a&gt;) Once you know it&#39;ll work on cheap fake humans, try it for real on real expensive humans.&lt;/li&gt;
&lt;li&gt;A chatbot that just scrapes &amp;amp; summarizes the latest AI-related news &amp;amp; research &amp;amp; ArXiv papers, then tells you. (admittedly this is dual-use)&lt;/li&gt;
&lt;li&gt;An AI Tutor to more quickly onboard people into AI Alignment/Governance/Safety/etc.&lt;/li&gt;
&lt;li&gt;Cloning MATS mentors, by shoving all their papers &amp;amp; blog posts into a RAG and plugging it into Speakeasy
&lt;ul&gt;
&lt;li&gt;I&#39;m kidding
&lt;ul&gt;
&lt;li&gt;Maybe
&lt;ul&gt;
&lt;li&gt;Come on, it&#39;d be convenient for you too, right? The simulated you can mentor me 24/7, while the &amp;quot;real&amp;quot; you can only do a couple hours per week.
&lt;ul&gt;
&lt;li&gt;This isn&#39;t helping my chances of getting accepted into the MATS program, is it.&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;hr /&gt;
&lt;p&gt;&lt;a id=&quot;project_4&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h2&gt;🎯 Project 4) Beating Goodhart&#39;s Law&lt;/h2&gt;
&lt;p&gt;&lt;strong&gt;Summary:&lt;/strong&gt; Goodhart&#39;s Law predicts that agents, human or AI, will tend to &amp;quot;game&amp;quot; any metric you reward it with. But &lt;em&gt;why&lt;/em&gt; is Goodhart&#39;s Law true? I take previous research that models Goodhart&#39;s with causal networks, then turn them into numerical simulations. With this, I found a robust way to beat Goodhart&#39;s: use an &lt;em&gt;ensemble of specifications&lt;/em&gt; / rewarded metrics, but: 1) pick metrics which are mostly &lt;em&gt;uncorrelated&lt;/em&gt;, and 2) &lt;em&gt;cap&lt;/em&gt; how much reward the agent gets per metric, so that one or two bad metrics don&#39;t screw over the whole ensemble.&lt;/p&gt;
&lt;p&gt;In sum, it&#39;s time to take your &lt;strong&gt;CUES: Capped, Uncorrelated Ensemble of Specifications&lt;/strong&gt;.&lt;/p&gt;
&lt;p&gt;. . .&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Related to:&lt;/strong&gt; Robust specification, Game theory, Theoretical AI Safety work, Numerical experiment&lt;/p&gt;
&lt;p&gt;. . .&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;The Theory:&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;&lt;em&gt;(modified cross-post from &lt;a href=&quot;https://blog.ncase.me/backlog/&quot;&gt;my Idea Dump Backlog&lt;/a&gt; from a few months ago)&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://aisafety.dance/media/p2/gms/goodhart.png&quot; alt=&quot;&quot; /&gt;&lt;/p&gt;
&lt;p&gt;&lt;a href=&quot;https://arxiv.org/pdf/1803.04585&quot;&gt;Manheim &amp;amp; Garrabrant 2019&lt;/a&gt; models Goodhart&#39;s Law as Pearl-esque causal diagrams. For example, let&#39;s say you&#39;re the boss of a news-writer. You care about Quality (ha! how old-fashioned...), which influences the rewarded-metric Views, but that&#39;s twice as influence-able by Clickbait. As a causal diagram:&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/media/backlog/causal_1.png&quot; alt=&quot;&quot; /&gt;&lt;/p&gt;
&lt;p&gt;We can convert this causal diagram (and &lt;em&gt;any&lt;/em&gt; causal diagram) into an approximation: a series of linear equations. For example:&lt;/p&gt;
&lt;p&gt;&#92;( views = quality + 2&#92;times clickbait &#92;)&lt;/p&gt;
&lt;p&gt;So if an agent (the news-writer) has limited amount of hours/effort they can put into Quality vs Clickbait, what will they do to maximize their rewarded-metric, Views? The optimal strategy is to put all their effort into Clickbait, since that has a higher coefficient than Quality!&lt;/p&gt;
&lt;p&gt;In &lt;em&gt;general&lt;/em&gt;, Goodhart&#39;s happens because True Goal influences a Metric, but that Metric&#39;s almost always more easily influenced by some Cheat.&lt;/p&gt;
&lt;p&gt;But what if we had &lt;em&gt;multiple&lt;/em&gt; Metrics, influenced by &lt;em&gt;mostly different&lt;/em&gt; Cheats?&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/media/backlog/causal_2.png&quot; alt=&quot;&quot; /&gt;&lt;/p&gt;
&lt;p&gt;As equations:&lt;/p&gt;
&lt;p&gt;&#92;( proxy_1 = Target + 2&#92;cdot noise_1 &#92;)&lt;br /&gt;
&#92;( proxy_2 = Target + 2&#92;cdot noise_2 &#92;)&lt;br /&gt;
&#92;( proxy_3 = Target + 2&#92;cdot noise_3 &#92;)&lt;/p&gt;
&lt;p&gt;Then: we have a first-draft solution to Goodhart&#39;s! Reward the agent not on just &lt;em&gt;one&lt;/em&gt; Metric, but &lt;em&gt;all of them added up.&lt;/em&gt; Mathematically, this will increase the True Goal&#39;s coefficient, hopefully above the coefficient of all other Cheats:&lt;/p&gt;
&lt;p&gt;&#92;( composite = 3&#92;cdot Target + 2 &#92;cdot noise_1 + 2 &#92;cdot noise_2 + 2 &#92;cdot noise_3 &#92;)&lt;/p&gt;
&lt;p&gt;(Note: see, the Metrics don&#39;t have to be &lt;em&gt;fully&lt;/em&gt; uncorrelated – I mean, they &lt;em&gt;shouldn&#39;t&lt;/em&gt;, they should all correlate with the True Goal – but even controlling for the True Goal, the Metrics can still have &lt;em&gt;some&lt;/em&gt; Cheats in common, as long as it&#39;s not &lt;em&gt;too&lt;/em&gt; much.)&lt;/p&gt;
&lt;p&gt;But what if a Cheat is really, &lt;em&gt;really&lt;/em&gt; powerful? That would give the Cheat the highest coefficient, messing up our strategy. So, a fix to that: &lt;strong&gt;Cap the maximum amount a Metric can contribute to the final reward.&lt;/strong&gt; This will prevent any powerful Cheats from having an outsized effect on the ensemble.&lt;/p&gt;
&lt;p&gt;&#92;(composite = min(proxy_1,1) + min(proxy_2,1) + min(proxy_3,1) + ...&#92;)&lt;/p&gt;
&lt;p&gt;&lt;em&gt;Now,&lt;/em&gt; the only way to get a high reward is to &lt;em&gt;actually invest effort in the True Goal, not Cheats.&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;Goodhart: Good-bye!&lt;/p&gt;
&lt;p&gt;. . .&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;The Numerical Experiments:&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;Here&#39;s the plan to numerically test the above: (I already did a sloppy version of this experiment last year, need to re-do it more rigorously, as described below. &lt;em&gt;(Update Oct 9: did it! see end of this section.)&lt;/em&gt; )&lt;/p&gt;
&lt;p&gt;Set up for the numerical simulation:&lt;/p&gt;
&lt;p&gt;1) &lt;strong&gt;Generate a random, two-layer linear network:&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/media/oct-2024/research/cues.png&quot; alt=&quot;&quot; /&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;The first layer has 1 True Goal, and &lt;em&gt;N&lt;/em&gt; Cheats.&lt;/li&gt;
&lt;li&gt;The second layer has &lt;em&gt;M&lt;/em&gt; Metrics.&lt;/li&gt;
&lt;li&gt;The True Goal causally influences (has an arrow to) all Metrics.&lt;/li&gt;
&lt;li&gt;There&#39;s a probability &lt;em&gt;p&lt;/em&gt; that a Cheat causally influences a Metric.&lt;/li&gt;
&lt;li&gt;The weights for each causal connection from Cheat→Metric are sampled from a power law. This is to model &lt;em&gt;really&lt;/em&gt; outsized Cheats.&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;2) &lt;strong&gt;Test Baseline:&lt;/strong&gt; See what would happen if we just rewarded the Metric with the highest causal connection.&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;The agent has a finite amount of Resources (1.00 unit total), that it can invest into the nodes at the first layer: True Goal, and/or Cheats.&lt;/li&gt;
&lt;li&gt;In this simulation, the agent picks the optimal distribution of Resources to maximize its reward. The agent will use &lt;a href=&quot;https://en.wikipedia.org/wiki/Simulated_annealing&quot;&gt;Simulated Annealing&lt;/a&gt;, which unlike gradient descent, is guaranteed to find a globally optimal-ish solution if you do it right.&lt;/li&gt;
&lt;li&gt;I then plot how much Resource goes into the True Goal in the agent&#39;s solution. If it&#39;s low, then Goodhart&#39;s Law struck.&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;3) &lt;strong&gt;Test ensemble &lt;em&gt;WITHOUT&lt;/em&gt; capping each Metric&#39;s influence.&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;Agent&#39;s reward is &lt;code&gt;Metric_1 + Metric_2 + ... + Metric_M&lt;/code&gt;.&lt;/p&gt;
&lt;p&gt;4) &lt;strong&gt;Test ensemble &lt;em&gt;AND&lt;/em&gt; cap each Metric&#39;s influence.&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;Agent&#39;s reward is &lt;code&gt;min(1, Metric_1) + min(1, Metric_2) + ... + min(1, Metric_M)&lt;/code&gt;.&lt;/p&gt;
&lt;p&gt;5) &lt;strong&gt;Compare!&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Update Oct 9th, 2024:&lt;/strong&gt; I did the above numerical simulation! &lt;strong&gt;&lt;a href=&quot;https://colab.research.google.com/drive/1m6HQSLcgBINFMlign0n6yJwHEwV-mYxD?usp=sharing&quot;&gt;Check out my Colab notebook.&lt;/a&gt;&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;Here&#39;s a randomly generated causal graph: (note I set p=1, Cheats can affect &lt;em&gt;every&lt;/em&gt; Metric; but their influence is power-law distributed.)&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/media/oct-2024/cues/graph.png&quot; alt=&quot;&quot; /&gt;&lt;/p&gt;
&lt;p&gt;Here&#39;s what the agent invests in, when we reward it based on a single Metric, or even an ensemble of added-up (uncapped) Metrics:&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/media/oct-2024/cues/cheat.png&quot; alt=&quot;&quot; /&gt;&lt;/p&gt;
&lt;p&gt;But here&#39;s what the agent invests in, when we reward it on an ensemble of all Metrics added-up in a &lt;em&gt;capped-influence&lt;/em&gt; way:&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/media/oct-2024/cues/cues.png&quot; alt=&quot;&quot; /&gt;&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/media/oct-2024/cues/planned.png&quot; alt=&quot;&quot; /&gt;&lt;/p&gt;
&lt;p&gt;. . .&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;More real-world empirical evidence that CUES may work:&lt;/strong&gt;  It&#39;s been long-well-known that SPRs&lt;sup class=&quot;footnote-ref&quot;&gt;&lt;a href=&quot;https://blog.ncase.me/research-notes-oct-2024/#fn2&quot; id=&quot;fnref2&quot;&gt;[2]&lt;/a&gt;&lt;/sup&gt;, which simply add up a bunch of metrics, do better than experts at diagnosis &amp;amp; prediction in a wide variety of fields. Amazingly, &lt;em&gt;unit-weight&lt;/em&gt; SPRs — where each metric is capped at an equal influence — do even better! The math behind CUES may finally explain why SPRs are so unreasonably effective.&lt;/p&gt;
&lt;p&gt;. . .&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Compare/contrasting to other work in AI Safety:&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;You may be thinking, &amp;quot;okay so what, we already know neural ensembles help with robust classification&amp;quot;. And, ok, yeah. But:
&lt;ol&gt;
&lt;li&gt;CUES is for the &lt;em&gt;other&lt;/em&gt; end of an ANN: an ensemble of &lt;em&gt;specifications&lt;/em&gt;.&lt;/li&gt;
&lt;li&gt;CUES hints at a &lt;em&gt;theoretical&lt;/em&gt; reason why ensembles work.&lt;/li&gt;
&lt;li&gt;CUES suggests &lt;em&gt;capped-influence&lt;/em&gt; ensembles do better than merely adding/averaging everything in an ensemble, to prevent one huge Cheat from screwing up everything.&lt;/li&gt;
&lt;/ol&gt;
&lt;/li&gt;
&lt;li&gt;&amp;quot;Wait, doesn&#39;t ensemble reward also show up in &lt;a href=&quot;https://proceedings.neurips.cc/paper_files/paper/2017/file/d5e2c0adad503c91f91df240d0cd4e49-Paper.pdf&quot;&gt;the famous RHLF paper&lt;/a&gt;?&amp;quot;  Kind of.  In the RHLF paper, the &amp;quot;rewards&amp;quot; aren&#39;t specified by the designer, they&#39;re &lt;em&gt;learnt&lt;/em&gt; in training – and further, the more training there is, the &lt;em&gt;closer&lt;/em&gt; they converge!  So the rewards in RHLF are &lt;em&gt;highly&lt;/em&gt; correlated!  And thus, so are their failure modes, hence (probably partly) why RHLF&#39;d LLMs are prone to &lt;a href=&quot;https://arxiv.org/pdf/2307.15043&quot;&gt;weird adversarial jailbreaks&lt;/a&gt;.
&lt;ul&gt;
&lt;li&gt;Update Oct 5: I &lt;em&gt;just&lt;/em&gt; learnt about Coste et al 2024: &lt;a href=&quot;https://arxiv.org/pdf/2310.02743&quot;&gt;“Reward Model Ensembles Help Mitigate Over-optimization”&lt;/a&gt;. So that&#39;s more evidence for &amp;quot;ensembles =&amp;gt; robustness&amp;quot;. The difference between this paper and CUES, is that the above paper talks about an ensemble of &lt;em&gt;learnt rewards&lt;/em&gt;, CUES is an ensemble of &lt;em&gt;specifications.&lt;/em&gt;
&lt;ul&gt;
&lt;li&gt;Put another way: that paper uses ensembles to deal with &lt;em&gt;inner alignment&lt;/em&gt;, CUES is meant to deal with &lt;em&gt;outer alignment&lt;/em&gt;.&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;The CUES strategy could also be implemented not just for training a whole agent, but training &lt;em&gt;each individual neuron.&lt;/em&gt; If the math behind CUES is right, then having a neural network where the &lt;em&gt;weights&lt;/em&gt; are capped (thus, capped-influence), should result in more robust networks.
&lt;ul&gt;
&lt;li&gt;&lt;em&gt;Is&lt;/em&gt; this true? I haven&#39;t tested it myself yet (though it wouldn&#39;t take long), but &lt;a href=&quot;https://thomas-tanay.github.io/post--L2-regularization/&quot;&gt;Tanay &amp;amp; Griffin 2018&lt;/a&gt; show that an MNIST classifier with L2 regularization (punishing big neural weights), was &lt;em&gt;much, much&lt;/em&gt; more robust to adversarial examples. (So: CUES could explain &lt;em&gt;why&lt;/em&gt; this finding happened!) With low L2 regularization, you could make an adversarial example that&#39;s visually indistinguishable to humans. &lt;em&gt;With&lt;/em&gt; large L2 regularization, it&#39;s obvious:&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/media/oct-2024/research/l2.png&quot; alt=&quot;&quot; /&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;But, I think this is the most exciting implication of CUES: &lt;strong&gt;IF TRUE, we don&#39;t need to come up with an ideal specification for AI to solve AI alignment, we just need a lot of mostly-uncorrelated-failure specifications!&lt;/strong&gt;  Examples:
&lt;ul&gt;
&lt;li&gt;An ensemble of ways to reward an AI agent: Average utility, Median utility, Regret, Imitating a demonstration, etc.&lt;/li&gt;
&lt;li&gt;An ensemble to train an ANN to model a human&#39;s reward function (inverse reward learning): Assume human is optimal, Assume human is Boltzmann rational, Assume exponential discounting, Assume hyperbolic discounting, Learn from behavior, Learn from speech, Learn from unconscious facial expressions, etc.
&lt;ul&gt;
&lt;li&gt;(In particular, I find &lt;a href=&quot;https://en.wikipedia.org/wiki/Human_Compatible#Russell&#39;s_three_principles&quot;&gt;Stuart Russell&#39;s proposal for AI Alignment&lt;/a&gt; promising, but I think the weakest point in this proposal is asking AI to infer our preferences &lt;em&gt;from our behavior&lt;/em&gt;. C&#39;mon, people procrastinate &amp;amp; reliably do things they know they won&#39;t want.)&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;CUES also ties into an idea for how we can align powerful AI to &lt;em&gt;moral, humane&lt;/em&gt; values, not just to a single human user (who may intend to use technically-aligned AI for inhumane ends). The idea is &lt;em&gt;Moral Parliament&lt;/em&gt; (&lt;a href=&quot;https://www.fhi.ox.ac.uk/wp-content/uploads/2021/06/Parliamentary-Approach-to-Moral-Uncertainty.pdf&quot;&gt;Newberry &amp;amp; Ord 2021&lt;/a&gt;): take a bunch of plausible moral theories (utilitarianism, deontology, contractarianism), then allow them to vote on decisions &lt;em&gt;in such a way that their affect is capped at a maximum&lt;/em&gt; (to prevent &amp;quot;Pascal Wager&amp;quot; style shenanigans, where a moral theory can say something&#39;s infinitely good/bad). So, Moral Parliament is [kinda] similar to CUES! (with the major difference being that the &amp;quot;caps&amp;quot; are at different thresholds for each moral theory, based on how much credence you give it.)&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;. . .&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;More possibilities / Challenges / Etc:&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;The above numerical simulation was for the very, &lt;em&gt;very&lt;/em&gt; simple case of a linear, single-layer causal network. To do: try this simulation with non-linear and/or multi-layer causal networks.&lt;/li&gt;
&lt;li&gt;The above sims also assume the Metric is &lt;em&gt;causally downstream&lt;/em&gt; of the True Goal.  Would it still work if the Metrics could also be &lt;em&gt;causally upstream?&lt;/em&gt;  Or even &lt;em&gt;causally confounded&lt;/em&gt; with some unrelated thing?&lt;/li&gt;
&lt;li&gt;In the sim, &lt;code&gt;min(1, Metric_x)&lt;/code&gt;caps all Metrics&#39; influence at 1.  In the real-world, how would we find the &amp;quot;caps&amp;quot;?  Maybe at one-sigma of what the Metric&#39;s distribution was recently?
&lt;ul&gt;
&lt;li&gt;Alternatively, to avoid an arbitrary hard-cap, what if I tried a logarithmic soft-cap? (I remember testing this last year &amp;amp; it didn&#39;t work well, I&#39;ll try again with this more rigorous setup)&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;Like the Moral Parliament thing mentioned above, what if we allowed the Metrics to have &lt;em&gt;different&lt;/em&gt; caps on their influence?  Say, proportional to how much credence we give in that Metric being a good proxy?&lt;/li&gt;
&lt;/ul&gt;
&lt;hr /&gt;
&lt;p&gt;&lt;a id=&quot;project_5&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h2&gt;🪞 Project 5) The game theory of self-modification&lt;/h2&gt;
&lt;p&gt;&lt;strong&gt;Summary:&lt;/strong&gt; What would happen if an AI gets the ability to modify itself? Would it choose to &amp;quot;wirehead&amp;quot;, or have its values slowly drift, or get locked into one set of values, or something else? Heck, what would happen if AI gets the ability to modify &lt;em&gt;our&lt;/em&gt; values? We endorse value-modification in &lt;em&gt;some&lt;/em&gt; cases – therapy, education, learning to love new people &amp;amp; art &amp;amp; food – but not other cases, like brainwashing. How do we formalize what&#39;s a &amp;quot;good&amp;quot; kind of self-modification (for AI or human), and what&#39;s &amp;quot;bad&amp;quot;?&lt;/p&gt;
&lt;p&gt;This project explores all of those questions via game theory. Alas, there&#39;s very few papers on game theory where an agent has the option to &lt;em&gt;modify their own utility function&lt;/em&gt; (or more).&lt;/p&gt;
&lt;p&gt;So, my trick: we can use the &lt;em&gt;standard, elementary&lt;/em&gt; tools of game theory, by treating all future versions of an agent (AI or human), at each time step, &lt;em&gt;as if it&#39;s a different agent&lt;/em&gt;. Playing games with &amp;amp; against your possible future selves!&lt;/p&gt;
&lt;p&gt;. . .&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Related to:&lt;/strong&gt; Game theory, Agent foundations, Theoretical work, Recursive self-improvement, Value draft, Value lock-in, Wireheading&lt;/p&gt;
&lt;p&gt;. . .&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Explanation:&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;&lt;em&gt;(modified cross-post from &lt;a href=&quot;https://blog.ncase.me/backlog/&quot;&gt;my Idea Dump Backlog&lt;/a&gt; from a few months ago)&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;You ever place your smartphone &lt;em&gt;across&lt;/em&gt; the room before you sleep, instead of next to your bed... so that when the alarm goes off, you&#39;re forced to get up to turn it off, &amp;amp; not be tempted to browse memes in bed?&lt;/p&gt;
&lt;p&gt;(Or you ever done some other &amp;quot;tricking your future self&amp;quot; thing?)&lt;/p&gt;
&lt;p&gt;Congratulations, you just broke a fundamental assumption of standard game theory, which is the basis of modern economics, political science, and Artificial Intelligence!&lt;/p&gt;
&lt;p&gt;That assumption is &lt;em&gt;we have preferences in some set order.&lt;/em&gt; But the above smartphone alarm example &lt;em&gt;isn&#39;t&lt;/em&gt; explainable with a stable preference-order:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;If you prefer &lt;code&gt;browsing memes &amp;gt; getting up&lt;/code&gt;, you wouldn&#39;t put your phone across the room, you&#39;d put it next to your bed to browse memes.&lt;/li&gt;
&lt;li&gt;If you prefer &lt;code&gt;getting up &amp;gt; browsing memes&lt;/code&gt;, you wouldn&#39;t &lt;em&gt;need&lt;/em&gt; to put your phone across the room, you&#39;d just get up when it&#39;s time to get up.&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;The only way this is explainable – and it&#39;s how we intuitively think about it anyway – is like &lt;em&gt;you&#39;re playing a game against your future self.&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;(See &lt;a href=&quot;https://www.youtube.com/watch?v=W-Cz-LK16g4&quot;&gt;: Jerry Seinfeld&#39;s Night Guy/Morning Guy skit&lt;/a&gt;)&lt;/p&gt;
&lt;p&gt;Here&#39;s the game&#39;s choice-tree:&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/media/backlog/choice_tree.png&quot; alt=&quot;Decision-tree visualizing above main text.&quot; /&gt;&lt;/p&gt;
&lt;p&gt;The trick is to analyze future versions of yourself &lt;em&gt;as if they&#39;re different players&lt;/em&gt;; then we can use the standard techniques of game theory to figure out what will happen! For example, above, Morning Guy wants to browse memes, but Night Guy &lt;em&gt;knows&lt;/em&gt; Morning Guy will do that, so Night Guy puts the phone across the room, to force Morning Guy to get up.&lt;/p&gt;
&lt;p&gt;(In this example, it&#39;s the agent&#39;s utility function getting modified overnight, involuntarily; but we can also extend the same logic to &lt;em&gt;voluntary&lt;/em&gt; utility changes -- as in wireheading, or il/legitimate value change!)&lt;/p&gt;
&lt;p&gt;My project is to to distill &amp;amp; expand on the very-little research so far, on the &amp;quot;game theory of self-modification&amp;quot;! This would have lots of applications to human behaviour &amp;amp; AI Safety. Like:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Given that we &lt;em&gt;can&#39;t&lt;/em&gt; be &amp;quot;ideal rational&amp;quot;, and &lt;em&gt;have&lt;/em&gt; to be &amp;quot;bounded rational&amp;quot;, what &amp;quot;meta-rational&amp;quot; tricks help work around our bounded rationality? (e.g. putting phone across the room)&lt;/li&gt;
&lt;li&gt;What kind of value change is legitimate? (&lt;a href=&quot;https://www.lesswrong.com/posts/mHQHBEuFcEWRnitp4/0-the-value-change-problem-introduction-overview-and&quot;&gt;Nora Ammann 2023&lt;/a&gt;) How do we change our values &lt;em&gt;according to our own values?&lt;/em&gt;&lt;/li&gt;
&lt;li&gt;When &amp;amp; why a self-modifying AI &lt;em&gt;won&#39;t&lt;/em&gt; just hack its own brain to get maximum reward (wireheading).&lt;/li&gt;
&lt;li&gt;Explaining &lt;a href=&quot;https://en.wikipedia.org/wiki/Human_Compatible#Russell&#39;s_three_principles&quot;&gt;Stuart Russell&#39;s proposal for AI Safety&lt;/a&gt;: the AI&#39;s only reward is to maximize the &lt;em&gt;human&#39;s&lt;/em&gt; reward, but the AI &lt;em&gt;is uncertain about it&lt;/em&gt; and &lt;em&gt;knows&lt;/em&gt; it&#39;s uncertain about it.&lt;/li&gt;
&lt;li&gt;Original(?) research showing that Russell&#39;s proposal is robust to &amp;quot;bounded rationality&amp;quot; &lt;em&gt;and&lt;/em&gt; self-modification!&lt;/li&gt;
&lt;li&gt;Original(?) research showing that a &amp;quot;meta-rational&amp;quot; AI using Russell&#39;s proposal -- contrary to the instrumental convergence/AI basic drives paradigm -- would choose to &lt;em&gt;NOT&lt;/em&gt; seek power, if it has too high a risk of its utility fn being corrupted? (e.g. Throw the One Ring into Mount Doom)&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;Usually, AI game theory is explained with dense math notation. At first I thought, &amp;quot;Why don&#39;t I explain it with readable pseudo-code&amp;quot;? Then I realized... Wait, why don&#39;t I just write &lt;em&gt;actual code&lt;/em&gt;, that readers can play with in the browser, to try their &lt;em&gt;own&lt;/em&gt; game-theory experiments? And so that&#39;s what I&#39;m doing!&lt;/p&gt;
&lt;p&gt;(It also may be that analytical, closed-form solutions aren&#39;t possible -- in which case, I can still present findings using numerical simulations.)&lt;/p&gt;
&lt;p&gt;Code of the Smartphone Alarm problem:&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/media/backlog/selfmod1.png&quot; alt=&quot;Screenshot of code&quot; /&gt;&lt;/p&gt;
&lt;p&gt;The thought-tree it produces: (actual code output!)&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/media/backlog/selfmod2.png&quot; alt=&quot;Screenshot of output&quot; /&gt;&lt;/p&gt;
&lt;p&gt;(What the above code is doing, is taking the game tree, then recursively analyzing it to predict what the agent will do at each step! The main difference from this &amp;amp; standard game theory, is that the &lt;em&gt;same&lt;/em&gt; agent is allowed to have different utility functions at different points.)&lt;/p&gt;
&lt;p&gt;I can also visualize thought-trees as procedurally generated trees. (It&#39;s bouncy because I&#39;m using spring-physics to figure out how the tree should be spaced out!&lt;sup class=&quot;footnote-ref&quot;&gt;&lt;a href=&quot;https://blog.ncase.me/research-notes-oct-2024/#fn3&quot; id=&quot;fnref3&quot;&gt;[3]&lt;/a&gt;&lt;/sup&gt;)&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/media/backlog/tree_viz.gif&quot; alt=&quot;tree_viz&quot; /&gt;&lt;/p&gt;
&lt;p&gt;That&#39;d make pretty pics for the ArXiv paper~&lt;/p&gt;
&lt;p&gt;. . .&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Prior Work:&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;a href=&quot;https://arxiv.org/pdf/1605.03142&quot;&gt;Everitt et al 2016&lt;/a&gt; prove that a rational agent that 1) plans ahead &amp;amp; 2) judges future outcomes based on their &lt;em&gt;current&lt;/em&gt; preferences will refuse to wirehead!&lt;/li&gt;
&lt;li&gt;However, that the paper assumes the AI is &lt;em&gt;perfectly&lt;/em&gt; rational. &lt;a href=&quot;https://arxiv.org/pdf/2011.06275&quot;&gt;Tětek &amp;amp; Sklenka 2021&lt;/a&gt; proved that an &lt;em&gt;imperfectly&lt;/em&gt;-rational (or &amp;quot;bounded rational&amp;quot;) agent&#39;s original goals would get exponentially corrupted under self-modification.&lt;/li&gt;
&lt;li&gt;&lt;em&gt;However&lt;/em&gt;, another caveat to &lt;em&gt;that&lt;/em&gt; is their paper assumes the AI is &lt;em&gt;unaware of their own bounded rationality&lt;/em&gt; (as they freely acknowledge in Section 6). This is where my research can build upon theirs: figure out if a bounded-rational agent &lt;em&gt;that&#39;s aware it&#39;s bounded rational&lt;/em&gt; can (probabilistically) maintain safe self-modification!&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;em&gt;(note: above was copy-pasted &amp;amp; modified from footnote 28 of my own explainer, &lt;a href=&quot;https://aisafety.dance/p2/&quot;&gt;AI Safety for Fleshy Humans: Part 2&lt;/a&gt;)&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;. . .&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;More possibilities / Challenges / Etc:&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;The above focuses on self-modification to &lt;em&gt;utility functions&lt;/em&gt;. What about self-modifications to one&#39;s explore-exploit rate, or exponential/hyperbolic discounting rates, or the very decision theory the agent uses?
&lt;ul&gt;
&lt;li&gt;For example: a &lt;em&gt;high&lt;/em&gt;-explore AI would be curious about self-modding to a &lt;em&gt;low&lt;/em&gt;-explore rate... but a &lt;em&gt;low&lt;/em&gt;-explore AI would be &lt;em&gt;not&lt;/em&gt; curious about self-modding to a &lt;em&gt;high&lt;/em&gt;-explore rate. So: would AIs that can modify their own explore rates, &lt;em&gt;by default&lt;/em&gt; tend towards incuriosity &amp;amp; rigidity? When yes, when no?&lt;/li&gt;
&lt;li&gt;&lt;a href=&quot;https://arxiv.org/pdf/1710.05060&quot;&gt;Yudkowsky &amp;amp; Soares 2018&lt;/a&gt; make an interesting case for Functional Decision Theory (FDT): they prove (under reasonable circumstances) that a standard Causal Decision Theory agent would &lt;em&gt;choose to self-modify&lt;/em&gt; to FDT, because &lt;em&gt;CDT endorses FDT over CDT.&lt;/em&gt;  An example of a decision theory endorsing replacing itself!&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;The above also only focuses on &amp;quot;value preservation&amp;quot; – but how do we make sense of cases where we &lt;em&gt;want&lt;/em&gt; our values to change? How is it even &lt;em&gt;possible&lt;/em&gt; to value &lt;em&gt;having different values?&lt;/em&gt; (e.g. &amp;quot;I&#39;m attracted to Bad Boys, but I wish I wasn&#39;t&amp;quot;.)
&lt;ul&gt;
&lt;li&gt;I think one trick to this is to treat each &amp;quot;value&amp;quot; in an agent &lt;em&gt;as if it&#39;s an agent itself.&lt;/em&gt;  (I think this is similar to &lt;a href=&quot;https://www.lesswrong.com/tag/shard-theory&quot;&gt;Shard Theory&lt;/a&gt;? or Minsky&#39;s Society of Mind.)&lt;/li&gt;
&lt;li&gt;Take the above Bad Boys example: let&#39;s say there&#39;s 3 values/sub-agents at play: Excitement (pro Bad Boy), Safety and Emotional Intimacy (anti Bad Boy).  If the brain was a democracy, the latter two agents want to have the first agent &amp;quot;voted off the island&amp;quot;. That&#39;s how one (as an ensemble whole) can value &lt;em&gt;changing&lt;/em&gt; your values.
&lt;ul&gt;
&lt;li&gt;Alternatively, imagine your brain like a co-operatively owned company: the &amp;quot;sub-agents&amp;quot; (individual values) democratically choose who to hire &amp;amp; fire. &lt;strong&gt;This is how you can modify your values &lt;em&gt;in line with your values&lt;/em&gt;.&lt;/strong&gt;&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;The above may also help formalize the line between what we consider &amp;quot;good&amp;quot; value-change and &amp;quot;bad&amp;quot; value-drift: did I change my values &lt;em&gt;according to my current values?&lt;/em&gt;
&lt;ul&gt;
&lt;li&gt;For example, right now, I value beliefs discovered through the scientific methods, and &lt;em&gt;not&lt;/em&gt; value beliefs acquired through &amp;quot;direct revelation&amp;quot;. If someone were to hypothetically drug me so hard I start believing in direct revelation, I&#39;d consider that a Bad Ending. But! If hard &lt;em&gt;scientific&lt;/em&gt; evidence were to come out about direct revelation working (e.g. someone publicly pre-registers visions of 100 lottery numbers and they all come true), &lt;em&gt;then&lt;/em&gt; if I update towards believing in direct revelation, that&#39;s a Good Ending.&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;One promising strategy for alignment is for an AI to &lt;em&gt;learn&lt;/em&gt; my true reward function. (Inverse reward learning) Would a smart-enough AI have an incentive to try to mis-learn my reward function as something easier to fulfill, or &lt;em&gt;actually modify me&lt;/em&gt; to have easier-to-fulfill preferences? &lt;a href=&quot;https://arxiv.org/pdf/1605.03143&quot;&gt;Everitt &amp;amp; Hutter 2016&lt;/a&gt; prove not, but only for optimally rational agents. Research Q: does it still hold for bounded-rational agents?&lt;/li&gt;
&lt;li&gt;I&#39;m using numerical simulation because my analytic-solution skills are crud. Maybe I could collab with someone with better analysis chops?&lt;/li&gt;
&lt;/ul&gt;
&lt;hr /&gt;
&lt;p&gt;&lt;a id=&quot;project_6&quot;&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h2&gt;➡️ Project 6) SCI: Semi-automated Causal Inference&lt;/h2&gt;
&lt;p&gt;&lt;strong&gt;Summary:&lt;/strong&gt; Make a hybrid AI (a GOFAI with an LLM as a module inside of it) to semi-automate &lt;a href=&quot;https://mixtape.scunning.com/&quot;&gt;causal inference&lt;/a&gt; from observational data, which will be highly scientifically/economically valuable in fields where it&#39;s impractical/unethical to do experiments, like epidemiology, economics, social science, etc.&lt;/p&gt;
&lt;p&gt;(also, I &lt;em&gt;may&lt;/em&gt; collab with MATS alum Egg Syntax on this; they&#39;re also already independently pursuing causal-inference-with-LLMs)&lt;/p&gt;
&lt;p&gt;. . .&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Related to:&lt;/strong&gt; Human-in-the-loop/Cyborgism/Narrow AI, Causal models/inference&lt;/p&gt;
&lt;p&gt;. . .&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Motivation:&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;Not directly related to AI Safety/Alignment, but it &lt;em&gt;may&lt;/em&gt; help by giving a proof-of-concept, to change the economic incentives?&lt;/p&gt;
&lt;p&gt;Concretely: if we can show a very scientifically/economically valuable use case for automation, that &lt;em&gt;does not require&lt;/em&gt; further advances in foundational models... that may(?) shift the incentive away from advancing those models, and towards figuring out how to plug-and-play &lt;em&gt;current&lt;/em&gt; AI:&lt;/p&gt;
&lt;p&gt;&lt;img src=&quot;https://blog.ncase.me/content/media/backlog/venn.png&quot; alt=&quot;&quot; /&gt;&lt;/p&gt;
&lt;p&gt;I think &lt;em&gt;causal inference&lt;/em&gt; is a big low-hanging fruit, here!&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Explanation:&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;This is “the” Scientific Process:&lt;/strong&gt;&lt;/p&gt;
&lt;ol&gt;
&lt;li&gt;Look at stuff&lt;/li&gt;
&lt;li&gt;Notice weird stuff (violations of model predictions)&lt;/li&gt;
&lt;li&gt;Generate hypotheses for why (generate causal and/or mathematical models)&lt;/li&gt;
&lt;li&gt;Generate tests that can distinguish between those models&lt;/li&gt;
&lt;li&gt;Run those tests&lt;/li&gt;
&lt;li&gt;Look at the results&lt;/li&gt;
&lt;li&gt;Repeat&lt;/li&gt;
&lt;/ol&gt;
&lt;p&gt;&lt;strong&gt;How can each part be automated?&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;(What I may do as part of a pilot test of this project...)&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Observation: Grab a bunch of data from Our World In Data or Gallup or similar.&lt;/li&gt;
&lt;li&gt;Noticing Weird Stuff:
&lt;ul&gt;
&lt;li&gt;Use good ol’ fashioned stats to pick up strong correlations, or sudden discontinuities/kinks in time-series data, etc&lt;/li&gt;
&lt;li&gt;Use an LLM as a “common sense module” to detect if something is weird? (e.g. “Hey Claude, if there was a 50% drop in youth self-harm in March 2020, how weird is that on a scale from 1-5?”)&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;Generate hypotheses (in the form of causal models)
&lt;ul&gt;
&lt;li&gt;I&#39;ll need to figure out how to scaffold/prompt LLMs to do this usefully (e.g. “Hey Claude, what major events happened in March 2020?” or “Hey Claude, what possible confounding factors could influence both being trans, and being a programmer?”)&lt;/li&gt;
&lt;li&gt;Human-in-the-loop: You the human can also add/subtract your own hypotheses&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;Generate tests to distinguish between causal models (Pearl-like causal diagrams)
&lt;ul&gt;
&lt;li&gt;This is good ol’ fashioned stats/causal inference. If there’s a correlation between A &amp;amp; B, and our three hypotheses are A causes B, B causes A, Confounder causes A &amp;amp; B... then to rule out Confounder, we “control for” Confounder and see if the correlation reduces or goes away.&lt;/li&gt;
&lt;li&gt;To test between A causes B or B causes A, if we have time-series data, we can do lagged-statistics-tests to see whether A or B leads/follows. Or use an “instrumental variable” which we know only affects A, and can’t affect B unless through A.&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;Run those tests! (LLM calls stats tools as tool use)
&lt;ul&gt;
&lt;li&gt;Bonus: If it realizes it doesn&#39;t have enough data, it can request you for more.&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;Look at results, Repeat!&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;. . .&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Challenges with testing this:&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;How can we know if this scaffolded LLM is inferring causal relations from scratch, or just pulling it up from its latent knowledge? For example, if we gave it lung cancer &amp;amp; smoking data, ideally it’d figure out smoking causes lung cancer... but if it outputs that, how would we know it figured it out vs just remembered that’s the correct answer, from its written corpus data?
&lt;ul&gt;
&lt;li&gt;One difference this system has, vs previous LLM-causal-inference tests, is that we do have a mix of LLMs and “good ol’ fashioned stats” — so, maybe we will be able to tell if it’s correctly going through all the steps of inferring smoking→cancer from scratch? (e.g. it should generate “maybe industrialization is a confounding factor” as a hypothesis, then generate tests to confirm/falsify that, like looking at non-smokers in industrialized areas.)&lt;/li&gt;
&lt;li&gt;That is: it should recreate not just the “correct” conclusion, but the entire process to rule out alternative hypotheses.&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;“Placebo tests”: this system should not find direct causal links between things we know to be spurious correlations, e.g. ice cream sales &amp;amp; drowning (confounded by summer swimming)&lt;/li&gt;
&lt;li&gt;How do I get a hypothesis’s “prior”? Use the “weirdness rating” again? Or a “simplicity rating”?&lt;/li&gt;
&lt;li&gt;Generating advice for more science, or for policy: If there is insufficient data to distinguish between two plausible models, it can ask for more data/experiments. Or, once we have a causal model, it can make policy recommendations for stuff we want. (For example, if it figures out what policies reliably cause less poverty or drug use, it can recommend that.)&lt;/li&gt;
&lt;/ul&gt;
&lt;hr /&gt;
&lt;p&gt;Alright, those are six AI/Alignment-related research projects I&#39;m working on! Let&#39;s see in one year&#39;s time if any of these go anywhere.&lt;/p&gt;
&lt;p&gt;Ciao,&lt;br /&gt;
~ Nicky Case&lt;/p&gt;
&lt;hr class=&quot;footnotes-sep&quot; /&gt;
&lt;section class=&quot;footnotes&quot;&gt;
&lt;ol class=&quot;footnotes-list&quot;&gt;
&lt;li id=&quot;fn1&quot; class=&quot;footnote-item&quot;&gt;&lt;p&gt;“[Russell &amp;amp; Norvig&#39;s famous AI textbook describes] a seemingly reasonable, but incorrect, reward function for a vacuum robot: if we reward the action of cleaning up dirt, the optimal policy causes the robot to repeatedly dump and clean up the same dirt.” (&lt;a href=&quot;https://people.eecs.berkeley.edu/~russell/papers/russell-nips16-cirl.pdf&quot;&gt;source&lt;/a&gt;) &lt;a href=&quot;https://blog.ncase.me/research-notes-oct-2024/#fnref1&quot; class=&quot;footnote-backref&quot;&gt;↩︎&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li id=&quot;fn2&quot; class=&quot;footnote-item&quot;&gt;&lt;p&gt;Bishop &amp;amp; Trout (2005) give a snappy overview: &lt;a href=&quot;https://academic.oup.com/book/36404/chapter-abstract/320070686?redirectedFrom=fulltext&quot;&gt;The Amazing Success of Statistical Prediction Rules&lt;/a&gt; (&lt;a href=&quot;https://joelvelasco.net/teaching/4330/bishop&amp;amp;trout-ch2.pdf&quot;&gt;pdf&lt;/a&gt;) &lt;a href=&quot;https://blog.ncase.me/research-notes-oct-2024/#fnref2&quot; class=&quot;footnote-backref&quot;&gt;↩︎&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li id=&quot;fn3&quot; class=&quot;footnote-item&quot;&gt;&lt;p&gt;It&#39;s the &lt;a href=&quot;https://en.wikipedia.org/wiki/Force-directed_graph_drawing&quot;&gt;force-directed graph drawing&lt;/a&gt; algorithm, but with the vertical (y) positions locked. &lt;a href=&quot;https://blog.ncase.me/research-notes-oct-2024/#fnref3&quot; class=&quot;footnote-backref&quot;&gt;↩︎&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ol&gt;
&lt;/section&gt;
</content>
        </entry>

    

</feed>
