From: Zack M. Davis Date: Tue, 29 Aug 2023 00:09:35 +0000 (-0700) Subject: check in X-Git-Url: http://534655.efjtl6rk.asia/source?a=commitdiff_plain;h=0cbfce99f1e04752dd07a30a7f7722f8534a65e5;p=Ultimately_Untrue_Thought.git check in --- diff --git a/content/2023/a-hill-of-validity-in-defense-of-meaning.md b/content/2023/a-hill-of-validity-in-defense-of-meaning.md index 1624c68..0ae6bf4 100644 --- a/content/2023/a-hill-of-validity-in-defense-of-meaning.md +++ b/content/2023/a-hill-of-validity-in-defense-of-meaning.md @@ -478,7 +478,7 @@ As it happened, the next day, we saw these Tweets from @ESYudkowsky, linking to (Why now? Maybe he saw the tag in my "tools have shattered" Tweet on Monday, or maybe the _Quillette_ article was just timely?) -The most obvious reading of these Tweets was as a political concession to me. The two-type taxonomy of MtF was the thing I was originally trying to talk about, back in 2016–2017, before getting derailed onto the present philosophy-of-language war, and here Yudkowsky was backing up my side on that. +The most obvious reading of these Tweets was as a political concession to me. The two-type taxonomy of MtF was the thing I was originally trying to talk about, back in 2016–2017, before getting derailed onto the present philosophy-of-language war, and here Yudkowsky was backing up my side on that. At this point, some readers might think that this should have been the end of the matter, that I should have been satisfied. I had started the recent drama flare-up because Yudkowsky had Tweeted something unfavorable to my agenda. But now, Yudkowsky was Tweeting something favorable to my agenda! Wouldn't it be greedy and ungrateful for me to keep criticizing him about the pronouns and language thing, given that he'd thrown me a bone here? Shouldn't I call it even? diff --git a/content/drafts/if-clarity-seems-like-death-to-them.md b/content/drafts/if-clarity-seems-like-death-to-them.md index 1637a85..714a5e3 100644 --- a/content/drafts/if-clarity-seems-like-death-to-them.md +++ b/content/drafts/if-clarity-seems-like-death-to-them.md @@ -13,19 +13,19 @@ Status: draft [^egan-paraphrasing]: The original quote says "one hundred thousand straights" ... "gay community" ... "gay and lesbian" ... "franchise rights on homosexuality" ... "unauthorized queer." -Recapping our Whole Dumb Story so far: in a previous post, "Sexual Dimorphism in Yudkowsky's Sequences, in Relation to My Gender Problems", I told you about how I've always (since puberty) had this obsessive erotic fantasy about being magically transformed into a woman and how I used to think it was immoral to believe in psychological sex differences, until I read these really great Sequences of blog posts by Eliezer Yudkowsky which incidentally pointed out how absurdly impossible my obsessive fantasy was ... +Recapping our Whole Dumb Story so far: in a previous post, ["Sexual Dimorphism in Yudkowsky's Sequences, in Relation to My Gender Problems"](/2021/May/sexual-dimorphism-in-the-sequences-in-relation-to-my-gender-problems/), I told you about how I've always (since puberty) had this obsessive erotic fantasy about being magically transformed into a woman and how I used to think it was immoral to believe in psychological sex differences, until I read these really great Sequences of blog posts by Eliezer Yudkowsky which [incidentally pointed out how absurdly impossible my obsessive fantasy was](https://www.lesswrong.com/posts/QZs4vkC7cbyjL9XA9/changing-emotions) ... -—none of which gooey private psychological minutiæ would be at all in the public interest to blog about _except that_, as I explained in a subsequent post, "Blanchard's Dangerous Idea and the Plight of the Lucid Crossdreamer", around 2016ish, everyone in the community that formed around the Sequences suddenly decided for political reasons that guys like me might actually be women in some unspecified metaphysical sense, and the cognitive dissonance of having to rebut all this nonsense coming from everyone I used to trust, drove me temporarily insane from stress and sleep deprivation ... +—none of which gooey private psychological minutiæ would be in the public interest to blog about _except that_, as I explained in a subsequent post, ["Blanchard's Dangerous Idea and the Plight of the Lucid Crossdreamer"](/2023/Jul/blanchards-dangerous-idea-and-the-plight-of-the-lucid-crossdreamer/), around 2016, everyone in the community that formed around the Sequences suddenly decided that guys like me might actually be women in some unspecified metaphysical sense, and the cognitive dissonance from having to rebut all this nonsense coming from everyone I used to trust drove me [temporarily](/2017/Mar/fresh-princess/) [insane](/2017/Jun/memoirs-of-my-recent-madness-part-i-the-unanswerable-words/) from stress and sleep deprivation ... -—which would have been the end of the story, _except that_, as I explained in a subsequent–subsequent post, "A Hill of Validity in Defense of Meaning", in late 2018, Eliezer Yudkowsky prevaricated about his own philosophy of language for the same political reasons, and my unsuccessful attempts to get him to clarify led me and allies to conclude that Yudkowsky and his "rationalists" were corrupt. +—which would have been the end of the story, _except that_, as I explained in a subsequent–subsequent post, ["A Hill of Validity in Defense of Meaning"](/2023/Jul/a-hill-of-validity-in-defense-of-meaning/), in late 2018, Eliezer Yudkowsky prevaricated about his own philosophy of language in the service of the gender-identity coalition, and my unsuccessful attempts to get him to clarify led me and allies to conclude that Yudkowsky and his "rationalists" were corrupt. Anyway, given that the "rationalists" were fake and that we needed something better, there remained the question of what to do about that, and how to relate to the old thing, and the operators of the marketing machine for the old thing. -_I_ had been hyperfocused on prosecuting my Category War, but the reason Michael and Ben and Jessica were willing to help me out on that, was not because they particularly cared about the gender and categories example, but because it seemed like a manifestation of a _more general_ problem of epistemic rot in "the community". +_I_ had been hyperfocused on prosecuting my Category War, but the reason Michael and Ben and Jessica were willing to help me out on that was not because they particularly cared about the gender and categories example, but because it seemed like a manifestation of a _more general_ problem of epistemic rot in "the community". Ben had [previously](http://benjaminrosshoffman.com/givewell-and-partial-funding/) [written](http://benjaminrosshoffman.com/effective-altruism-is-self-recommending/) a lot [about](http://benjaminrosshoffman.com/openai-makes-humanity-less-safe/) [problems](http://benjaminrosshoffman.com/against-responsibility/) [with](http://benjaminrosshoffman.com/against-neglectedness/) Effective Altruism. Jessica had had a bad time at MIRI, as she had told me back in March, and would [later](https://www.lesswrong.com/posts/KnQs55tjxWopCzKsk/the-ai-timelines-scam) [write](https://www.lesswrong.com/posts/MnFqyPLqbiKL8nSR7/my-experience-at-and-around-miri-and-cfar-inspired-by-zoe) [about](https://www.lesswrong.com/posts/pQGFeKvjydztpgnsY/occupational-infohazards). To what extent were my thing, and Ben's thing, and Jessica's thing, manifestations of "the same" underlying problem? Or had we all become disaffected with the mainstream "rationalists" for our own idiosyncratic reasons, and merely randomly fallen into each other's, and Michael's, orbit? -I believed that there _was_ a real problem, but didn't feel like I had a good grasp on what it was specifically. Cultural critique is a fraught endeavor: if someone tells an outright lie, you can, maybe, with a lot of effort, prove that to other people, and get a correction on that specific point. (Actually, as we had just discovered, even that might be too much to hope for.) But _culture_ is the sum of lots and lots of little micro-actions by lots and lots of people. If your _entire culture_ has visibly departed from the Way that was taught to you in the late 'aughts, how do you demonstrate that to people who, to all appearances, are acting like they don't remember the old Way, or that they don't think anything has changed, or that they notice some changes but think the new way is better? It's not as simple as shouting, "Hey guys, Truth matters!"—any ideologue or religious person would agree with _that_. It's not feasible to litigate every petty epistemic crime in something someone said, and if you tried, someone who thought the culture was basically on track could accuse you of cherry-picking. If "culture" is a real thing at all—and it certainly seems to be—we are condemned to grasp it unclearly, relying on the brain's pattern-matching faculties to sum over thousands of little micro-actions as a [_gestalt_](https://en.wiktionary.org/wiki/gestalt), rather than having the kind of robust, precise representation a well-designed AI could compute plans with. +If there _was_ a real problem, I didn't feel like I had a good grasp on what it was specifically. Cultural critique is a fraught endeavor: if someone tells an outright lie, you can, maybe, with a lot of effort, prove that to other people, and get a correction on that specific point. (Actually, as we had just discovered, even that might be too much to hope for.) But _culture_ is the sum of lots and lots of little micro-actions by lots and lots of people. If your _entire culture_ has visibly departed from the Way that was taught to you in the late 'aughts, how do you demonstrate that to people who, to all appearances, are acting like they don't remember the old Way, or that they don't think anything has changed, or that they notice some changes but think the new way is better? It's not as simple as shouting, "Hey guys, Truth matters!"—any ideologue or religious person would agree with _that_. It's not feasible to litigate every petty epistemic crime in something someone said, and if you tried, someone who thought the culture was basically on track could accuse you of cherry-picking. If "culture" is a real thing at all—and it certainly seems to be—we are condemned to grasp it unclearly, relying on the brain's pattern-matching faculties to sum over thousands of little micro-actions as a [_gestalt_](https://en.wiktionary.org/wiki/gestalt), rather than having the kind of robust, precise representation a well-designed AI could compute plans with. Ben called the _gestalt_ he saw the Blight, after the rogue superintelligence in Vernor Vinge's _A Fire Upon the Deep_: the problem wasn't that people were getting dumber; it was that there was locally coherent coordination away from clarity and truth and towards coalition-building, which was validated by the official narrative in ways that gave it a huge tactical advantage; people were increasingly making decisions that were better explained by their political incentives rather than acting on coherent beliefs about the world—using and construing claims about facts as moves in a power game, albeit sometimes subject to genre constraints under which only true facts were admissible moves in the game. @@ -37,7 +37,7 @@ I thought explaining the Blight to an ordinary grown-up was going to need _eithe The schism introduced new pressures on my social life. On 20 April 2019, I told Michael that I still wanted to be friends with people on both sides of the factional schism (in the frame where recent events were construed as a factional schism), even though I was on this side. Michael said that we should unambiguously regard Anna and Eliezer as criminals or enemy combatants (!!), that could claim no rights in regards to me or him. -I don't think I "got" the framing at this time. War metaphors sounded Scary and Mean: I didn't want to shoot my friends! But the point of the analogy (which Michael explained, but I wasn't ready to hear until I did a few more weeks of emotional processing) was specifically that soliders on the other side of a war _aren't_ particularly morally blameworthy as individuals:[^soldiers] their actions are being directed by the Power they're embedded in. +I don't think I "got" the framing at this time. War metaphors sounded Scary and Mean: I didn't want to shoot my friends! But the point of the analogy (which Michael explained, but I wasn't ready to hear until I did a few more weeks of emotional processing) was specifically that soliders on the other side of a war _aren't_ necessarily morally blameworthy as individuals:[^soldiers] their actions are being directed by the Power they're embedded in. [^soldiers]: At least, not blameworthy _in the same way_ as someone who committed the same violence as an individual. @@ -55,23 +55,23 @@ I may have subconsciously pulled off an interesting political thing. In my final > If we can't even get a public consensus from our _de facto_ leadership on something _so basic_ as "concepts need to carve reality at the joints in order to make probabilistic predictions about reality", then, in my view, there's _no point in pretending to have a rationalist community_, and I need to leave and go find something else to do (perhaps whatever Michael's newest scheme turns out to be). I don't think I'm setting [my price for joining](https://www.lesswrong.com/posts/Q8evewZW5SeidLdbA/your-price-for-joining) particularly high here? -And as it happened, on 4 May 2019, Yudkowsky [re-Tweeted Colin Wright on the "univariate fallacy"](https://twitter.com/ESYudkowsky/status/1124751630937681922)—the point that group differences aren't a matter of any single variable—which was _sort of_ like the clarification I had been asking for. (Empirically, it made me feel a lot less personally aggrieved.) Was I wrong to interpet this as another "concession" to me? (Again, notwithstanding that the whole mindset of extracting "concessions" was corrupt and not what our posse was trying to do.) +And as it happened, on 4 May 2019, Yudkowsky [re-Tweeted Colin Wright on the "univariate fallacy"](https://twitter.com/ESYudkowsky/status/1124751630937681922)—the point that group differences aren't a matter of any single variable—which was _sort of_ like the clarification I had been asking for. (Empirically, it made me feel a lot less personally aggrieved.) Was I wrong to interpet this as [another "concession" to me](/2023/Jul/a-hill-of-validity-in-defense-of-meaning/#proton-concession)? (Again, notwithstanding that the whole mindset of extracting "concessions" was corrupt and not what our posse was trying to do.) -Separately, on 30 April 2019, I visited the house where "Susan" and her husband Mike and Kelsey Piper and a few other people lived, which I'll call "Arcadia",[^named-houses] saying, essentially (and sincerely), Oh man oh jeez, Ben and Michael want me to join in a rationalist civil war against the corrupt mainstream-rationality establishment, and I'd really rather not, and I don't like how they keep using scary hyperbolic words like 'cult' and 'war' and 'criminal', but on the other hand, they're _the only ones backing me up_ on this _incredibly basic philosophy thing_ and I don't feel like I have anywhere else to _go_. The ensuing group conversation made some progress, but was mostly pretty horrifying. +Separately, on 30 April 2019, I visited the house where "Meredith" and her husband Mike and Kelsey Piper and a few other people lived, which I'll call "Arcadia",[^named-houses] saying, essentially (and sincerely), Oh man oh jeez, Ben and Michael want me to join in a rationalist civil war against the corrupt mainstream-rationality establishment, and I'd really rather not, and I don't like how they keep using scary hyperbolic words like 'cult' and 'war' and 'criminal', but on the other hand, they're _the only ones backing me up_ on this _incredibly basic philosophy thing_ and I don't feel like I have anywhere else to _go_. The ensuing group conversation made some progress, but was mostly pretty horrifying. [^named-houses]: It was common practice in our subculture to name group houses. My apartment was "We'll Name It Later." -In an adorable twist, Mike and "Susan"'s two-year-old son was reportedly saying the next day that Kelsey doesn't like his daddy, which was confusing until it was figured out he had heard Kelsey talking about why she doesn't like Michael _Vassar_.[^mike-pseudonym] +In an adorable twist, Mike and "Meredith"'s two-year-old son was reportedly saying the next day that Kelsey doesn't like his daddy, which was confusing until it was figured out he had heard Kelsey talking about why she doesn't like Michael _Vassar_.[^mike-pseudonym] [^mike-pseudonym]: I'm not giving Mike a pseudonym because his name is needed for this adorable anecdote to make sense, and this Whole Dumb Story isn't otherwise saying sensitive things about him. And as it happened, on 7 May 2019, Kelsey wrote [a Facebook comment displaying evidence of understanding my point](/images/piper-spending_social_capital_on_talking_about_trans_issues.png). -These two datapoints led me to a psychological hypothesis (which was maybe "obvious", but I hadn't thought about it before): when people see someone wavering between their coalition and a rival coalition, they're motivated to offer a few concessions to keep the wavering person on their side. Kelsey could _afford_ (_pace_ [Upton Sinclair](https://www.goodreads.com/quotes/21810-it-is-difficult-to-get-a-man-to-understand-something)) to not understand the thing about sex being a natural category ("I don't think 'people who'd get surgery to have the ideal female body' cuts anything at the joints"!!) when it was just me freaking out alone, but "got it" almost as soon as I could credibly threaten to _walk_ (defect to a coalition of people she dislikes) ... and maybe my "closing thoughts" email had a similar effect on Yudkowsky (assuming he otherwise wouldn't have spontaneously tweeted something about the univariate fallacy two weeks later)?? This probably wouldn't work if you repeated it (or tried to do it consciously)? +These two datapoints led me to a psychological hypothesis (which was maybe "obvious", but I hadn't thought about it before): when people see someone wavering between their coalition and a rival coalition, they're motivated to offer a few concessions to keep the wavering person on their side. Kelsey could _afford_ (_pace_ [Upton Sinclair](https://www.goodreads.com/quotes/21810-it-is-difficult-to-get-a-man-to-understand-something)) to not understand the thing about sex being a natural category when it was just me freaking out alone, but "got it" almost as soon as I could credibly threaten to _walk_ (defect to a coalition of people she dislikes) ... and maybe my "closing thoughts" email had a similar effect on Yudkowsky (assuming he otherwise wouldn't have spontaneously tweeted something about the univariate fallacy two weeks later)?? This probably wouldn't work if you repeated it (or tried to do it consciously)? ---- -I started drafting a "why I've been upset for five months and have lost faith in the so-called 'rationalist' community" memoir-post. Ben said that the target audience to aim for was people like I was a few years ago, who hadn't yet had the experiences I had—so they wouldn't have to freak out to the point of being imprisoned and demand help from community leaders and not get it; they could just learn from me. That is, the actual sympathetic-but-naïve people could learn. Not the people messing with me. +I started drafting a "why I've been upset for five months and have lost faith in the so-called 'rationalist' community" memoir-post. Ben said that the target audience to aim for was sympathetic-but-naïve people like I was a few years ago, who hadn't yet had the experiences I had—so they wouldn't have to freak out to the point of being imprisoned and demand help from community leaders and not get it; they could just learn from me. I didn't know how to continue it. I was too psychologically constrained; I didn't know how to tell the Whole Dumb Story without (as I perceived it) escalating personal conflicts or leaking info from private conversations. @@ -79,7 +79,7 @@ I decided to take a break from the religious civil war [and from this blog](/201 My dayjob performance had been suffering terribly for months. The psychology of the workplace is ... subtle. There's a phenomenon where some people are _way_ more productive than others and everyone knows it, but no one is cruel enough [to make it _common_ knowledge](https://slatestarcodex.com/2015/10/15/it-was-you-who-made-my-blue-eyes-blue/), which is awkward for people who simultaneously benefit from the culture of common-knowledge-prevention allowing them to collect the status and money rents of being a $150K/yr software engineer without actually [performing at that level](http://zackmdavis.net/blog/2013/12/fortune/), while also having [read enough Ayn Rand as a teenager](/2017/Sep/neither-as-plea-nor-as-despair/) to be ideologically opposed to subsisting on unjustly-acquired rents rather than value creation. The "everyone knows I feel guilty about underperforming, so they don't punish me because I'm already doing enough internalized domination to punish myself" dynamic would be unsustainable if it were to evolve into a loop of "feeling gulit _in exchange for_ not doing work" rather than the intended "feeling guilt in order to successfully incentivize work". I didn't think they would actually fire me, but I was worried that they _should_. -I asked my boss to temporarily take on some easier tasks, that I could make steady progress on even while being psychologically impaired from a religious war. (We had a lot of LaTeX templating of insurance policy amendments that needed to get done.) If I was going to be psychologically impaired _anyway_, it was better to be upfront about how I could best serve the company given that impairment, rather than hoping that the boss wouldn't notice. +I asked my boss to temporarily take on some easier tasks that I could make steady progress on even while being psychologically impaired from a religious war. (We had a lot of LaTeX templating of insurance policy amendments that needed to get done.) If I was going to be psychologically impaired _anyway_, it was better to be upfront about how I could best serve the company given that impairment, rather than hoping that the boss wouldn't notice. My "intent" to take a break from the religious war didn't take. I met with Anna on the UC Berkeley campus, and read her excerpts from some of Ben's and Jessica's emails. (She had not acquiesced to my request for a comment on "... Boundaries?", including in the form of two paper postcards that I stayed up until 2 _a.m._ on 14 April 2019 writing; I had figured that spamming people with hysterical and somewhat demanding physical postcards was more polite (and funnier) than my usual habit of spamming people with hysterical and somewhat demanding emails.) While we (my posse) were aghast at Yudkowsky's behavior, she was aghast at ours: reaching out to try to have a conversation with Yudkowsky, and then concluding he was a fraud because we weren't satisfied with the outcome was like hiding soldiers in an ambulance, introducing a threat against Yudkowsky in context where he had a right to be safe. @@ -89,31 +89,17 @@ It was true that she had tried to warn me for years, and (not yet having gotten My frustration bubbled out into follow-up emails: -> To: Anna Salamon <[redacted]> -> Date: 7 May 2019 12:53 _p.m._ -> Subject: Re: works cited -> > I'm also still pretty _angry_ about how your response to my "I believed our own propaganda" complaint is (my possibly-unfair paraphrase) "what you call 'propaganda' was all in your head; we were never _actually_ going to do the unrestricted truthseeking thing when it was politically inconvenient." But ... no! **I _didn't_ just make up the propaganda! The hyperlinks still work! I didn't imagine them! They were real! You can still click on them:** ["A Sense That More Is Possible"](https://www.lesswrong.com/posts/Nu3wa6npK4Ry66vFp/a-sense-that-more-is-possible), ["Raising the Sanity Waterline"](https://www.lesswrong.com/posts/XqmjdBKa4ZaXJtNmf/raising-the-sanity-waterline) -> -> Can you please _acknowledge that I didn't just make this up?_ Happy to pay you $200 for a reply to this email within the next 72 hours -

- -> To: Anna Salamon <[redacted]> -> Date: 7 May 2019 3:35 _p.m._ -> Subject: Re: works cited -> -> Or see ["A Fable of Science and Politics"](https://www.lesswrong.com/posts/6hfGNLf4Hg5DXqJCF/a-fable-of-science-and-politics), where the editorial tone is pretty clear that we're supposed to be like Daria or Ferris, not Charles. +[TODO: summarize smooth over cheerful-price bit, with a history footnote— -(This being a parable about an underground Society polarized into factions with different beliefs about the color of the unseen sky, and how different types of people react to the discovery of a passageway to the overworld which reveals that the sky is blue. Daria (formerly of the Green faction) steels herself to accept the unpleasant truth. Ferris reacts with delighted curiosity. Charles, thinking only of preserving the existing social order and unconcerned with what the naïve would call "facts", _blocks off the passageway_.) +> Can you please _acknowledge that I didn't just make this up?_ Happy to pay you $200 for a reply to this email within the next 72 hours -> To: Anna Salamon <[redacted]> -> Date: 7 May 2019 8:26 _p.m._ -> Subject: Re: works cited -> > But, it's kind of bad that I'm thirty-one years old and haven't figured out how to be less emotionally needy/demanding; feeling a little bit less frame-locked now; let's talk in a few months (but offer in email-before-last is still open because rescinding it would be dishonorable) -Anna said she didn't want to receive monetary offers from me anymore; previously, she had regarded my custom of throwing money at people to get what I wanted as good-faith libertarianism between consenting adults, but now she was afraid that if she accepted, it would be portrayed in some future Ben Hoffman essay as an instance of her _using_ me. She agreed that someone could have gotten the ideals I had gotten out of "A Sense That More Is Possible", "Raising the Sanity Waterline", _&c._, but there was also evidence from that time pointing the other way (_e.g._, ["Politics Is the Mind-Killer"](https://www.lesswrong.com/posts/9weLK2AJ9JEt2Tt8f/politics-is-the-mind-killer)), that it shouldn't be surprising if people steered clear of controversy. +] + +Anna said she didn't want to receive cheerful price offers from me anymore; previously, she had regarded my custom of throwing money at people to get what I wanted as good-faith libertarianism between consenting adults, but now she was afraid that if she accepted, it would be portrayed in some future Ben Hoffman essay as an instance of her _using_ me. She agreed that someone could have gotten the ideals I had gotten out of "A Sense That More Is Possible", "Raising the Sanity Waterline", _&c._, but there was also evidence from that time pointing the other way (_e.g._, ["Politics Is the Mind-Killer"](https://www.lesswrong.com/posts/9weLK2AJ9JEt2Tt8f/politics-is-the-mind-killer)), that it shouldn't be surprising if people steered clear of controversy. I replied: but when forming the original let's-be-apolitical vision in 2008, we did not anticipate that _whether or not I should cut my dick off_ would _become_ a political issue. That was _new evidence_ about whether the original vision was wise! I wasn't trying to do politics with my idiosyncratic special interest; I was trying to _think seriously_ about the most important thing in my life and only do the minimum amount of politics necessary to protect my ability to think. If 2019-era "rationalists" were going to commit a trivial epistemology mistake that interfered with my ability to think seriously about the most important thing in my life, but couldn't correct the mistake even after it was pointed out, then the 2019-era "rationalists" were _worse than useless_ to me personally. This probably didn't matter causally (I wasn't an AI researcher, therefore I didn't matter), but it might matter timelessly (if I was part of a reference class that includes AI researchers). @@ -219,7 +205,7 @@ Michael said that part of the reason this worked was because it represented a cl ------ -On 4 July 2019, Scott Alexander published ["Some Clarifications on Rationalist Blogging"](https://slatestarcodex.com/2019/07/04/some-clarifications-on-rationalist-blogging/), disclaiming any authority as a "rationalist" leader. ("I don't want to claim this blog is doing any kind of special 'rationality' work beyond showing people interesting problems [...] Insofar as [_Slate Star Codex_] makes any pretensions to being 'rationalist', it's a rationalist picnic and not a rationalist monastery.") I assumed this was inspired by Ben's request back in March that Scott "alter the beacon" so as to not confuse people about what the current-year community was. I appreciated it. +On 4 July 2019, Scott Alexander published ["Some Clarifications on Rationalist Blogging"](https://slatestarcodex.com/2019/07/04/some-clarifications-on-rationalist-blogging/), disclaiming any authority as a "rationalist" leader. ("I don't want to claim this blog is doing any kind of special 'rationality' work beyond showing people interesting problems [...] Insofar as [_Slate Star Codex_] makes any pretensions to being 'rationalist', it's a rationalist picnic and not a rationalist monastery.") I assumed this was inspired by [Ben's request back in March](/2023/Jul/a-hill-of-validity-in-defense-of-meaning/#alter-the-beacon) that Scott "alter the beacon" so as to not confuse people about what the current-year community was. I appreciated it. ------- @@ -299,7 +285,7 @@ I said that for me and my selfish perspective, the main outcome was finally shat ------- -I still wanted to finish the memoir-post mourning the "rationalists", but I still felt psychologically constrained; I was still bound by internal silencing-chains. So instead, I mostly turned to a combination of writing bitter and insulting comments whenever I saw someone praise the "rationalists" collectively, and—more philosophy blogging! +I wanted to finish the memoir-post mourning the "rationalists", but I still felt psychologically constrained, bound by internal silencing-chains. So instead, I mostly turned to a combination of writing bitter and insulting comments whenever I saw someone praise the "rationalists" collectively, and—more philosophy blogging! In August 2019's ["Schelling Categories, and Simple Membership Tests"](https://www.lesswrong.com/posts/edEXi4SpkXfvaX42j/schelling-categories-and-simple-membership-tests), I explained a nuance that had only merited a passion mention in "... Boundaries?": sometimes you might want categories for different agents to _coordinate_ on, even at the cost of some statistical "fit." (This was of course generalized from a "pro-trans" argument that had occured to me, [that self-identity is an easy Schelling point when different people disagree about what "gender" they perceive someone as](/2019/Oct/self-identity-is-a-schelling-point/).) @@ -313,7 +299,7 @@ I was _furious_ when "Against Lie Inflation" came out. (Furious at what I percei ----- -While visiting "Arcadia" on 7 August 2019, Mike and "Susan"'s son (age 2¾ years) asked me, "Why are you a boy?" +While visiting "Arcadia" on 7 August 2019, Mike and "Meredith"'s son (age 2¾ years) asked me, "Why are you a boy?" After a long pause, I said, "Yes," as if I had misheard the question as "Are you a boy?" I think it was a motivated mishearing: it was only after I answered that I consciously realized that's not what the kid asked. diff --git a/notes/critical_acclaim.md b/notes/critical_acclaim.md index bf9cff2..cd94a1a 100644 --- a/notes/critical_acclaim.md +++ b/notes/critical_acclaim.md @@ -213,4 +213,4 @@ https://twitter.com/adrusi/status/1681782160984883201 > > where 2017-zack... was on a different level -https://www.datasecretslox.com/index.php/topic,1553.msg38755.html +https://www.reddit.com/r/SneerClub/comments/10w6r33/sequel_to_a_beloved_siskind_classic_from_a/ diff --git a/notes/memoir-sections.md b/notes/memoir-sections.md index fbf68e3..e12e651 100644 --- a/notes/memoir-sections.md +++ b/notes/memoir-sections.md @@ -1,3 +1,30 @@ +Notes from pt. 3–6 readthrough (post publication of pt. 1–2)— +_ fullname Taylor and Hoffman at start of pt. 3 +_ be more specific about Ben's anti-EA and Jessica's anti-MIRI things, perhaps in footnotes +_ Ben on "locally coherent coordination": use direct quotes for Ben's language—maybe rewrite in my own language (footnote?) as an understanding test +_ set context for "EA Has a Lying Problem" (written by Sarah, likely with Michael's influence—maybe ask Sarah) +_ clarify schism (me and Vassar bros leaving the EA/rat borg?) +_ set context for Anna on first mention in the postq +_ more specific on "mostly pretty horrifying" and group conversation with the whole house +_ paragraph to explain the cheerful price bit +_ cut words from the "Yes Requires" slapfight? +_ better introduction of Steven Kaas +_ "Not the Incentives"—rewrite given that I'm not shielding Ray +_ cut many words from "Social Reality" scuffle +_ is "long May 2020" link still good? +_ better context on "scam" &c. earlier +_ Ben's "financial fraud don't inquire as to the conscious motives of the perp" claim may be false +_ later thoughts on jump to evaluation, translating between different groups' language +_ mention that I was miffed about "Boundaries?" not getting Curated, while one of Euk's animal posts did +_ examples of "bitter and insulting" comments about rationalists +_ cut words from descriptions of other posts! (if people want to read them, they can click through) +_ explicitly mention http://benjaminrosshoffman.com/bad-faith-behavior-not-feeling/ + + + + + +TODO blocks— _ "Lenore" psychiatric disaster - Eliezerfic fight conclusion _ Michael Vassar and the Theory of Optimal Gossip @@ -1380,6 +1407,10 @@ Still citing it (19 Sep 22): https://twitter.com/ingalala/status/156839169106472 https://arbital.greaterwrong.com/p/logical_dt/?l=5gc It even leaked into Big Yud!!! "Counterfactuals were made for humanity, not humanity for counterfactuals." +At least I don't have to link the rebuttal myself every time: +https://www.datasecretslox.com/index.php/topic,1553.msg38755.html +https://old.reddit.com/r/slatestarcodex/comments/10vx6gk/the_categories_were_made_for_man_not_man_for_the/j7k8fjc/ + ------ If you _have_ intent-to-inform and occasionally end up using your megaphone to say false things (out of sloppiness or motivated reasoning in the passion of the moment), it's actually not that big of a deal, as long as you're willing to acknowledge corrections. (It helps if you have critics who personally hate your guts and therefore have a motive to catch you making errors, and a discerning audience who will only reward the critics for finding real errors and not fake errors.) In the long run, the errors cancel out. @@ -2742,3 +2773,7 @@ https://twitter.com/aditya_baradwaj/status/1694355639903080691 Scott November 2020: "I think we eventually ended up on the same page" https://www.datasecretslox.com/index.php/topic,1553.msg38799.html#msg38799 + +> Or see ["A Fable of Science and Politics"](https://www.lesswrong.com/posts/6hfGNLf4Hg5DXqJCF/a-fable-of-science-and-politics), where the editorial tone is pretty clear that we're supposed to be like Daria or Ferris, not Charles. + +(This being a parable about an underground Society polarized into factions with different beliefs about the color of the unseen sky, and how different types of people react to the discovery of a passageway to the overworld which reveals that the sky is blue. Daria (formerly of the Green faction) steels herself to accept the unpleasant truth. Ferris reacts with delighted curiosity. Charles, thinking only of preserving the existing social order and unconcerned with what the naïve would call "facts", _blocks off the passageway_.) diff --git a/notes/memoir_wordcounts.csv b/notes/memoir_wordcounts.csv index 992682f..d54f64d 100644 --- a/notes/memoir_wordcounts.csv +++ b/notes/memoir_wordcounts.csv @@ -466,3 +466,35 @@ 07/26/2023,114379,0 07/27/2023,114276,-103 07/28/2023,113854,-422 +07/29/2023,113854,0 +07/30/2023,113854,0 +07/31/2023,113854,0 +08/01/2023,113854,0 +08/02/2023,113854,0 +08/03/2023,113854,0 +08/04/2023,113854,0 +08/05/2023,113854,0 +08/06/2023,113854,0 +08/07/2023,113854,0 +08/08/2023,113854,0 +08/09/2023,113854,0 +08/10/2023,113854,0 +08/11/2023,113854,0 +08/12/2023,113854,0 +08/13/2023,113854,0 +08/14/2023,113854,0 +08/15/2023,113854,0 +08/16/2023,113854,0 +08/17/2023,113854,0 +08/18/2023,113854,0 +08/19/2023,113854,0 +08/20/2023,113854,0 +08/21/2023,113854,0 +08/22/2023,113854,0 +08/23/2023,113854,0 +08/24/2023,113854,0 +08/25/2023,113854,0 +08/26/2023,113854,0 +08/27/2023,113660,-194 +08/28/2023,, +