From: M. Taylor Saotome-Westlake Date: Sun, 17 Jul 2022 16:35:47 +0000 (-0700) Subject: Sunday memoir confrontation 1: pass over existing text X-Git-Url: http://534655.efjtl6rk.asia/source?a=commitdiff_plain;h=b71aa035779ef150c5c98d633e9eae6568fcc2eb;p=Ultimately_Untrue_Thought.git Sunday memoir confrontation 1: pass over existing text --- diff --git a/content/drafts/a-hill-of-validity-in-defense-of-meaning.md b/content/drafts/a-hill-of-validity-in-defense-of-meaning.md index f489c4e..0e3dd82 100644 --- a/content/drafts/a-hill-of-validity-in-defense-of-meaning.md +++ b/content/drafts/a-hill-of-validity-in-defense-of-meaning.md @@ -10,11 +10,13 @@ Status: draft Recapping our story so far—in a previous post, ["Sexual Dimorphism in Yudkowsky's Sequences, in Relation to My Gender Problems"](/2021/May/sexual-dimorphism-in-the-sequences-in-relation-to-my-gender-problems/), I told the the part about how I've "always" (since puberty) had this obsessive sexual fantasy about being magically transformed into a woman and also thought it was immoral to believe in psychological sex differences, until I got set straight by these really great Sequences of blog posts by Eliezer Yudkowsky, which taught me (incidentally, among many other things) how absurdly unrealistic my obsessive sexual fantasy was given merely human-level technology, and that it's actually immoral _not_ to believe in psychological sex differences given that psychological sex differences are actually real. In a subsequent post, ["Blanchard's Dangerous Idea and the Plight of the Lucid Crossdreamer"](/2022/TODO/blanchards-dangerous-idea-and-the-plight-of-the-lucid-crossdreamer/), I told the part about how, in 2016, everyone in my systematically-correct-reasoning community up to and including Eliezer Yudkowsky suddenly starting claiming that guys like me might actually be women in some unspecified metaphysical sense, and insisted on playing dumb when confronted with alternative explanations of the relevant phenomena or even just asked what that means, until I eventually had a stress- and sleep-deprivation-induced delusional nervous breakdown, got sent to psychiatric prison once, and then went crazy again a couple months later. -That's not the really egregious part of the story. The thing is, psychology is a complicated empirical science: no matter how "obvious" I might think something is, I have to admit that I could be wrong—[not just as an obligatory profession of humility, but _actually_ wrong in the real world](https://www.lesswrong.com/posts/GrDqnMjhqoxiqpQPw/the-proper-use-of-humility). If my fellow rationalists merely weren't sold on the autogynephilia and transgender thing, I would certainly be disappointed, but it's definitely not grounds to denounce the entire community as a failure or a fraud. And indeed, I _did_ [end up moderating my views somewhat](/2022/Jul/the-two-type-taxonomy-is-a-useful-approximation-for-a-more-detailed-causal-model/) compared to the extent to which my thinking in 2016–7 took Blanchard–Bailey–Lawrence as received truth. At the same time, I don't particularly regret saying what I said in 2016–7, because Blanchard–Bailey–Lawrence is still very obviously _directionally_ correct compared to the nonsense everyone else was telling me. +That's not the really egregious part of the story. The thing is, psychology is a complicated empirical science: no matter how "obvious" I might think something is, I have to admit that I could be wrong—[not just as an obligatory profession of humility, but _actually_ wrong in the real world](https://www.lesswrong.com/posts/GrDqnMjhqoxiqpQPw/the-proper-use-of-humility). If my fellow rationalists merely weren't sold on the autogynephilia and transgender thing, I would certainly be disappointed, but it's definitely not grounds to denounce the entire community as a failure or a fraud. And indeed, I _did_ [end up moderating my views](/2022/Jul/the-two-type-taxonomy-is-a-useful-approximation-for-a-more-detailed-causal-model/) compared to the extent to which my thinking in 2016–7 took Blanchard–Bailey–Lawrence as received truth. At the same time, I don't particularly regret saying what I said in 2016–7, because Blanchard–Bailey–Lawrence is still very obviously _directionally_ correct compared to the nonsense everyone else was telling me. But a striking pattern in my attempts to argue with people about the two-type taxonomy in late 2016 and early 2017 was the tendency for the conversation to get _derailed_ on some variation of, "Well, the word _woman_ doesn't necessarily mean that," often with a link to ["The Categories Were Made for Man, Not Man for the Categories"](https://slatestarcodex.com/2014/11/21/the-categories-were-made-for-man-not-man-for-the-categories/), a 2014 post by Scott Alexander arguing that because categories exist in our model of the world rather than the world itself, there's nothing wrong with simply _defining_ trans people to be their preferred gender, in order to alleviate their dysphoria. -But this is wrong. To be clear, it's true _that_ categories exist in our model of the world, rather than the world itself—categories "map", not "territory"—and it's true that trans women might be women _with respect to_ some genuinely useful definition of the word "woman." However, the Scott Alexander piece that people kept linking to me goes much further, claiming that we can redefine gender categories _in order to make trans people feel better_: +After Yudkowsky had stepped away from full-time writing, Alexander had emerged as our subculture's preeminent writer. Most people in an intellectual scene "are writers" in some sense, but Alexander was the one "everyone" reads: you could reference a _Slate Star Codex_ post in conversation and expect people to familiar with the idea, either from having read it, or by osmosis, by other people having referenced the idea in conversation. The frequency with which "... Not Man for the Categories" was cited at me, seemed to suggest it had become our subculture's "party line" on trans issues. + +But the post is wrong in very obvious ways. To be clear, it's true _that_ categories exist in our model of the world, rather than the world itself—categories are "map", not "territory"—and it's true that trans women might be women _with respect to_ some genuinely useful definition of the word "woman." However, Alexander goes much further, claiming that we can redefine gender categories _in order to make trans people feel better_: > I ought to accept an unexpected man or two deep inside the conceptual boundaries of what would normally be considered female if it'll save someone's life. There's no rule of rationality saying that I shouldn't, and there are plenty of rules of human decency saying that I should. @@ -42,9 +44,13 @@ This is wrong because categories exist in our model of the world _in order to_ c In the case of Alexander's bogus argument about gender categories, the relevant principle ([#30](https://www.lesswrong.com/posts/d5NyJ2Lf6N22AD9PB/where-to-draw-the-boundary) on [the list of 37](https://www.lesswrong.com/posts/FaJaCgqBKphrDzDSj/37-ways-that-words-can-be-wrong)) is that if group things together in your map that aren't actually similar in the territory, you're going to be misled into making bad predictions. -Importantly, this is a very general point about how language itself works _that has nothing to do with gender_. No matter what you believe about politically controversial empirical questions, intellectually honest people should be able to agree that "I ought to accept an unexpected [X] or two deep inside the conceptual boundaries of what would normally be considered [Y] if [positive consequence]" is not the correct philosophy of language, _independently of the particular values of X and Y_. +Importantly, this is a very general point about how language itself works _that has nothing to do with gender_. No matter what you believe about politically-controversial empirical questions, intellectually honest people should be able to agree that "I ought to accept an unexpected [X] or two deep inside the conceptual boundaries of what would normally be considered [Y] if [positive consequence]" is not the correct philosophy of language, _independently of the particular values of X and Y_. + +Also, this ... really wasn't what I was trying to talk about. _I_ thought I was trying to talk about autogynephilia as an _empirical_ theory of psychology, the truth or falsity of which obviously cannot be altered by changing the meanings of words. But at this point I still trusted people in my robot cult to be basically intellectually honest, rather than fucking with me because of their political incentives, so I endeavored to respond to the category-boundary argument as if it were a serious argument: when I quit my dayjob in March 2017 in order to have more time to study and work on this blog, the capstone of my sabbatical was an exhaustive response to Alexander, ["The Categories Were Made for Man to Make Predictions"](/2018/Feb/the-categories-were-made-for-man-to-make-predictions/) (which Alexander [graciously included in his next links post](https://archive.ph/irpfd#selection-1625.53-1629.55)). A few months later, I followed it up with ["Reply to _The Unit of Caring_ on Adult Human Females"](/2018/Apr/reply-to-the-unit-of-caring-on-adult-human-females/), responding to a similar argument. I'm proud of those posts: I think Alexander's and _Unit of Caring_'s arguments were incredibly dumb, and with a lot of effort, I think I did a pretty good job of explaining exactly why to anyone who was interested and didn't, at some level, prefer not to understand. + +Of course, a niche blogger explaining why + -Also, this ... really wasn't what I was trying to talk about. _I_ thought I was trying to talk about autogynephilia as an _empirical_ theory of psychology, the truth or falsity of which obviously cannot be altered by changing the meanings of words. But at this point I still trusted people in my robot cult to be basically intellectually honest, rather than fucking with me because of their political incentives, so I endeavored to respond to the category-boundary argument as if it were a serious argument. So when I quit my dayjob in March 2017 in order to have more time to study and work on this blog, the capstone of my sabbatical was an exhaustive response to Alexander, ["The Categories Were Made for Man to Make Predictions"](/2018/Feb/the-categories-were-made-for-man-to-make-predictions/) (which Alexander [graciously included in his next links post](https://archive.ph/irpfd#selection-1625.53-1629.55)). A few months later, I followed it up with ["Reply to _The Unit of Caring_ on Adult Human Females"](/2018/Apr/reply-to-the-unit-of-caring-on-adult-human-females/), responding to a similar argument. I'm proud of those posts: I think Alexander's and _Unit of Caring_'s arguments were incredibly dumb, and with a lot of effort, I think I did a pretty good job of explaining exactly why to anyone who didn't, at some level, prefer not to understand. At this point, I was _disappointed_ with the limited impact of my work, but not to the point of bearing much hostility to "the community". People had made their arguments, and I had made mine; I didn't think I was _entitled_ to anything more than that.