Debiasing by rationalizing your own motives

Some time back, I saw somebody express an opinion that I disagreed with. Next, my mind quickly came up with emotional motives the other person might have for holding such an opinion, that would let me safely justify dismissing that opinion.

Now, it’s certainly conceivable that they did have such a reason for holding the opinion. People do often have all kinds of psychological, non-truth-tracking reasons for believing in something. So I don’t know whether this guess was correct or not.

But then I recalled something that has stayed with me: a slide from a presentation that Stuart Armstrong held several years back, that showed the way that we tend to think of our own opinions as being based on evidence, reasoning, etc.. And at the same time, we don’t see any of the evidence that caused other people to form their opinion, so instead we think of the opinions of others as being only based on rationalizations and biases.

Yes, it was conceivable that this person I was disagreeing with, held their opinion because of some bias. But given how quickly I was tempted to dismiss their view, it was even more conceivable that I had some similar emotional bias making me want to hold on to my opinion.

And being able to imagine a plausible bias that could explain another person’s position, is a Fully General Counterargument. You can dismiss any position that way.

So I asked myself: okay, I have invented a plausible bias that would explain the person’s commitment to this view. Can I invent some plausible bias that would explain my own commitment to my view?

I could think of several, right there on the spot. And almost as soon as I could, I felt my dismissive attitude towards the other person’s view dissolve, letting me consider their arguments on their own merits.

So, I’ll have to remember this. New cognitive trigger-action plan: if I notice myself inventing a bias that would explain someone else’s view, spend a moment to invent a bias that would explain my opposing view, in order to consider both more objectively.

The muted signal hypothesis of online outrage

Everyone, it sometimes seems, has their own pet theory of why social media and the Internet often seem like so unpleasant and toxic places. Let me add one more.

People want to feel respected, loved, appreciated, etc. When we interact physically, you can easily experience subtle forms of these feelings. For instance, even if you just hang out in the same physical space with a bunch of other people and don’t really interact with them, you often get some positive feelings regardless. Just the fact that other people are comfortable having you around, is a subtle signal that you belong and are accepted.

Similarly, if you’re physically in the same space with someone, there are a lot of subtle nonverbal things that people can do to signal interest and respect. Meeting each other’s gaze, nodding or making small encouraging noises when somebody is talking, generally giving people your attention. This kind of thing tends to happen automatically when we are in each other’s physical presence.

Online, most of these messages are gone: a thousand people might read your message, but if nobody reacts to it, then you don’t get any signal indicating that you were seen. Even getting a hundred likes and a bunch of comments on a status, can feel more abstract and less emotionally salient than just a single person nodding at you and giving you an approving look when you’re talking.

So there’s a combination of two things going on. First, many of the signals that make us feel good “in the physical world” are relatively subtle. Second, online interaction mutes the intensity of signals, so that subtle ones barely even register.

Depending on how sensitive you are, and how good you are generally feeling, you may still feel the positive signals online as well. But if your ability to feel good things is already muted, because of something like depression or just being generally in a bad mood, you may not experience the good things online at all. So if you want to consistently feel anything, you may need to ramp up the intensity of the signals.

Anger and outrage are emotional reactions with a very strong intensity, strong enough that you can actually feel them even in online interactions. They are signals that can consistently get similar-minded people rallied on your side. Anger can also cause people to make sufficiently strongly-worded comments supporting your anger that those comments will register emotionally. A shared sense of outrage isn’t the most pleasant way of getting a sense of belonging, but if you otherwise have none, it’s still better than nothing.

And if it’s the only way of getting that belonging, then the habit of getting enraged will keep reinforcing itself, as it will give all of the haters some of what they’re after: pleasant emotions to fill an emotional void.

So to recap:

When interacting physically, we don’t actually need to do or experience much in order to experience positive feelings. Someone nonverbally acknowledging our presence or indicating that they’re listening to us, already feels good. And we can earn the liking and respect of others, by doing things that are as small as giving them nonverbal signals of liking and respect.

Online, all of that is gone. While things such as “likes” or positive comments serve some of the same function, they often fail to produce much of a reaction. Only sufficiently strong signals can consistently break through and make us feel like others care about us, and outrage is one of the strongest emotional reactions around, so many people will learn to engage in more and more of it.

The parliamentary model as the correct ethical model

In 2009, Nick Bostrom brought up the possibility of dealing with moral uncertainty with a “parliamentary model” of morality. Suppose that you assign (say) 40% probability to some form particular of utilitarianism being correct, and 20% probability to some other form of utilitarianism being correct, and 20% probability to some form of deontology being true. Then in the parliamentary model, you imagine yourself as having a “parliament” that decides on what to do, with the first utilitarian theory having 40% of the delegates, the other form having 20% of the delegates, and the deontological theory having 20% of the delegates. The various delegates then bargain with each other and vote on different decisions. Bostrom explained:

The idea here is that moral theories get more influence the more probable they are; yet even a relatively weak theory can still get its way on some issues that the theory think are extremely important by sacrificing its influence on other issues that other theories deem more important. For example, suppose you assign 10% probability to total utilitarianism and 90% to moral egoism (just to illustrate the principle). Then the Parliament would mostly take actions that maximize egoistic satisfaction; however it would make some concessions to utilitarianism on issues that utilitarianism thinks is especially important. In this example, the person might donate some portion of their income to existential risks research and otherwise live completely selfishly.

As I noted, the model was proposed for dealing with a situation where you’re not sure of which ethical theory is correct. I view this somewhat differently. I lean towards the theory that the parliamentary model itself is the most correct ethical theory, as the brain seems to contain multiple different valuation systems that get activated in different situations, as well as multiple competing subsystems that feed inputs to these higher-level systems. (E.g. there exist both systems that tend to produce more deontological judgments, and systems that tend to produce more consequentialist judgments.)

Over time, I’ve settled upon something like a parliamentary model for my own decision-making. Different parts of me clearly tend towards different kinds of ethical frameworks, and rather than collapse into constant infighting, the best approach seems to go for a compromise where the most dominant parts get their desires most of the time, but less dominant parts also get their desires on issues that they care particularly strongly about. For example, a few days back I was considering the issue of whether I want to have children; several parts of my mind subscribed to various ethical theories which felt that the idea of having them felt a little iffy. But then a part of my mind piped up that clearly cared very strongly about the issue, and which had a strong position of “YES. KIDS”. Given that the remaining parts of my mind only had ambivalent or weak preferences on the issue, they decided to let the part with the strongest preference to have its way, in order to get its support on other issues.

There was a time when I had a strong utilitarian faction in my mind which did not want to follow a democratic process and tried to force its will on all the other factions. This did not work very well, and I’ve felt much better after it was eventually overthrown.

Confidence and patience don’t feel like anything in particular

After doing my self-concept work, I’ve been expecting to feel confident in social situations. And observing myself in them or after them, I have been more confident. But I haven’t felt particularly confident.

The thing is, being confident doesn’t feel like much in particular. I was pretty confident in my ability to open my laptop and write this post. I’m also confident in my ability to go to the shower and wash my hair, and I’m confident in my ability to go to the grocery store to buy stuff.

But writing this, or washing my hair, or going to the grocery store, aren’t things that would fill me with any particular “feeling of confidence”. They’re just things that I do, without thinking about them too much.

Similarly, being confident in a social situation doesn’t mean you’d actually have any strong feeling of confidence. It just means you don’t have any feeling of unconfidence.

Which is obvious when I think about it. So why did I expect otherwise?

I think the explanation is, the only times when I have previously paid conscious attention to my confidence, have been in situations where I’ve felt unconfident. And if you lack confidence, you try to psych yourself up. You try to summon some *other* emotion to flood your mind and push the feeling of unconfidence away.

If you are successfully suppressing your lack of confidence with some other emotion, you do “feel confident”. You are feeling whatever the other emotion is, that’s temporarily allowing you to be confident.

But if you don’t have any uncertainties that are actively surfacing, you don’t need to summon any other emotion to temporarily suppress them. Just those uncertainties not being around, is enough by itself. And something that’s just not around, doesn’t feel like anything.

Another similar thing is “patience”. If we feel impatient with someone, we might struggle to “try to be patient”. But if you actually are patient with someone, it usually doesn’t feel like anything in particular. You don’t have a glow of patience as you think about how badly the other person is getting on your nerves but how you withstand it anyway; rather the other person’s behavior just doesn’t bother you very much in the first place.

Edited to add: somebody pointed out that there exists good feeling of “you’ve got this” that one can feel. That’s true, and I agree that this could sensibly be called “confidence”. What I was trying to say was less “there’s no sensation that could reasonably be called confidence” but more “most everyday confidence doesn’t feel like anything in particular”. Paradoxically, even if confidence wouldn’t usually feel like anything, the lack of a feel can make you unconfident if you think that you should feel something to be confident. Somebody else mentioned that they do also have an actual feeling of patience; I’m not sure if I’ve experienced this myself, but the same thing applies.