Tag Archives: ethics

Do what your parents say?

Should you feel compelled to pay any heed to what your parents want in your adult choices? I used to say a definite ‘no’. My mother said I should do whatever I liked, and I vowed to ignore her. From a preference utilitarian perspective, I guess that virtually all aspects of a person’s lifestyle make much more difference to a given person than to their parents. If you feel a sense of obligation in return for your parents giving you life, why? You made no agreement, your parents took their chances in full knowledge you might grow up to be anyone.

However what if fewer parents do take their chances with a greater risk of children being less satisfactory to them? The biggest effect of taking your parents’ preferences into account more could be via increasing the perception that children are worth having to other parents. It may be a small effect, but the value of life is high.

I’m not sure how much of a difference expected agreeableness of childen makes to people’s choices to have them. At first it may seem negligible. Most people seem to like their children a lot regardless of what they do. However if a person were guaranteed that their child would grow up to be exactly the opposite of what they admire, I would be surprised if there were no effect, so I must expect some gradient. I haven’t seen any data on this except my mother’s (joking?) claim that she would’ve aborted me had she thought I would be an economist. I’m not about to give up economics, but I do visit sometimes, and I painted the new living room and helped with my grandmother’s gardening since getting here this time. See how great descendent are? I would be interested if anyone has better data.

Choose pain-free utilitarianism

Some of my friends are hedonic utilitarians, or close human approximations (people whose excuse to talk excitedly in bars and on the internet is sometimes hedonic utilitarianism). I am a preference utilitarian, so I would like to talk excitedly on the internet about how they are wrong.

Robert Wiblin sums up a big motivation for hedonic utilitarianism:

“I am hedonic rather than a preference utilitarian because if I were aware of a being that wanted things but had no experiences I would not care about it as its welfare could not be affected”

Something like this seems a common reason. What makes a thing good or bad if not someone experiencing it as good or bad? And how can you consciously experience something as good or bad if it’s not some variation on pleasure and pain? If your wanting chocolate isn’t backed by being pleased by chocolate, why would I want you to have chocolate more than I would want any old unconscious chocolate-getting mechanism to have chocolate? Pleasure and pain are the distinctive qualia that connect normativeness to consciousness and make it all worthwhile.

This must be wrong. Pain at least can have no such importance, as empirically it can be decomposed into a sensation and a desire to not have the sensation. This is demonstrated by the medical condition pain asymbolia and by the effects of morphine for example. In both cases people say that they can still feel the sensation of the pain they had, but they no longer care about it.

To say that the sensation of pain is inherently bad then is no different than to say that the sensation of seeing the color red is inherently bad.  The leftover contender for making pain bad is the preference not to have pain. You may still care only about the sensation of having or fulfilling a preference, and not about preferences that are fulfilled outside of knowledge. The feeling of preferring could still be that sought after sensation inherently imbued with goodness or badness. It must be some variation on preferences though; hedonism’s values are built of them.

Natural cultural relativists?

When given the same ability to punish anyone, cooperative people want to punish members of groups they identify with more than they do outsiders, while less cooperative people want to punish outsiders more. From the Journal of Evolution and Human Behavior:

One of the most critical features of human society is the pervasiveness of cooperation in social and economic exchanges. Moreover, social scientists have found overwhelming evidence that such cooperative behavior is likely to be directed toward in-group members. We propose that the group-based nature of cooperation includes punishment behavior. Punishment behavior is used to maintain cooperation within systems of social exchange and, thus, is directed towards members of an exchange system. Because social exchanges often take place within groups, we predict that punishment behavior is used to maintain cooperation in the punisher’s group. Specifically, punishment behavior is directed toward in-group members who are found to be noncooperators. To examine this, we conducted a gift-giving game experiment with third-party punishment. The results of the experiment (N=90) support the following hypothesis: Participants who are cooperative in a gift-giving game punish noncooperative in-group members more severely than they punish noncooperative out-group members.

..[W]e predict that … punishment behavior is directed toward in-group members who are found to be noncooperators. To examine this, we conducted a gift-giving game experiment with third-party punishment. The results of the experiment (N=90) support the following hypothesis: Participants who are cooperative in a gift-giving game punish noncooperative in-group members more severely than they punish noncooperative out-group members.

The researchers’ conclusion is that punishment is just an extension of cooperation, and so applies in the same areas. They were not expecting, and haven’t got a good explanation for, uncooperative people’s interest in specifically punishing outsiders.

This provides a potential explanation for something I was wondering about. Middle class people often seem to talk about poor people and people from other cultures in terms of their actions being caused by bad external influences, in contrast to the language of free will and responsibility for their own kind. Discussion of Aboriginals in Australia regularly exemplifies this. e.g. SMH:

More than half the Aboriginal male inmates in prison for violent crimes are suffering from post traumatic stress disorder, an academic says.

And without effective intervention, the “stressors” for the disorder will be passed on to other generations, perpetuating the cycles of crime.

Dr Caroline Atkinson said most violent inmates had suffered from some form of family violence, alcohol and drug use, as well as profound grief and loss…

“It was a confronting experience being inside a cell with someone who has committed murder, but I quickly realised they are the ones with the answers and they had such amazing insight,” she said.

This is quite unlike news coverage I have seen of middle class white murderers. When we see faults as caused by external factors rather than free will or personal error, we aren’t motivated to punish. Is the common practice of coolly blaming circumstance when we talk about situations like violence in Aboriginal communities because the good, cooperative people who write about these things don’t identify with the groups they are talking about?

On a side note, is our ‘widening moral circle’ linked to greater desire to reform other cultures?

Everyone else prefers laws to values

How do you tell what a superhuman AI's values are? ( picture: ittybittiesforyou - see bottom)

How do you tell what a superhuman AI's values are? ( picture: ittybittiesforyou - see bottom)

Robin Hanson says that it is more important to have laws than shared values. I agree with him when ‘shared values’ means that shared indexical values remain about different people, e.g. If you and I share a high value of orgasms, you value you having orgasms and I value me having orgasms. Unless we are dating it’s all the same to me if you prefer croquet to orgasms. I think the singularitarians aren’t talking about this though. They want to share values in such a way that AI wants them to have orgasms. In principle this would be far better than having different values and trading. Compare gains from trading with the world economy to gains from the world economy’s most heartfelt wish being to please you. However I think that laws will get far more attention than values overall in arranging for an agreeable robot transition, and rightly so. Let me explain, then show you how this is similar to some more familiar situations.

Greater intelligences are unpredictable

If you know exactly what a creature will do in any given situation before it does it, you are at least as smart as it (if we don’t include it’s physical power as intelligence). Greater intelligences are inherently unpredictable. If you know the intelligence is trying to do, then you know what kind of outcome to expect, but guessing how it will get there is harder. This should be less so for lesser intelligences, and more so for more different intelligences. I will have less trouble guessing what a ten year old will do in chess against me than a grand master, though I can guess the outcome in both cases. If I play someone with a significantly different way of thinking about the game they may also be hard to guess.

Unpredictability is dangerous

This unpredictability is a big part of the fear of a superhuman AI. If you don’t know what path an intelligence will take to the goal you set it, you don’t know whether it will affect other things that you care about. This problem is most vividly illustrated by the much discussed case where the AI in question is suddenly very many orders of magnitude smarter than a human. Imagine we initially gave it only a subset of our values, such as our yearning to figure out whether P = NP, and we assume that it won’t influence anything outside its box. It might determine that the easiest way to do this is to contact outside help, build powerful weapons, take more resources by force, and put them toward more computing power. Because we weren’t expecting it to consider this option, we haven’t told it about our other values that are relevant to this strategy, such as the popular penchant for being alive.

I don’t find this type of scenario likely, but others do, and the problem could arise at a lesser scale with weaker AI. It’s a bit like the problem that every genie owner in fiction has faced. There are two solutions. One is to inform the AI about all of human values, so it doesn’t matter how wide it’s influence is. The other is to restrict its actions. SIAI interest seems to be in giving the AI human values (whatever that means), then inevitably surrendering control to it. If the AI will inevitably likely be so much smarter than humans that it will control everything fovever almost immediately, I agree that values are probably the thing to focus on. But consider the case where AI improves fast but by increments, and no single agent becomes more powerful than all of human society for a long time.

Unpredictability also makes it hard to use values to protect from unpredictability

When trying to avoid the dangers of unpredictability, the same unpredictability causes another problem for using values as a means of control. If you don’t know what an entity will do with given values, it is hard to assess whether it actually has those values. It is much easier to assess whether it is following simpler rules. This seems likely to be the basis for human love of deontological ethics and laws. Utilitarians may get better results in principle, but from the perspective of anyone else it’s not obvious whether they are pushing you in front of a train for the greater good or specifically for the personal bad. You would have to do all the calculations yourself and trust their information. You also can’t rely on them to behave in any particular way so that you can plan around them, unless you make deals with them, which is basically paying them to follow rules, so is more evidence for my point.

‘We’ cannot make the AI’s values safe.

I expect the first of these things to be a particular problem with greater than human intelligences. It might be better in principle if an AI follows your values, but you have little way to tell whether it is. Nearly everyone must trust the judgement, goodness and competency of whoever created a given AI, be it a person or another AI. I suspect this gets overlooked somewhat because safety is thought of in terms of what to do when *we* are building the AI. This is the same problem people often have thinking about government. They underestimate the usefulness of transparency there because they think of the government as ‘we’. ‘We should redistribute wealth’ may seem unproblematic, whereas ‘I should allow an organization I barely know anything about to take my money on the vague understanding that they will do something good with it’ does not. For people to trust AIs the AIs should have simple enough promised behavior that people using them can verify that they are likely doing what they are meant to.

This problem gets worse the less predictable the agents are to you. Humans seem to naturally find rules more important for more powerful people and consequences more important for less powerful people. Our world also contains some greater than human intelligences already: organizations. They have similar problems to powerful AI. We ask them to do something like ‘cheaply make red paint’ and often eventually realize their clever ways to do this harm other values, such as our regard for clean water. The organization doesn’t care much about this because we’ve only paid it to follow one of our values while letting it go to work on bits of the world where we have other values. Organizations claim to have values, but who can tell if they follow them?

To control organizations we restrict them with laws. It’s hard enough to figure out whether a given company did or didn’t give proper toilet breaks to its employees. It’s virtually impossible to work out whether their decisions on toilet breaks are as close to optimal according some popularly agreed set of values.

It may seem this is because values are just harder to influence, but this is not obvious. Entities follow rules because of the incentives in place rather than because they are naturally inclined to respect simple constraints. We could similarly incentivise organizations to be utilitarian if we wanted. We just couldn’t assess whether they were doing it. Here we find rules more useful and values less for these greater than human intelligences than we do for humans.

We judge and trust friends and associates according to what we perceive to be their values. We drop a romantic partner because they don’t seem to love us enough even if they have fulfilled their romantic duties. But most of us will not be put off using a product because we think the company doesn’t have the right attitude, though we support harsh legal punishments for breaking rules. Entities just a bit superhuman are too hard to control with values.

You might point out here that values are not usually programmed specifically in organizations, whereas in AI they are. However this is not a huge difference from the perspective of everyone who didn’t program the AI. To the programmer, giving an AI all of human values may be the best method of avoiding assault on them. So if the first AI is tremendously powerful, so nobody but the programmer gets a look in, values may matter most. If the rest of humanity still has a say, as I think they will, rules will be more important.

Mistakes with nonexistent people

Who is better off if you live and I die? Is one morally obliged to go around impregnating women? Is the repugnant conclusion repugnant? Is secret genocide OK? Does it matter if humanity goes extinct? Why shouldn’t we kill people? Is pity for the dead warranted?

All these discussions come down to the same question often: whether to care about the interests of people who don’t exist but could.

I shan’t directly argue either way; care about whatever you like. I want to show that most of the arguments against caring about the non-existent which repeatedly come up in casual discussion rely on two errors.

Here are common arguments (paraphrased from real discussions):

  1. There are infinitely many potential people, so caring about them is utterly impractical.
  2. The utility that a non-existent person experiences is undefined, not zero. You are calculating some amount of utility and attributing it to zero people. This means utility per person is x/0 = undefined.
  3. Causing a person to not exist is a victimless crime. Stop pretending these people are real just because you imagine them!
  4. If someone doesn’t exist, they don’t have preferences, so you can’t fulfil them. This includes not caring if they exist or not. The dead do not suffer, only their friends and relatives do that.
  5. Life alone isn’t worth anything – what matters is what happens in it, so creating a new life is a neutral act.
  6. You can’t be dead. It’s not something you can be. So you can’t say whether life is better.
  7. Potential happiness is immeasurable; the person could have been happy, they could have been sad. Their life doesn’t exist, so it doesn’t have characteristics.
  8. How can you calculate loss of future life? Maybe they’d live another hundred years, if you’re going to imagine they don’t die now.

All of these arguments spring from two misunderstandings:

Thinking of value as being a property of particular circumstances rather than of the comparison between choices of circumstances.

People who won't exist under any of our choices are of no importance (picture: Michelangelo)

People who won't exist under any of our choices are of no importance (picture: Michelangelo)

We need never be concerned with the infinite people who don’t exist. All those who won’t exist under any choice we might make are irrelevant.  The question is whether those who do exist under one choice we can make and don’t exist under another would be better off existing.

2, 3 and 4 make this mistake too. The utility we are talking about accrues in the possible worlds where the person does exist, and has preferences. Saying someone is worse off not existing is saying that in the worlds where they do exist they have more utility. It is not saying that where they don’t exist they experience suffering, or that they can want to exist when they do not.

Assuming there is nothing to be known about something that isn’t the case.

If someone doesn’t exist, you don’t just not know about their preferences. They actually don’t have any. So how can you say anything about them? If a person died now, how can you say anything about how long they would have lived? How good it could have been? It’s all imaginary. This line of thought underlies arguments 4-8.

But in no case are we discussing characteristics of something that doesn’t exist. We are discussing which characteristics are likely in the case where it does exist. This is very different.

If I haven’t made you a cake, the cake doesn’t have characteristics. To ask whether it is chocolate flavoured is silly. You can still guess that conditional on my making it it is more likely chocolate flavoured than fish flavoured. Whether I’ve made it already is irrelevant. Similarly you can guess that if a child were born it would be more likely to find life positive (as most people seem to) and to like music and food and sex and other things it’s likely to be able to get, and not to have an enourmous unsatisfiable desire for six to be prime. You can guess that conditional on someone’s life continuing, it would probably continue until old age. These are the sorts of things we uncontroversially guess all the time about our own futures, which are of course also conditional on choices we make, so I can’t see why they would become a problem when other potential people are involved.

Are there any good arguments that don’t rely on these errors for wanting to ignore those who don’t currently exist in consequentialist calculations?