Why personal experimentation?

Cross posted from Overcoming Bias. Comments there.

***

I experiment with many things, as do those around me. Some of this is randomization and explicit records, more is just trying different things, muttering ‘VOI‘ and repeating what felt good. I refer here to everything between a cautious banana-mustard-ham sandwich and polyamory.

Robin has suggested that I over-invest in such exploration. That most new things should be bad, and so most experimentation a private loss for public gain. What’s more, there shouldn’t be lots of low hanging fruit in trying things out. Most of the things humans frequently want to do (eat, sleep, change moods, organize time, learn, interact with others) should have been well figured out in ancient times. And anything that does still need checking out should be divided between many people.

Nonetheless, it looks to me like experimentation is worth it. Lots of the things we do seem barely satisfactory, there seem likely to be better alternatives, it seems hard to learn what has been tried for what ends, or what is good from listening to others or reading, and I and my friends seem to actually find good things by looking. e.g. Beeminderexplicit charity evaluation, unusual degrees of honesty, workflowy and explicit organization seem to often add value over the defaults, not to mention many tiny things, like banana-mustard-ham sandwiches.

If it is true that a lot of experimentation is worth it, we have a slight puzzle: if there is valuable information I might glean by experimentation, why hasn’t it been worth it for others in the past to collect it and put it where I can see it?

I will try to answer this over the next few posts. Before that, what do you think?

More signaling

Cross posted from Overcoming Bias. Comments there.

***

Centurion: Where is Brian of Nazareth?
Brian: You sanctimonious bastards!
Centurion: I have an order for his release!
Brian: You stupid bastards!
Mr. Cheeky: Uh, I’m Brian of Nazareth.
Brian: What?
Mr. Cheeky: Yeah, I – I – I’m Brian of Nazareth.
Centurion: Take him down!
Brian: I’m Brian of Nazareth!
Victim #1: Eh, I’m Brian!
Mr. Big Nose: I’m Brian!
Victim #2: Look, I’m Brian!
Brian: I’m Brian!
Victims: I’m Brian!
Gregory: I’m Brian, and so’s my wife!

– Monty Python’s Life of Brian

It’s easy for everyone to claim to be Brian. What Brian (and those who wish to identify him) need is a costly signal: an action that’s only worth doing if you are Brian, given that anyone who does the act will be released. In Brian’s life-or-death situation it is pretty hard to arrange such a thing. But in many other situations, costly signals can be found. An unprotected posture can be a costly signal of confidence in your own fighting ability, if this handicap is small for a competent fighter but dangerous for a bad fighter. College can act as a costly signal of diligence, if lazy, disorganized people who don’t care for the future would find attending college too big a cost for the improved job prospects.

A situation requires costly signaling when one party wishes to treat two types of people differently, but both types of people want to be treated in the better way. An analogous way to think of this as a game is that Nature decides between A or -A, then the sender looks at Nature’s choice, and gives a signal to the receiver, B or -B. Then the receiver takes an action, C or -C. The sender always wants the receiver to do C, but the receiver wants to do C if A and -C if -A. To stop the sender from lying, you can modify the costs to the sender of B and -B.

Suppose instead that the sender and the receiver perfectly agreed: either both wanted C always, or both wanted C if A and -C if -A. Then the players can communicate perfectly well even if all of the signals are costless – the sender has every reason to tell the receiver the truth.

If players can have these two kinds of preferences, and you have two players, these are the two kinds of signaling equilibria you can have (if the receiver always wants C, then he doesn’t listen to signals anyway).

Most of the communication in society involves far more than two players. But you might suppose it can be basically decomposed into two player games. That is, if two players who talk to each other both want C iff A, you might suppose they can communicate costlessly, regardless of who the first got the message from and where the message goes to. If the first one always wants C, you might expect costly signaling. If the second does, you might expect the message to be unable to pass that part in the chain. This modularity is important, because we mostly want to model little bits of big communication networks using simple models.

Surprisingly, this is not how signaling pairs fit together. To see this, consider the simplest more complicated case: a string of three players, playing Chinese Whispers. Nature chooses, the sender sees and tells an intermediary, who tells a receiver, who acts. Suppose the sender and the intermediary both always want C, while the receiver wants to act appropriately to Nature’s choice. By the above modular thesis, there will be a signaling equilibrium where the first two players talk honestly for free, and the second and third use costly signals between them.

Suppose everyone is following this strategy: the sender tells the intermediary whatever she sees, and the intermediary also tells the receiver honestly, because when he would like to lie the signal to do so is too expensive. Suppose you are the sender, and looking at Nature you see -A. You know that the other players follow the above strategy. So if you tell the intermediary -A, he will transmit this to the receiver, though he would rather not modulo signal prices. And that’s too bad for you, because you want C.

Suppose instead you lie and say A. Then the intermediary will pay the cost to send this message to the receiver, since he assumes you too are following the above set of strategies. Then the receiver will do what you want: C. So of course you lie to the intermediary, and send the message you want with all the signaling costs of doing so accruing to the intermediary. Your values were aligned with his before taking into account signaling costs, but now they are so out of line you can’t talk to each other at all. Given that you behave this way, he will quickly stop listening to you. There is no signaling equilibrium here.

In fact to get the sender to communicate honestly with the intermediary, you need the signals between the sender and the intermediary to be costly too. Just as costly as the ones between the intermediary and the receiver, assuming the other payoffs involved are the same for each of them. So if you add an honest signaling game before a costly signaling game, you get something that looks like two costly signaling games.

For example, take a simple model where scientists observe results, and tell journalists, who tell the public. The scientist and the journalist might want the public to be excited regardless of the results, whereas the public might want to keep their excitement for exciting results. In order for journalists who have exciting news to communicate it to the public, they need to find a way of sending signals that can’t be cheaply imitated by the unlucky journalists. However now that the journalists are effectively honest, scientists have reason to misrepresent results to them. So before information can pass through the whole chain, the scientists need to use costly signals too.

If you have an arbitrarily long chain of people talking to each other in this way, with any combination of these two payoff functions among the intermediaries, everyone who starts off always wanting C must face costly signals, of the same size as if they were in an isolated two player signaling game. Everyone who wants C iff A can communicate for free. It doesn’t matter whether communicating pairs are cooperative or not, before signaling costs. So for instance a whole string of people who apparently agree with one another can end up using costly signals to communicate because the very last one talks to someone who will act according to the state of the world.

So such things are not modular in the way you might first expect, though they are easily predicted by other simple rules. I’m not sure what happens in more complicated networks than strings. The aforementioned results might influence how networks form, since in practice it should be effectively cheaper overall to direct information through smaller numbers of people with the wrong type of payoffs. Anyway, this is something I’ve been working on lately. More here.

Filters and bottlenecks

Cross posted from Overcoming Bias. Comments there.

***

Lots of processes have filters: a certain proportion of the time they fail at that stage. There are filters in the path from dead stars to booming civilizations. There are filters in the path from being a baby to being an old person. There are filters in the path from having an idea to having a thriving business.

Lots of processes also have bottlenecks. These look similar, in that many things fail at that point. For instance the path to becoming a Nobel Prize winner is bottlenecked by there only being so many Nobel Prizes ever year. Rather than a fixed fraction of people getting past that barrier, a fixed number of people do.

It’s worth noticing if something is a filter or a bottleneck, because you should treat them differently often. Either way you can increase the fraction reaching the end by widening the filter or bottleneck to let more past. But for the filter it might be worth doing this at any other stage in the process, whereas for the bottleneck it is pointless at all earlier stages. You can’t get more Nobel Prize winners by improving education, but you might get more thriving businesses.

Can a tiny bit of noise destroy communication?

Cross posted from Overcoming Bias. Comments there.

***

If everyone knows a tenth of the population dishonestly claims to observe alien spaceships, this can make it very hard for the honest alien-spaceship-observer to communicate fact that she has actually seen an alien spaceship.

In general, if the true state of the world is seen as not much more likely than you sending the corresponding false message somehow, it’s hard to communicate the true state.

You might think there needs to be quite a bit of noise relative to true claims, or for acting on true claims to be relatively unimportant, for the signal to get drowned out. Yet it seems to me that a relatively small amount of noise could overwhelm communication, via feedback.

Suppose you have a network of people communicating one-on-one with one another. There are two possible mutually exclusive states of the world – A and B – which individuals occasionally get some info about directly. They can tell each other about info they got directly, and also about info they heard from others. Suppose that everyone likes for they and others to believe the truth, but they also like to say that A is true (or to suggest that it is more likely). However making pro-A claims is a bit costly for some reason, so it’s not worthwhile if A is false. Then everyone is honest, and can trust what one another says.

Now suppose that the costs people experience from making claims about A vary among the population. In the lowest reaches of the distribution, it’s worth lying about A. So there is a small amount of noise from people falsely claiming A. Also suppose that nobody knows anyone else’s costs specifically, just the distribution that costs are drawn from.

Now when someone gives you a pro-A message, there’s a small chance that it’s false. This slightly reduces the benefits to you of passing on such pro-A messages, since the value from bringing others closer to the truth is diminished. Yet you still bear the same cost. If the costs of sending pro-A messages were near the threshold of being too high for you, you will now stop sending pro-A messages.

From the perspective of other people, this decreases the probability that a given message of A is truthful, because some of the honest A messages have been removed. This makes passing on messages of A even less valuable, so more people further down the spectrum of costs find it not worthwhile. And so on.

At the same time as the value of passing on A-claims declines due to their likely falsehood, it also declines due to others anticipating their falsehood and thus not listening to them. So even if you directly observe evidence of A in nature, the value of passing on such claims declines (though it is still higher than for passing on an indirect claim).

I haven’t properly modeled this, but I guess for lots of distributions of costs this soon reaches an equilibrium where everyone who still claims A honestly finds it worthwhile. But it seems that for some, eventually nobody ever claims A honestly (though sometimes they would have said A either way, and in fact A happened to be true).

In this model the source of noise was liars at the bottom of the distribution of costs. These should also change during the above process. As the value of passing on A-claims declines, the cost threshold below which it is worth lying about such claims lowers. This would offset the new liars at the top of the spectrum, so lead to equilibrium faster. If the threshold becomes lower than the entire population, lying ceases. If others knew that this had happened, they could trust A-claims again. This wouldn’t help them with dishonest B-claims, which could potentially be rife, depending on the model. However they should soon lose interest in sending false B-claims, so this would be fixed in time. However by that time it will be worth lying about A again. This is less complicated if the initial noise is exogenous.

Could risk aversion be from friend thresholds?

Cross posted from Overcoming Bias. Comments there.

***

If you are going for a job that almost nobody is going to get, it’s worth trying to be unusual. Better that one in a hundred employers loves you and the rest hate you than all of them think you’re mediocre.

On the other hand, if you are going for a job that almost everybody who applies is going to get, best to be as close to normal as possible.

In general, if you expect to fall on the bad side of some important threshold, it’s good to increase your variance and maybe make it over. If you expect to fall on the good side, it’s good to decrease your variance and stay there. This is assuming you can change your variance without changing your mean too much.

This suggests people should be risk seeking sometimes, and risk averse other times, depending on where the closest or most important thresholds are for them.

Prospect theory and its collected evidence says that people are generally risk averse for gains, and risk seeking for losses. That is, if you offer them fifty dollars for sure or half a chance of a hundred, they’ll take the sure fifty. If you offer them minus fifty dollars for sure, or half a chance of minus one hundred, they’ll take the gamble. The proposed value function looks something like this:

 

The zero point is a ‘reference point’, usually thought to be something like expectations or the status quo. This means people feel differently about gaining fifty dollars vs. a fifty percent of one hundred, and being given one hundred then later offered minus fifty or a fifty percent chance of minus one hundred, even though these things are equivalent in payoffs.

Risk aversion in gains and risk seeking in losses is what you would expect if people were usually sitting right near an important threshold, regardless of how much they had gained or lost in the past. What important threshold might people always be sitting on top of, regardless of their movement?

One that occurs to me is their friends’ and acquaintances’ willingness to associate with them. Which I will explain in a minute.

Robin has suggested that people should have high variance when they are getting to know someone, to make it over the friend threshold. Then they should tone it down if they make it over, so they don’t fall back under again.

This was in terms of how much information a person should reveal. But suppose people take into account how successful your life is in deciding whether they want to associate with you. For a given friend’s admiration, you don’t have that much to gain by getting a promotion say, because you are already good enough to be their friend. You have more to lose by being downgraded in your career, because there is some chance they will lose interest in associating with you.

Depending on how good the friend is, the threshold will be some distance below you. But never above you, because I specified friends, not potential friends. This is relevant, because it is predominantly friends, not potential friends, who learn about details of your life. Because of this selection effect, most of the small chances you take run the risk of sending bad news to existing friends more than sending good news to potential friends.

If you think something is going to turn out well, you should be risk averse because there isn’t much to gain sending better news to existing friends, but there is a lot to lose from maybe sending bad news. If you think something is going to go a tiny bit badly, you still want to be risk averse, as long as you are a bit above the thresholds of all your acquaintances. But if you think it’s going to go more badly, a small chance of it not going badly at all might be more valuable than avoiding it going more badly.

This is less clear when things go badly, because the thresholds for each of your friends can be spread out in the space below you, so there might be quite a distance where losing twice as much loses you twice as many friends. But it is less clear that people are generally risk seeking in losses. They do buy insurance for instance. It’s also plausible that most of the thresholds are not far below you, if people try to associate with the best people who will have them.

Another feature of the prospect theory value function is that the loss region is steeper than the gain region. That also fits with the present theory, where mostly you just have things to lose.

In sum, people’s broad patterns of risk aversion according to prospect theory seem explicable in terms of  thresholds of association with a selection effect.

Can you think of a good way to test that?