Skip to content

Theory and Reality - Peter Godfrey-Smith

Bayesianism and modern theories of evidence

Here is the formula as stated by Bayes P(he)=P(eh)P(h)P(e)P(h|e) = \frac{P(e|h)P(h)}{P(e)}

Here it is again but in a philisophical sense P(he)=P(eh)P(h)P(eh)P(h)+P(e¬h)P(¬h)P(h|e) = \frac{P(e|h)P(h)}{P(e|h)P(h)+P(e|\neg h)P(\neg h)}

P(he)P(h|e) is the probability of h, given e. Which in this sense is understood to mean the probability of the hypothesis given the event(or evidence).

Given that we can find the probability of P(he)P(h|e) we can then question whether we can confirm the hypothesis. Where P(he)>P(h)P(h|e) > P(h) we could say that given the evidence, updates the posterior probability to be greater the prior, we can confirm the hypothesis. In this system proposed by bayesianism, there is the updating of the priors, in light of new evidence and there is the confirmation of h.

The following evidence is given. You hold a belief that your friend is at a party in a house. You are not around, but you're travelling to the party and before you get there you have a 70% (0.7) confidence that they will be there. We can say P(h)0.7P(h) - 0.7. You get to the house and you see your friends car. But seeing your friends car does not mean they are there. They could have parked there, went to the party and left by foot. So this is a probability of the car firstly, before the person. What is the proabability that the car is there given they are there P(eh)P(e|h) and the car is there but they are not P(e¬h)P(e|\neg h). These do not need to add up to any value. So we can say that P(eh)=0.9P(e|h) = 0.9 and P(e¬h)=0.1P(e|\neg h) = 0.1. From this, we can compute the probability that your friend is at the party in light of this new evidence. P(he)=0.90.7(0.90.7)+(0.10.3)=0.954545P(h|e) = \frac{0.9 \cdot 0.7}{(0.9 \cdot 0.7) + (0.1 \cdot 0.3)} = 0.954545. So we update our belief that they are at the party.

All of this is assuming that scientific theories are supposed to predict what we will see. Godfrey points out that we have to have a prior probability in order to update the posterior. He follows that there is a need to interpret bayesian priors into a system of scientific theories. Which is where the idea of subjectivist interpretations comes in. In this system, probabilities are degrees of belief about some hypothesis. This was pioneered by Frank Ramsey and Bruno de Finetti. This approach asks idealised gambling people to reveal their confidence in a belief through staking some money on that event. You want to set up a gambling system so that a person is punished for being wrong. At a 1:1 odds of a belief they will owe $1 for being wrong and for being right. This isn't saying that they think the theory useless. They hold some belief in it occuring. If they set the odds at 1000000:1 then they are saying that they are so confident they are right that they would risk losing a million dollars just to gain 1. But if you have one belief, you may have another, and they may be related. As we consider more and more beliefs, we can start to consider them a network of subjective probabilities. If there is coherency between all of these network of beliefs, then there is rationality according to bayesians.

The fundamentals of probability theory were developed by Kolmogorov. Kolmogorov had to create a proof for probability theory, which was before then a system of empirical observations (such as the law of large numbers). No proofs had been given before then. Subjectivist bayesians use the following 4 axioms according to Godfrey:

  1. All probabilities are between 0 and 1
  2. If a proposition is true(tautology), then the probability of it being true is 1
  3. If hh and hh' are exclusive alternatives(cannot both be true) then P(hh)=P(h)+P(h)P(h \lor h') =P(h)+P(h')
  4. P(hj)=P(hj)P(j)P(h|j) = \frac{P(h \land j)}{P(j)} provided that P(j)>0P(j) > 0

The Dutch Book Argument

The centrality of this argument is on beliefs that do not conform to the axioms. If a person holds two views that are exclusives, at a total probability of greater than 1, then they are being irrational, by violating axiom 3. For example, consider a coin toss. A person holds the following beliefs P(heads)=0.6P(heads) = 0.6 and P(tails)=0.6 P(tails) = 0.6. If taking a bet, of 1.5:1 they will lose $15 dollars for every $10 they gain. If a person(A Dutch Bookie) was offering a bet on this, the system the person holds is fully exploitable and losing. The subjectivist Beyesian is saying that so long as you do not violate the axioms, your subjective beliefs can be rational. They must also update their beliefs upon new evidence in accordance with bayes rule. So yesterdays posteriors become tomorrow's priors Pi1(he)=Pi(h)P_{i-1}(h|e) = P_{i}(h).

Some assessments of bayesianism

There is some free choice when picking a prior, as there is nothing to inform you in the first place. This can lead to two people picking vastly different priors. Person A can choose the priors P(h)=0.999999P(h) = 0.999999 and P(¬h)=0.000001P(\neg h) = 0.000001, while person B can choose the priors P(h)=0.000001P(h) = 0.000001 and P(¬h)=0.999999P(\neg h) = 0.999999. If they continue to update their priors, they will eventually reach some convergence. But if one person chooses a massive prior probability and updates accordingly, then this is still acceptable. In the case where they do come to some agreement, there will still be some small degree of disagreement about some theory.

There also exists a problem of inductive conclusions of priors. Goodman's riddle of induction would ask whether there is any difference between a person looking at all the evidence and inferring the moon is made of cheese and a person inferring it is composed of silica and such. Bayesian thinking wouldn't seem to argue there is a difference between either.

Elimination of alternatives

There exists an idea going forward about the role of confirmation in science, which concerns whether you can eliminate alternative hypotheses, sometimes called eliminative induction. This is the sort of reasoning that Sherlock Holmes does. You see how far an argument takes you only to rule it out, and finally arrive at some other alternative. John Platt argued for a modified Popperian view that would say that science does this eliminative inference, which can have a deductive or nondeductive form. There may be less decisive ruling out of alternatives through probabilities.

Criticisms of Popper

Popper's attempt to establish a criterion on demarcation has been heavily criticized, particularly with regard to the concept of measurement and observation. A universal statement may be rendered falsified by a specific statement, but how do we come to a conclusion about the statement? If we are measuring some effect, we must make a decision on whether to accept that observation. We might have issues with an observation that falsifies a theory or a law. For example, there have been instances of measurement errors when measuring light that have resulted in an example of matter traveling faster than light. Of course, there is no problem with this issue being overturned, but we still must make a decision about the observations, whether they will be accepted as relevant or not. The test can always be blamed, and there is no reason to abandon a theory due to logic, even with surprising evidence.

We must then make a decision whether the observation is to be accepted or not. Scientific theories can be dealt with as if they weren't scientific. It is the difference in decision-making that allows for some continuation of falsification. We can decide to reject certain theories after they do not meet some criteria, but there is no logical reason for doing so.

A prediction that something is unlikely is also non-scientific according to Popper's definition because it does not state that something will not happen, but rather that the probability is very low. The hypothesis that a coin is fair may be held even after a million occurrences of a head. This is the difference between falsification in practice and reality. Popper accepted that probability was not a constituent of a logically coherent theory of science. The consequence is that falsification can occur without a deductive logic between observations and theories. For example, consider a bridge being built with two theories of construction, one old and tested, and the other new and untested. Neither is falsified, but the problem comes when we consider which theory to choose. Godfrey-Smith takes it that Popper rejected the idea that probability of this kind could play a role in scientific decisions about theories of this kind. There is no reason why the untested one is any better than the tried and tested one. The ideas of Platt on strong inference are modifications of Popper's idea.

Risk and Science

The idea of science being an endeavor of risk is praised, as we ought to be subjecting our theories to more and more risk. This is seen as a virtuous act. Theories can appear to have a lot of contact with observation, but in reality, it is often pseudo contact. However, the question of how we deal with this exposure to new ideas is still in need of more thought.