Joint probabilities
Abstract
When combining information from multiple sources and attempting to estimate the probability of a conclusion, we often find ourselves in the position of knowing the probability of the conclusion conditional on each of the individual sources, but we have no direct information about the probability of the conclusion conditional on the combination of sources. The probability calculus provides no way of computing such joint probabilities. This paper introduces a new way of combining probabilistic information to estimate joint probabilities. It is shown that on a particular conception of objective probabilities, clear sense can be made of second-order probabilities (probabilities of probabilities), and these can be related to combinatorial theorems about proportions in finite sets as the sizes of the sets go to infinity. There is a rich mathematical theory consisting of such theorems, and the theorems generate corresponding theorems about secondorder probabilities. Among the latter are a number of theorems to the effect that certain inferences from probabilities to probabilities, although not licensed by the probability calculus, have probability 1 of producing correct results. This does not mean that they will always produce correct results, but the set of cases in which the inferences go wrong form a set of measure 0. Among these inferences are some enabling us to reasonably estimate the values of joint probabilities in a wide range of cases. A function called the Y-function is defined. The central theorem is the Y-Theorem, which tells us that if we know the individual probabilities for the different information sources and estimate the joint probability using the Y-function, the second-order probability of getting the right answer is 1. This mathematical result is tested empirically using a simple multi-sensor example. The Y-theorem agrees with Dempster's rule of combination in special cases, but not in general.