Philosophy 5340

Topic 5: The Justification of Induction


1.  Hume’s Skeptical Challenge to Induction

        In the section entitled “Sceptical Doubts Concerning the Operations of the Understanding” in his Enquiry Concerning Human Understanding, Hume offers, in the following passage, an argument for the conclusion that inductive reasoning cannot be justified:

All reasonings may be divided into two kinds, namely, demonstrative reasoning, or that concerning relations of ideas, and moral reasoning, or that concerning matter of fact and existence. That there are no demonstrative arguments in the case seems evident; since it implies no contradiction that the course of nature may change, and that an object, seemingly like those which we have experienced, may be attended with different or contrary effects. May I not clearly and distinctly conceive that a body, falling from the clouds, and which, in all other respects, resembles snow, has yet the taste of salt or feeling of fire? Is there any more intelligible proposition than to affirm, that all the trees will flourish in December and January, and decay in May and June? Now whatever is intelligible, and can be distinctly conceived, implies no contradiction, and can never be proved false by any demonstrative argument or abstract reasoning a priori.

If we be, therefore, engaged by arguments to put trust in past experience, and make it the standard of our future judgment, these arguments must be probable only, or such as regard matter of fact and real existence according to the division above mentioned. But that there is no argument of this kind, must appear, if our explication of that species of reasoning be admitted as solid and satisfactory. We have said that all arguments concerning existence are founded on the relation of cause and effect; that our knowledge of that relation is derived entirely from experience; and that all our experimental conclusions proceed upon the supposition that the future will be conformable to the past. To endeavour, therefore, the proof of this last supposition by probable arguments, or arguments regarding existence, must be evidently going in a circle, and taking that for granted, which is the very point in question.

(Epistemology – Contemporary Readings, ed. Michael Huemer, pages 303-4)

        Hume’s argument here can be summarized as follows:

(1) It is logically possible that the future does not resemble the past.

Therefore,

(2) There can be no deductive justification for inductive inference.

(3) Any attempt to justify induction by appealing to an inductive inference would be circular, and would beg the question, since it would assume that inductive inference is justified.

(4) Circular justification is not acceptable.

Therefore,

(5) There can be no inductive justification for inductive inference.

(6) The only possible ways of justifying some proposition are either by deduction or by induction.

Therefore,

(7) There is no way of justifying inductive inference.


2.  Logical Probability and Hume’s Skeptical Challenge to Induction

    One response to Hume’s argument, advanced by David Stove in his book The Rationality of Induction (Oxford: Oxford University Press, 1986), is that Hume overlooks the possibility of the idea of logical probability.

    What is logical probability?  First of all, logical probability is a type of probability, where probability is something that obeys certain axioms.  If we use “Pr(p) = k” to say that the probability that proposition p is true is equal to k, then here is one standard way of formulating axioms for probability:

Axiom 1 (Normativity):             For any p, 0 ≤ Pr(p) ≤ 1.

Axiom 2 (Necessary truths):     If it is necessarily true that p, then Pr(p) = 1.

Axiom 3 (Additivity):               If p and q are logically incompatible, then Pr(p v q)  = Pr(p) + Pr(q).

Two theorems that follow from these three axioms are these:

Theorem 1:                        Pr(p) + Pr(~p) = 1.

Theorem 2 (Overlap):        Pr(p v q)  = Pr(p) + Pr(q) – Pr(p & q).

    Next, let us introduce the idea of conditional probability, using “Pr(q/p) = k” to say that the probability that proposition q is true given only that proposition p is true is equal to k.  Conditional probability is then defined as follows:

If Pr(p) > 0, then Pr(q/p) = def. Pr(q & p)/Pr(p).

(The restriction Pr(p) > 0 is needed since division by zero is mathematically undefined.)

    Given the definition of conditional probability, one can now prove a number of other useful theorems, including the following:

Theorem 3 (Multiplication):    Pr(q & p)= Pr(q/p) x Pr(p)

Theorem 4 (Total Probability):    If Pr(p) > 0, then Pr(p) = Pr(p/q) x Pr(q) + Pr(p/~q) x Pr(~q)

Theorem 5 (Logical Consequence):  If p => q, then Pr(p) ≤ Pr(q).

    In general, then, probability is anything that satisfies the axioms of probability.  But what is logical probability?  The answer is that the concept of logical probability is the concept of (1) a relation between a proposition and a number that (2) is a necessary relation, rather than a contingent one, and that (3) satisfies the axioms for the general concept of probability.

    So viewed, one can think of logical probability as akin to the relation of entailment between propositions, with logical probability being a more general relation.  In particular, one can think of (a) proposition p’s entailing proposition q as corresponding to its being the case that the logical probability of q given p is equal to one, and (a) proposition p’s entailing proposition ~q as corresponding to its being the case that the logical probability of q given p is equal to zero.

    Whether there is such a relation of logical probability is a controversial matter, with a number of philosophers having argued that there is no such relation.

3. The Justification of Induction: Two Very Different Types of Cases

        As Hume posed the problem of induction, it involved showing that it is reasonable to believe that regularities that have held in the past will continue to hold in the future.  Since the regularities that Hume had in mind were those associated with laws of nature, I prefer to say that the problem of induction upon which Hume focused was concerned with the question of how, if at all, one can prove that there are laws underlying the regularities in question.  (If one holds, as Hume did, that laws are nothing more than certain cosmic regularities, one can rephrase this by saying that the problem is how, if at all, one can demonstrate that certain regularities that have held in the past are in fact cosmic regularities, or cosmic regularities of the right sort.)

        Situations are very common, however, where one moves from information about events of a certain sort having some property, P, to a conclusion to the effect that further events of that sort are also likely to have property P, where one does not think that the reason that this is the case is that there is some underlying law.  An urn contains marbles, and after shaking up the urn, a marble is drawn which turns out to be red.  This action is repeated, say, 99 times, and in each case, the marble drawn is red.  It is then widely thought that the fact that the first 100 marbles drawn from the urn were all red makes it more likely than it initially was that the next marble will be red.  But one does not think that this is so because there is some underlying law that makes it the case either that all the marbles in the urn red, or that the probability that a marble in the urn is red is equal to k.

4. David Stove’s Approach to the Justification of Induction

        I think that induction is justified in both types of cases.  But I also think that the justification for the inductive inference is very different in the two cases.  In the marble-and-urn sort of case, I believe that one of the proofs of the justification of induction that David Stove offers in his book The Rationality of Induction (1986), which is based upon the approach of D. C. Williams in his book The Ground of Induction (1947), and which makes use of the statistical law of large numbers, is in principle sound.  Stove advanced that argument to prove, of course, not that inductive inferences were justified in marble-and-urn cases, but that they were justified when what one is attempting to arrive at are laws – or, if one prefers, relevant cosmic regularities.  In this, I think that Stove was mistaken.  In the marble-and-urn case, one is arriving at a conclusion about the next marble drawn from the urn based on a random selection of marbles from the urn.  Any marble in the urn could have been among the marbles drawn earlier, and I think that this is crucial.   But when one is attempting to show that it is reasonable to believe that there are certain laws of nature, one’s observations are not being selected randomly from the totality of events, since there is no possibility of future events being part of one’s sample.

        A way of thinking about this is to consider a case where there are two urns containing marbles, and all of the drawings of marbles are from the same urn.  The law of large numbers, properly formulated, can surely not be used to arrive at any conclusions concerning marbles in the other urn.  But if this is right, then how could it be otherwise when one replaces urns with distinct spatiotemporal regions – namely, the past and the future?

        In short, arriving at statistical conclusions about a population that one has sampled extensively, and arriving at laws, or nomologically based probabilities, are very different things.  So, for example, if observations of drawings involving one urn led to the conclusion that some law was involved – imagine, for example, that the drawing of a red marble was always followed by a drawing of a green marble, and vice versa – then one could project that finding from the one urn to the other.  But statistical information about the colors of marbles in one urn cannot be thus projected, and it seems to me that the same is true when references to urns are replaced by references to distinct spatiotemporal regions, however large the latter may be.

5.  Thomas Bayes and the Justification of Induction

    The idea that the concept of logical probability is relevant to the justification of induction goes back a long way – in particular, it goes back to Hume’s own time. Thus, of the possible responses to Hume’s skepticism concerning induction, the most interesting and the most important, in my opinion, is found in Thomas Bayes’ posthumously published “An Essay Towards Solving a Problem in the Doctrine of Chances” (1763).  But Bayes does not refer to David Hume, who was born in 1711 and died in 1776, and I suspect that Bayes was unaware of Hume’s argument for the conclusion that induction is unjustified.  But equally, one suspects, Hume was completely unaware of Bayes’ argument – as are almost all present-day philosophers, Bayesians included.

    In brief, after proving a number of minor things, including the theorem that now bears his name, Bayes went on to consider what was once often referred to as the problem of ‘inverse probability’, where this is the matter of determining the probable distribution of some unobserved (or unobservable) variable given information about the distribution of some known variable that depends upon the unknown variable.  So, for example, given information about how many times an uneven, weighted coin has landed heads in a given number of tosses, if one thought that the outcome of such tosses depended upon an unobservable propensity of the coin to land heads, determining the inverse probabilities would be a matter of determining the probabilities of different possible propensities.

6.  The Problem of Justifying Induction

Given the idea of inference to the best explanation, briefly discussed in an earlier lecture, it is natural to think that if one wants to find a justification for induction, starting out from the idea of inference to the best explanation is the way to go.  It seems to me, however, that that is not the route that one should travel, since it seems to me that a principle of inference to the best explanation is not at all a plausible candidate for a fundamental principle of inductive logic.  (I shall not, at this point, explain why I think that that is so, but this is something that we can discuss later.)

6.1 Rudolf Carnap and Inductive Logic

    If it is a mistake to think of some principle of inference to the best explanation as a fundamental principle, how should one approach questions concerning inductive inference? Here I have been strongly influenced by Rudolf Carnap’s book Logical Foundations of Probability, and it seems to me that what one needs to do is to think in terms of the concept of logical probability.

    Any such system of logical probability, to be satisfactory, has to rest upon fundamental principles of equiprobability.  Carnap, in his approach, thought in terms of two main alternatives.  One involved treating what he called “state descriptions” as equally probable.  The other involved treating what he called “structure descriptions” as equally probable.  Let me describe a very simple type of world that will give those of you who are not familiar with these notions an intuitive grasp of the difference, since these two notions will be important in what follows.

    Consider possible worlds where there are only three things – a, b, and c – and only two properties P and Q, where P and Q are incompatible with one another, and where everything must have one property or the other.  Then, for each object, there are two possibilities: either it has property P or it has property Q.  For the three objects, then, there are the following (2 x 2 x 2) = 8 possibilities:

State description 1:  a has P and b has P and c has P

State description 2:  a has P and b has P and c has Q

State description 3:  a has P and b has Q and c has P

State description 4:  a has Q and b has P and c has P

State description 5:  a has P and b has Q and c has Q

State description 6:  a has Q and b has P and c has Q

State description 7:  a has Q and b has Q and c has P

State description 8:  a has Q and b has Q and c has Q

Each of these eight possibilities is a state description.

    Next, there is the idea of a structure description.  The basic idea is that a structure description indicates only how many things have various properties and combinations of properties, but does not indicate which particular objects have the various properties.  So continuing with the example of worlds that contain just the three particulars a, b, and c and the two incompatible properties P and Q, one has the following four possible structure descriptions:

Structure description 1:  All three things have property P

Structure description 2:  Two things have property P, and one has property Q

Structure description 3:  One thing has property P, and two have property Q

Structure description 4:  All three things have property Q.

    Given these two ideas, one possible way of defining logical probability is by treating all state descriptions as equally likely, while another is to treat all structure descriptions as equally likely.  These two different choices will lead to different results, as is clear from the fact that while structure description 1 corresponds to just the one state description – namely, state description 1 – structure description 3 includes state descriptions 5, 6, and 7.

    Of these two approaches, the first seems very natural, while the second has no evident rationale.  But Carnap set out an argument, which we shall consider shortly, for the conclusion that if one treated state descriptions as equally likely, the result would be that one could never learn from experience.  As a result, he went with a definition of logical probability based on the idea that all structure descriptions are equally likely.   

6.2 Thomas Bayes’ Essay

        Thomas Bayes is, of course, a very well known figure, with a very familiar theorem that bears his name.  But that theorem is trivial, and was only a miniscule part of his essay.  Bayes was interested in proving much more substantial – in particular, theorems bearing upon the following problem which he states at the very beginning of his essay: 

Given the number of times in which an unknown event has happened and failed:

Required the chance that the probability of its happening in a single trial lies somewhere between any two degrees of probability that can be named.

    But if one can solve this problem, if one can establish a formula relating the probability that the objective chance of a certain sort of event lies between certain bounds, given information about the relative frequency of events of the sort in question, then one has done something very substantial indeed, for one has then solved the problem of justifying induction.  The question, then, is whether Bayes solved the problem of justifying induction, and did so in David Hume’s own lifetime, and by a method that Hume never considered.

    Now Hume, had he been aware of what Bayes had done, might well have objected that one could not make sense of the notion of chance with which Bayes was working.  Hume would have been right that there was no way of analyzing that notion at that time.  But it now seems clear that we are able to analyze that notion.  If so, that sort of Humean objection can no longer be sustained, and we need to confront the question of whether Bayes did succeed in justifying induction.

    My own view is that Bayes did not quite succeed, but that he was very much on the right track.  Bayes’ basic approach involved introducing the metaphysical idea of chances, or propensities, and then he adopted an equiprobability principle according to which, to put it a bit loosely, any two propensities of the same general type are equally likely.

    My objection to this type of approach grows out of my interest in laws of nature.  It is that I think that there are good reasons for holding that objective chances cannot be ultimate properties of things.  Instead, objective chances logically supervene on, and must be reduced to, causal laws of nature plus categorical properties and relations.  But if that is right, then a solution to the problem of justifying induction should be set out in terms of equiprobability principles that are formulated, not in terms of propensities, but, instead, in terms of laws of nature.    

7.  Analysis and New Alternatives in Metaphysics

    The availability of a method of analyzing theoretical terms that is compatible with a non-reductionist interpretation of those terms opened the door not only to the possibility, for example, of defending indirect realism as an account of perceptual knowledge, but also to the possibility of non-reductionist analyses of a number of very important metaphysical notions, including the ideas of causation, of propensities and objective chances, of dispositional properties, and of laws of nature.  Given the breakthrough in analysis, non-reductionist analyses of all of those concepts can now be given.

    The possibility of setting out such analyses does not, of course, show that the concepts in question are metaphysically unproblematic.  Thus it could turn out that just as in the case of the concept of a logically necessary person, where the vast majority of philosophers think that, although an analysis of that concept can be given, it turns out that the concept is such that it is logically impossible for there to be anything answering to that concept, so one might think, for example, that though one can offer non-reductionist analyses of causation, of propensities, and of laws of nature, it turns out, for some or all of those concepts, that it is logically impossible for there to be anything to which those concepts, thus analyzed, truly apply.

I shall not consider that issue here.  What I want to do here, instead, is simply to explore the relevance of this issue, in the case of laws of nature, to the problem of justifying induction, and what I shall argue is that is that the justification of induction stands or falls with whether it is possible to set out a coherent non-reductionist account of laws of nature.

8.  Reductionist Versus Non-Reductionist Accounts of Laws of Nature

    One of the great divides in contemporary metaphysics is that between philosophers who defend reductionist approaches to such things as laws of nature and causation, and those who defend non-reductionist approaches.  So let us consider this divide, in the case of laws of nature.

    What is involved in a reductionist approach to laws of nature?  There are various ways of explaining this, but here I think it will do simply to say that reductionist views of laws of nature involve the acceptance of something like the following thesis of Humean Supervenience:

All matters of fact logically supervene on states of affairs that consist of particulars having non-dispositional properties and standing in spatial, temporal, and spatiotemporal relations that do not involve causation.

    A reductionist approach to laws of nature, then, is an approach that holds that laws of nature logically supervene upon those sorts of states of affairs involving particulars.  A non-reductionist approach to laws of nature rejects this supervenience claim.

    Can a non-reductionist approach be characterized in a more positive way, rather than simply in terms of a rejection of Humean Supervenience?  The answer is that it can be, and the sort of account that I favor is essentially as follows:

Laws of nature are atomic states of affairs consisting of second-order relations between properties (universals) that, first of all, are not entailed by any set of Humean states of affairs, and that, secondly, in the case of non-probabilistic laws of nature, entail that some specific regularity involving Humean states of affairs obtains.

9. Non-Reductionism, Reductionism, and the Epistemological Challenge

9.1 The Challenge to Non-Reductionist Views of Laws of Nature


    A common objection to non-reductionist approaches to laws of nature is that, in postulating the existence of states of affairs that involve something more than Humean states of affairs, there is no way of justifying the belief in the existence of the extra ontological items that are being postulated.  Thus Barry Loewer, for example, in his paper “Humean Supervenience”, claims, “The metaphysics and epistemology of Humean laws, and more specifically, Lewis-laws, are in much better shape than the metaphysics and epistemology of the main anti-Humean alternatives.”   Loewer himself does not really offer much support for the epistemological part of this claim.  But this type of epistemological objection is certainly defended by others, most notably, perhaps, by John Earman and John T. Roberts, who devote a two-part, 56-page paper in Philosophy and Phenomenological Research to an attempt to establish this objection.                                                                                                                                                                 

9.2 The Challenge Reversed

        A crucial claim, then, which reductionists with regard to laws of nature advance, is that non-reductionist approaches to laws of nature face a serious epistemological challenge:  How can one possibly be justified in believing in the existence of anything more that cosmic regularities?  How can one be justified in believing in the existence of strong laws of nature, understood as atomic states of affairs involving second-order relations between universals that are supposed to underlie, and provide a basis for, regularities?  Reductionists with regard to laws of nature generally are confident, moreover, that this challenge cannot be met.

        In what follows, I shall attempt to do two things.  First of all, I shall refer to some things that can be proven that together show that if strong laws of nature are not logically possible, then a belief in reductionist laws of nature cannot only not be justified: it can be shown to be unjustified.  Secondly, I shall then cite other results that can also be proved, which show that if, on the contrary, strong laws of nature are logically possible, then it can be shown that certain inductive inferences are justified.

10. Reductionist Approaches to Laws of Nature and Inductive Skepticism

        The results that I have in mind depend upon whether, in formulating inductive logic, one assumes that all state descriptions are equally probable, or whether, as Carnap thought, all structure descriptions are equally probable.  I shall not, at this point, offer a technical explanation of those two technical notions, since I think that the mini-world example offered earlier should suffice for present purposes.

        Let me, then, simply state some results.  In doing so, I shall often refer to a concrete case of the relevant theorem, rather than formulating it in a general and very abstract way.  Finally, all of these results are predicated on the assumption that strong laws of nature are not logically possible.

        The first two results are based on the assumption that the correct equiprobability assumption on which to base one’s inductive logic is that it is state descriptions that are equally likely.  Given that assumption, one has the following two results

Result 1

        Suppose, for concreteness, that there is an urn that contains a million marbles, each of which is either red or green.  Given no information at all, what is the probability that the millionth marble drawn from the urn is red?  The answer is  .  Suppose, now, that 999,999 marbles have been drawn from the urn, and that all of them are red. What is the probability, given that information, that the millionth marble drawn from the urn is red?  The answer is still  .

Conclusion: If strong laws of nature are logically impossible, and all state descriptions are equally probable, then one cannot learn from experience.

Result 2

    Suppose, first, that, through the total history of the world, there are an infinite number of things that are F.  Suppose, further, that a billion things that are F have been observed, and that all of them were G.  What is the probability that that all Fs are Gs?  The answer is that that probability is either equal to zero, or, if one accepts infinitesimals, infinitesimally close to zero.

Suppose that, somewhat disheartened by those two theorems, a reductionist with regard to laws of nature follows Carnap’s lead, and defines logical probability based on the proposition that it is structure descriptions, not state descriptions, which are equally likely.  Then one has the following theorem:

Result 3

If property G does not belong to a family of positive, incompatible properties, then given only the information there are n Fs, and that all of them have property G, the probability that the next F will also have property G is equal to  (n + 1)/(n + 2).

(This is Laplace’s famous Rule of Succession.)

        Now this is a cheering result.  The more Fs one observes, all of which have property G, the more likely it is that the next F has property G.  So one can learn from experience.

        But one also has the following theorem:

Result 4

        Suppose, once again, that, through the total history of the world, there are an infinite number of things that are F.  Suppose, further, that a billion things that are F have been observed, and that all of them were G.  What is the probability that that all Fs are Gs?  The answer, once again, is that that probability is either equal to zero, or, if one accepts infinitesimals, infinitesimally close to zero.

    Results 2 and 4 look depressing if one holds that strong laws of nature are logically impossible.  But Hans Reichenbach offered an interesting argument for the following theorem:

Result 5:  If probabilistic laws of nature are logically possible, then no evidence can ever make it likely that a non-probabilistic law obtains.

        If Reichenbach is right, then one can never confirm any non-probabilistic law, and so Results 2 and 4 need not trouble the reductionist with regard to laws.

        Happiness for the reductionist, however, is short-lived.  For, first of all, if one returns to the idea of defining logical probabilities based on the proposition that all state descriptions are equally likely, one can then prove the following theorem:

Result 6

        Suppose, for concreteness, 1000 Fs have been examined, and all 1000 have turned out to be Gs.  What is the probability that, if 1000 more Fs are examined, 90% of the combined set of 2000 Fs will be Gs?  The answer is that it is 2.04944 x 10-86.

        This is a rather small number.  What it illustrates is that the combination of a reductionist approach to laws of nature with a state description approach to logical probability is not going to allow one to be able confirm the existence of some law or other to the effect that the probability that an F is a G is equal to k, where k falls in some moderate interval in the vicinity of the number one.

    Suppose, finally, that one shifts, once again, from a formulation of logical probability that treats state descriptions as equally likely to a formulation that treats structure descriptions as equally likely.  Does that save the reductionist?  The answer is that it does not, since one can prove the following theorems:

Result 7

        Suppose that 1000 Fs have been examined, and all of them have turned out to be Gs.  What is the probability that, if 1000 more Fs are examined, 90% of the combined total of 2000 Fs will be Gs?  The answer is that the probability is just over 20%.

Result 8

        Suppose, finally, that one billion Fs have been examined, and all of them have turned out to be Gs.  Suppose, further, that in the total history of the universe, there are an infinite number of Fs.  What is the probability that 90% of all the Fs will be Gs?  The answer is that the probability is equal to 10%.

        The moral, I suggest, seems clear: if one embraces a reductionist approach to laws of nature, then regardless of whether one adopts a state description approach to inductive logic or a structure description approach, one will not be able to avoid the following conclusion:  No interesting scientific hypothesis concerning laws of nature can be confirmed.

11.  Families of Properties and the Epistemology of Strong Laws of Nature

        So how are things epistemologically if strong laws of nature are logically possible?  The answer to that question depends upon the idea that the most basic equiprobability principle is one that is formulated, not in terms of either state descriptions or structure descriptions, but in terms of families of properties.  One way of formulating such a principle is as follows:

Equiprobability and Families of Properties

        Given any family of incompatible properties, if P and Q are any two members of such a family, then the a priori probability that a has property P is equal to the a priori probability that a has property Q.

    Given such a principle, if strong laws of nature are logically possible, and if they can take the form of certain atomic states of affairs consisting of irreducible second-order relations among universals, then certain sets of such second-order relations will be families of relations.  Accordingly, one can apply the equiprobability principle just stated to such families of nomic relations, thereby generating equiprobability conclusions concerning laws of nature.

        Given this starting point, one can then work out the probability that a strong law of nature, falling within a certain range, does obtain, given information about events that would fall under such a law if it did exist.  Doing this involves a somewhat complicated calculation, especially because one has to take into account the possibility of probabilistic laws connecting being F with being G.  But I can set out a table that will make it evident that if one accepts the idea of strong laws of nature, the epistemology of laws of nature is in reasonably good shape.

        First of all, however, I need to explain some notation:

“ aL1“ means “There is some number k such that it is a law that the probability that something that has property F has property G is equal to k, where k lies in the range from a to 1.”

“Qn “means “n particular things that have property F all have property G.”

“M2 “means “G does not belong to a family of positive properties, so that the only possibilities are either having property G, or not having it.”

“Pr(q/p) = m “ means “The logical probability that q is the case given that p is the case is equal to m.”

So “Pr(aL1/Qn & M2) = m ” means “The probability that there is some number k such that it is a law that the probability that something that has property F has property G is equal to k, where k lies in the range from a to 1, given that n particular things that have property F all have property G, and that G does not belong to a family of positive properties is equal to m.”

Result 9

If a = 0.99, then the value of  Pr(aL1/Qn & M2) is given by the following table:

Value of n            

     1        0.000199   
   10        0.003140791   
 100        0.159292545
 200        0.404718644   
 300        0.60791531   
 400        0.750963679
 500        0.844923366
1000        0.986867122

    What this table shows is that if it is possible for there to be strong laws of nature, then the probability that it is either a deterministic law that all Fs are Gs or else that there is a probabilistic law to the effect that the probability that something that is F is also G is equal to k, where k lies in the range from 0.99 to 1, can be raised to quite high values by a relatively small number of instances.

Summing Up

        A very promising approach to the solution of the problem of justifying induction involves making use of the idea of logical probability.  But the prospects for such an approach depend crucially upon the metaphysics of laws of nature.  The reason, as we have just seen, is that, on the one hand, there are theorems that provide excellent reason for thinking that if strong, governing laws of nature are not logically possible, then no laws of nature can ever be confirmed, while, on the other hand, there are other theorems that provide excellent reason for thinking that that if strong laws of nature are logically possible, then the existence of such laws can be confirmed, and thus that induction can be justified.