Jump to content

Occam's razor

From Wikipedia, the free encyclopedia
(Redirected fromParsimonious)

Inphilosophy,Occam's razor(also spelledOckham's razororOcham's razor;Latin:novacula Occami) is the problem-solving principle that recommends searching for explanations constructed with the smallest possible set of elements. It is also known as theprinciple of parsimonyor thelaw of parsimony(Latin:lex parsimoniae). Attributed toWilliam of Ockham,a 14th-century Englishphilosopherandtheologian,it is frequently cited asEntia non sunt multiplicanda praeter necessitatem,which translates as "Entities must not be multiplied beyond necessity",[1][2]although Occam never used these exact words. Popularly, the principle is sometimes paraphrased as "The simplest explanation is usually the best one."[3]

Thisphilosophical razoradvocates that when presented with competinghypothesesabout the same prediction and both hypotheses have equal explanatory power, one should prefer the hypothesis that requires the fewest assumptions,[4]and that this is not meant to be a way of choosing between hypotheses that make different predictions. Similarly, in science, Occam's razor is used as anabductiveheuristicin the development of theoretical models rather than as a rigorous arbiter between candidate models.[5][6]

History

[edit]

The phraseOccam's razordid not appear until a few centuries after William of Ockham's death in 1347.Libert Froidmont,in hisOn Christian Philosophy of the Soul,gives him credit for the phrase, speaking of "novacula occami".[7]Ockham did not invent this principle, but its fame—and its association with him—may be due to the frequency and effectiveness with which he used it.[8]Ockham stated the principle in various ways, but the most popular version, "Entities are not to be multiplied without necessity" (Non sunt multiplicanda entia sine necessitate) was formulated by the IrishFranciscanphilosopherJohn Punchin his 1639 commentary on the works ofDuns Scotus.[9]

Formulations before William of Ockham

[edit]
Part of a page fromJohn Duns Scotus's bookCommentaria oxoniensia ad IV libros magistri Sententiarus,showing the words: "Pluralitas non est ponenda sine necessitate",i.e.," Plurality is not to be posited without necessity "

The origins of what has come to be known as Occam's razor are traceable to the works of earlier philosophers such asJohn Duns Scotus(1265–1308),Robert Grosseteste(1175–1253),Maimonides(Moses ben-Maimon, 1138–1204), and evenAristotle(384–322 BC).[10][11]Aristotle writes in hisPosterior Analytics,"We may assume the superiorityceteris paribus[other things being equal] of the demonstration which derives from fewer postulates or hypotheses. "Ptolemy(c. AD 90– c. 168) stated, "We consider it a good principle to explain the phenomena by the simplest hypothesis possible."[12]

Phrases such as "It is vain to do with more what can be done with fewer" and "A plurality is not to be posited without necessity" were commonplace in 13th-centuryscholasticwriting.[12]Robert Grosseteste, inCommentary on[Aristotle's]the Posterior Analytics Books(Commentarius in Posteriorum Analyticorum Libros) (c. 1217–1220), declares: "That is better and more valuable which requires fewer, other circumstances being equal... For if one thing were demonstrated from many and another thing from fewer equally known premises, clearly that is better which is from fewer because it makes us know quickly, just as a universal demonstration is better than particular because it produces knowledge from fewer premises. Similarly in natural science, in moral science, and in metaphysics the best is that which needs no premises and the better that which needs the fewer, other circumstances being equal."[13]

TheSumma TheologicaofThomas Aquinas(1225–1274) states that "it is superfluous to suppose that what can be accounted for by a few principles has been produced by many." Aquinas uses this principle to construct an objection toGod's existence,an objection that he in turn answers and refutes generally (cf.quinque viae), and specifically, through an argument based oncausality.[14]Hence, Aquinas acknowledges the principle that today is known as Occam's razor, but prefers causal explanations to other simple explanations (cf. alsoCorrelation does not imply causation).

William of Ockham

[edit]
Manuscriptillustration of William of Ockham

William of Ockham(circa1287–1347) was an English Franciscan friar andtheologian,an influential medieval philosopher and anominalist.His popular fame as a great logician rests chiefly on the maxim attributed to him and known as Occam's razor. The termrazorrefers to distinguishing between two hypotheses either by "shaving away" unnecessary assumptions or cutting apart two similar conclusions.

While it has been claimed that Occam's razor is not found in any of William's writings,[15]one can cite statements such asNumquam ponenda est pluralitas sine necessitate( "Plurality must never be posited without necessity" ), which occurs in his theological work on theSentences of Peter Lombard(Quaestiones et decisiones in quattuor libros Sententiarum Petri Lombardi;ed. Lugd., 1495, i, dist. 27, qu. 2, K).

Nevertheless, the precise words sometimes attributed to William of Ockham,Entia non sunt multiplicanda praeter necessitatem(Entities must not be multiplied beyond necessity),[16]are absent in his extant works;[17]this particular phrasing comes fromJohn Punch,[18]who described the principle as a "common axiom" (axioma vulgare) of the Scholastics.[9]William of Ockham himself seems to restrict the operation of this principle in matters pertaining to miracles and God's power, considering a plurality of miracles possible in theEucharist[further explanation needed]simply because it pleases God.[12]

This principle is sometimes phrased asPluralitas non est ponenda sine necessitate( "Plurality should not be posited without necessity" ).[19]In hisSumma Totius Logicae,i. 12, William of Ockham cites the principle of economy,Frustra fit per plura quod potest fieri per pauciora( "It is futile to do with more things that which can be done with fewer"; Thorburn, 1918, pp. 352–53;Knealeand Kneale, 1962, p. 243.)

Later formulations

[edit]

To quoteIsaac Newton,"We are to admit no more causes of natural things than such as are both true and sufficient to explain their appearances. Therefore, to the same natural effects we must, as far as possible, assign the same causes."[20][21] In the sentencehypotheses non fingo,Newton affirms the success of this approach.

Bertrand Russelloffers a particular version of Occam's razor: "Whenever possible, substitute constructions out of known entities for inferences to unknown entities."[22]

Around 1960,Ray Solomonofffounded thetheory of universal inductive inference,the theory of prediction based on observations – for example, predicting the next symbol based upon a given series of symbols. The only assumption is that the environment follows some unknown but computable probability distribution. This theory is a mathematical formalization of Occam's razor.[23][24][25]

Another technical approach to Occam's razor isontological parsimony.[26]Parsimony means spareness and is also referred to as the Rule of Simplicity. This is considered a strong version of Occam's razor.[27][28]A variation used in medicine is called the "Zebra":a physician should reject an exotic medical diagnosis when a more commonplace explanation is more likely, derived fromTheodore Woodward's dictum "When you hear hoofbeats, think of horses not zebras".[29]

Ernst Machformulated the stronger version of Occam's razor intophysics,which he called the Principle of Economy stating: "Scientists must use the simplest means of arriving at their results and exclude everything not perceived by the senses."[30]

This principle goes back at least as far as Aristotle, who wrote "Nature operates in the shortest way possible."[27]The idea of parsimony or simplicity in deciding between theories, though not the intent of the original expression of Occam's razor, has been assimilated into common culture as the widespread layman's formulation that "the simplest explanation is usually the correct one."[27]

Justifications

[edit]

Aesthetic

[edit]

Prior to the 20th century, it was a commonly held belief that nature itself was simple and that simpler hypotheses about nature were thus more likely to be true.This notion was deeply rooted in the aesthetic value that simplicity holds for human thought and the justifications presented for it often drew fromtheology.[clarification needed]Thomas Aquinasmade this argument in the 13th century, writing, "If a thing can be done adequately by means of one, it is superfluous to do it by means of several; for we observe that nature does not employ two instruments [if] one suffices."[31]

Beginning in the 20th century,epistemologicaljustifications based oninduction,logic,pragmatism,and especiallyprobability theoryhave become more popular among philosophers.[7]

Empirical

[edit]

Occam's razor has gained strong empirical support in helping to converge on better theories (seeUsessection below for some examples).

In the related concept ofoverfitting,excessively complex models are affected bystatistical noise(a problem also known as thebias–variance tradeoff), whereas simpler models may capture the underlying structure better and may thus have betterpredictiveperformance. It is, however, often difficult to deduce which part of the data is noise (cf.model selection,test set,minimum description length,Bayesian inference,etc.).

Testing the razor

[edit]

The razor's statement that "other things being equal, simpler explanations are generally better than more complex ones" is amenable to empirical testing. Another interpretation of the razor's statement would be that "simpler hypotheses are generally better than the complex ones". The procedure to test the former interpretation would compare the track records of simple and comparatively complex explanations. If one accepts the first interpretation, the validity of Occam's razor as a tool would then have to be rejected if the more complex explanations were more often correct than the less complex ones (while the converse would lend support to its use). If the latter interpretation is accepted, the validity of Occam's razor as a tool could possibly be accepted if the simpler hypotheses led to correct conclusions more often than not.

Even if some increases in complexity are sometimes necessary, there still remains a justified general bias toward the simpler of two competing explanations. To understand why, consider that for each accepted explanation of a phenomenon, there is always an infinite number of possible, more complex, and ultimately incorrect, alternatives. This is so because one can always burden a failing explanation with anad hoc hypothesis.Ad hoc hypotheses are justifications that prevent theories from being falsified.

Possible explanations can become needlessly complex. It might be coherent, for instance, to add the involvement ofleprechaunsto any explanation, but Occam's razor would prevent such additions unless they were necessary.

For example, if a man, accused of breaking a vase, makessupernaturalclaims thatleprechaunswere responsible for the breakage, a simple explanation might be that the man did it, but ongoing ad hoc justifications (e.g., "... and that's not me breaking it on the film; they tampered with that, too" ) could successfully prevent complete disproof. This endless supply of elaborate competing explanations, called saving hypotheses, cannot be technically ruled out – except by using Occam's razor.[32][33][34]

Any more complex theory might still possibly be true. A study of the predictive validity of Occam's razor found 32 published papers that included 97 comparisons of economic forecasts from simple and complex forecasting methods. None of the papers provided a balance of evidence that complexity of method improved forecast accuracy. In the 25 papers with quantitative comparisons, complexity increased forecast errors by an average of 27 percent.[35]

Practical considerations and pragmatism

[edit]

Mathematical

[edit]

One justification of Occam's razor is a direct result of basicprobability theory.By definition, all assumptions introduce possibilities for error; if an assumption does not improve the accuracy of a theory, its only effect is to increase the probability that the overall theory is wrong.

There have also been other attempts to derive Occam's razor from probability theory, including notable attempts made byHarold JeffreysandE. T. Jaynes.The probabilistic (Bayesian) basis for Occam's razor is elaborated byDavid J. C. MacKayin chapter 28 of his bookInformation Theory, Inference, and Learning Algorithms,[36]where he emphasizes that a prior bias in favor of simpler models is not required.

William H. JefferysandJames O. Berger(1991) generalize and quantify the original formulation's "assumptions" concept as the degree to which a proposition is unnecessarily accommodating to possible observable data.[37]They state, "A hypothesis with fewer adjustable parameters will automatically have an enhanced posterior probability, due to the fact that the predictions it makes are sharp."[37]The use of "sharp" here is not only a tongue-in-cheek reference to the idea of a razor, but also indicates that such predictions are moreaccuratethan competing predictions. The model they propose balances the precision of a theory's predictions against their sharpness, preferring theories that sharply make correct predictions over theories that accommodate a wide range of other possible results. This, again, reflects the mathematical relationship between key concepts inBayesian inference(namelymarginal probability,conditional probability,andposterior probability).

Thebias–variance tradeoffis a framework that incorporates the Occam's razor principle in its balance between overfitting (associated with lower bias but higher variance) and underfitting (associated with lower variance but higher bias).[38]

Other philosophers

[edit]

Karl Popper

[edit]

Karl Popperargues that a preference for simple theories need not appeal to practical or aesthetic considerations. Our preference for simplicity may be justified by itsfalsifiabilitycriterion: we prefer simpler theories to more complex ones "because their empirical content is greater; and because they are better testable".[39]The idea here is that a simple theory applies to more cases than a more complex one, and is thus more easily falsifiable. This is again comparing a simple theory to a more complex theory where both explain the data equally well.

Elliott Sober

[edit]

The philosopher of scienceElliott Soberonce argued along the same lines as Popper, tying simplicity with "informativeness": The simplest theory is the more informative, in the sense that it requires less information to a question.[40]He has since rejected this account of simplicity, purportedly because it fails to provide anepistemicjustification for simplicity. He now believes that simplicity considerations (and considerations of parsimony in particular) do not count unless they reflect something more fundamental. Philosophers, he suggests, may have made the error of hypostatizing simplicity (i.e., endowed it with asui generisexistence), when it has meaning only when embedded in a specific context (Sober 1992). If we fail to justify simplicity considerations on the basis of the context in which we use them, we may have no non-circular justification: "Just as the question 'why be rational?' may have no non-circular answer, the same may be true of the question 'why should simplicity be considered in evaluating the plausibility of hypotheses?'"[41]

Richard Swinburne

[edit]

Richard Swinburneargues for simplicity on logical grounds:

... the simplest hypothesis proposed as an explanation of phenomena is more likely to be the true one than is any other available hypothesis, that its predictions are more likely to be true than those of any other available hypothesis, and that it is an ultimatea prioriepistemic principle that simplicity is evidence for truth.

— Swinburne 1997

According to Swinburne, since our choice of theory cannot be determined by data (seeUnderdeterminationandDuhem–Quine thesis), we must rely on some criterion to determine which theory to use. Since it is absurd to have no logical method for settling on one hypothesis amongst an infinite number of equally data-compliant hypotheses, we should choose the simplest theory: "Either science is irrational [in the way it judges theories and predictions probable] or the principle of simplicity is a fundamental synthetic a priori truth."[42]

Ludwig Wittgenstein

[edit]

From theTractatus Logico-Philosophicus:

  • 3.328 "If a sign is not necessary then it is meaningless. That is the meaning of Occam's Razor."
(If everything in the symbolism works as though a sign had meaning, then it has meaning.)
  • 4.04 "In the proposition, there must be exactly as many things distinguishable as there are in the state of affairs, which it represents. They must both possess the same logical (mathematical) multiplicity (cf. Hertz's Mechanics, on Dynamic Models)."
  • 5.47321 "Occam's Razor is, of course, not an arbitrary rule nor one justified by its practical success. It simply says that unnecessary elements in a symbolism mean nothing. Signs which serve one purpose are logically equivalent; signs which serve no purpose are logically meaningless."

and on the related concept of "simplicity":

  • 6.363 "The procedure of induction consists in accepting as true the simplest law that can be reconciled with our experiences."

Uses

[edit]

Science and the scientific method

[edit]
Andreas Cellarius's illustration of the Copernican system, from theHarmonia Macrocosmica(1660). Future positions of the sun, moon and other solar system bodies can be calculated using ageocentricmodel (the earth is at the centre) or using aheliocentric model(the sun is at the centre). Both work, but the geocentric model arrives at the same conclusions through a much more complex system of calculations than the heliocentric model. This was pointed out in a preface toCopernicus' first edition ofDe revolutionibus orbium coelestium.

Inscience,Occam's razor is used as aheuristicto guide scientists in developing theoretical models rather than as an arbiter between published models.[5][6]Inphysics,parsimony was an important heuristic inAlbert Einstein's formulation ofspecial relativity,[43][44]in the development and application of theprinciple of least actionbyPierre Louis MaupertuisandLeonhard Euler,[45]and in the development ofquantum mechanicsbyMax Planck,Werner HeisenbergandLouis de Broglie.[6][46]

Inchemistry,Occam's razor is often an important heuristic when developing a model of areaction mechanism.[47][48]Although it is useful as a heuristic in developing models of reaction mechanisms, it has been shown to fail as a criterion for selecting among some selected published models.[6]In this context, Einstein himself expressed caution when he formulated Einstein'sConstraint:"It can scarcely be denied that the supreme goal of all theory is to make the irreducible basic elements as simple and as few as possible without having to surrender the adequate representation of a single datum of experience."[49][50][51]An often-quoted version of this constraint (which cannot be verified as posited by Einstein himself)[52]reduces this to "Everything should be kept as simple as possible, but not simpler."

In thescientific method,Occam's razor is not considered an irrefutable principle oflogicor a scientific result; the preference for simplicity in the scientific method is based on thefalsifiabilitycriterion. For each accepted explanation of a phenomenon, there may be an extremely large, perhaps even incomprehensible, number of possible and more complex alternatives. Since failing explanations can always be burdened withad hochypothesesto prevent them from being falsified, simpler theories are preferable to more complex ones because they tend to be moretestable.[53][54][55]As a logical principle, Occam's razor would demand that scientists accept the simplest possible theoretical explanation for existing data. However, science has shown repeatedly that future data often support more complex theories than do existing data. Science prefers the simplest explanation that is consistent with the data available at a given time, but the simplest explanation may be ruled out as new data become available.[5][54]That is, science is open to the possibility that future experiments might support more complex theories than demanded by current data and is more interested in designing experiments to discriminate between competing theories than favoring one theory over another based merely on philosophical principles.[53][54][55]

When scientists use the idea of parsimony, it has meaning only in a very specific context of inquiry. Several background assumptions are required for parsimony to connect with plausibility in a particular research problem.[clarification needed]The reasonableness of parsimony in one research context may have nothing to do with its reasonableness in another. It is a mistake to think that there is a single global principle that spans diverse subject matter.[55]

It has been suggested that Occam's razor is a widely accepted example of extraevidential consideration, even though it is entirely a metaphysical assumption. Most of the time, however, Occam's razor is a conservative tool, cutting out "crazy, complicated constructions" and assuring "that hypotheses are grounded in the science of the day", thus yielding "normal" science: models of explanation and prediction.[6]There are, however, notable exceptions where Occam's razor turns a conservative scientist into a reluctant revolutionary. For example,Max Planckinterpolated between theWienandJeansradiation laws and used Occam's razor logic to formulate the quantum hypothesis, even resisting that hypothesis as it became more obvious that it was correct.[6]

Appeals to simplicity were used to argue against the phenomena of meteorites,ball lightning,continental drift,andreverse transcriptase.[56]One can argue for atomic building blocks for matter, because it provides a simpler explanation for the observed reversibility of bothmixing[clarification needed]and chemical reactions as simple separation and rearrangements of atomic building blocks. At the time, however, theatomic theorywas considered more complex because it implied the existence of invisible particles that had not been directly detected.Ernst Machand the logical positivists rejectedJohn Dalton'satomic theoryuntil the reality of atoms was more evident inBrownian motion,as shown byAlbert Einstein.[57]

In the same way, postulating theaetheris more complex than transmission of light through avacuum.At the time, however, all known waves propagated through a physical medium, and it seemed simpler to postulate the existence of a medium than to theorize about wave propagation without a medium. Likewise,Isaac Newton's idea of light particles seemed simpler thanChristiaan Huygens's idea of waves, so many favored it. In this case, as it turned out, neither the wave—nor the particle—explanation alone suffices, aslight behaves like waves and like particles.

Three axioms presupposed by the scientific method are realism (the existence of objective reality), the existence of natural laws, and the constancy of natural law. Rather than depend on provability of these axioms, science depends on the fact that they have not been objectively falsified. Occam's razor and parsimony support, but do not prove, these axioms of science. The general principle of science is that theories (or models) of natural law must be consistent with repeatable experimental observations. This ultimate arbiter (selection criterion) rests upon the axioms mentioned above.[54]

If multiple models of natural law make exactly the same testable predictions, they are equivalent and there is no need for parsimony to choose a preferred one. For example,Newtonian,HamiltonianandLagrangianclassical mechanics are equivalent. Physicists have no interest in using Occam's razor to say the other two are wrong. Likewise, there is no demand for simplicity principles to arbitrate between wave and matrix formulations of quantum mechanics. Science often does not demand arbitration or selection criteria between models that make the same testable predictions.[54]

Biology

[edit]

Biologists or philosophers of biology use Occam's razor in either of two contexts both inevolutionary biology:the units of selection controversy andsystematics.George C. Williamsin his bookAdaptation and Natural Selection(1966) argues that the best way to explainaltruismamong animals is based on low-level (i.e., individual) selection as opposed to high-level group selection. Altruism is defined by some evolutionary biologists (e.g., R. Alexander, 1987; W. D. Hamilton, 1964) as behavior that is beneficial to others (or to the group) at a cost to the individual, and many posit individual selection as the mechanism that explains altruism solely in terms of the behaviors of individual organisms acting in their own self-interest (or in the interest of their genes, via kin selection). Williams was arguing against the perspective of others who propose selection at the level of the group as an evolutionary mechanism that selects for altruistic traits (e.g., D. S. Wilson & E. O. Wilson, 2007). The basis for Williams' contention is that of the two, individual selection is the more parsimonious theory. In doing so he is invoking a variant of Occam's razor known asMorgan's Canon:"In no case is an animal activity to be interpreted in terms of higher psychological processes, if it can be fairly interpreted in terms of processes which stand lower in the scale of psychological evolution and development." (Morgan 1903).

However, more recent biological analyses, such asRichard Dawkins'The Selfish Gene,have contended that Morgan's Canon is not the simplest and most basic explanation. Dawkins argues the way evolution works is that the genes propagated in most copies end up determining the development of that particular species, i.e., natural selection turns out to select specific genes, and this is really the fundamental underlying principle that automatically gives individual and group selection asemergentfeatures of evolution.

Zoologyprovides an example.Muskoxen,when threatened bywolves,form a circle with the males on the outside and the females and young on the inside. This is an example of a behavior by the males that seems to be altruistic. The behavior is disadvantageous to them individually but beneficial to the group as a whole and was thus seen by some to support the group selection theory. Another interpretation is kin selection: if the males are protecting their offspring, they are protecting copies of their own alleles. Engaging in this behavior would be favored by individual selection if the cost to the male musk ox is less than half of the benefit received by his calf – which could easily be the case if wolves have an easier time killing calves than adult males. It could also be the case that male musk oxen would be individually less likely to be killed by wolves if they stood in a circle with their horns pointing out, regardless of whether they were protecting the females and offspring. That would be an example of regular natural selection – a phenomenon called "the selfish herd".

Systematicsis the branch ofbiologythat attempts to establish patterns of relationship among biological taxa, today generally thought to reflect evolutionary history. It is also concerned with their classification. There are three primary camps in systematics: cladists, pheneticists, and evolutionary taxonomists. Cladists hold that classification should be based onsynapomorphies(shared, derived character states), pheneticists contend that overall similarity (synapomorphies and complementarysymplesiomorphies) is the determining criterion, while evolutionary taxonomists say that both genealogy and similarity count in classification (in a manner determined by the evolutionary taxonomist).[58][59]

It is among the cladists that Occam's razor is applied, through the method ofcladistic parsimony.Cladistic parsimony (ormaximum parsimony) is a method of phylogenetic inference that yieldsphylogenetic trees(more specifically, cladograms).Cladogramsare branching, diagrams used to represent hypotheses of relative degree of relationship, based onsynapomorphies.Cladistic parsimony is used to select as the preferred hypothesis of relationships the cladogram that requires the fewest implied character state transformations (or smallest weight, if characters are differentially weighted). Critics of the cladistic approach often observe that for some types of data, parsimony could produce the wrong results, regardless of how much data is collected (this is called statistical inconsistency, orlong branch attraction). However, this criticism is also potentially true for any type of phylogenetic inference, unless the model used to estimate the tree reflects the way that evolution actually happened. Because this information is not empirically accessible, the criticism of statistical inconsistency against parsimony holds no force.[60]For a book-length treatment of cladistic parsimony, seeElliott Sober'sReconstructing the Past: Parsimony, Evolution, and Inference(1988). For a discussion of both uses of Occam's razor in biology, see Sober's article "Let's Razor Ockham's Razor" (1990).

Other methods for inferring evolutionary relationships use parsimony in a more general way.Likelihoodmethods for phylogeny use parsimony as they do for all likelihood tests, with hypotheses requiring fewer differing parameters (i.e., numbers or different rates of character change or different frequencies of character state transitions) being treated as null hypotheses relative to hypotheses requiring more differing parameters. Thus, complex hypotheses must predict data much better than do simple hypotheses before researchers reject the simple hypotheses. Recent advances employinformation theory,a close cousin of likelihood, which uses Occam's razor in the same way. The choice of the "shortest tree" relative to a not-so-short tree under any optimality criterion (smallest distance, fewest steps, or maximum likelihood) is always based on parsimony.[61]

Francis Crickhas commented on potential limitations of Occam's razor in biology. He advances the argument that because biological systems are the products of (an ongoing) natural selection, the mechanisms are not necessarily optimal in an obvious sense. He cautions: "While Ockham's razor is a useful tool in the physical sciences, it can be a very dangerous implement in biology. It is thus very rash to use simplicity and elegance as a guide in biological research."[62]This is an ontological critique of parsimony.

Inbiogeography,parsimony is used to infer ancient vicariant events ormigrationsofspeciesorpopulationsby observing the geographic distribution and relationships of existingorganisms.Given the phylogenetic tree, ancestral population subdivisions are inferred to be those that require the minimum amount of change.[citation needed]

Religion

[edit]

In thephilosophy of religion,Occam's razor is sometimes applied to the existence of God. William of Ockham himself was aChristian.He believed in God, and in theauthorityofChristian scripture;he writes that "nothing ought to be posited without a reason given, unless it is self-evident (literally, known through itself) or known by experience or proved by the authority of Sacred Scripture."[63]Ockham believed that an explanation has no sufficient basis in reality when it does not harmonize with reason, experience, or theBible.Unlike many theologians of his time, though, Ockham did not believe God could be logically proven with arguments. To Ockham, science was a matter of discovery;theologywas a matter ofrevelationandfaith.He states: "Only faith gives us access to theological truths. The ways of God are not open to reason, for God has freely chosen to create a world and establish a way of salvation within it apart from any necessary laws that human logic or rationality can uncover."[64]

Thomas Aquinas,in theSumma Theologica,uses a formulation of Occam's razor to construct an objection to the idea that God exists, which he refutes directly with a counterargument:[65]

Further, it is superfluous to suppose that what can be accounted for by a few principles has been produced by many. But it seems that everything we see in the world can be accounted for by other principles, supposing God did not exist. For all natural things can be reduced to one principle which is nature; and all voluntary things can be reduced to one principle which is human reason, or will. Therefore there is no need to suppose God's existence.

In turn, Aquinas answers this with thequinque viae,and addresses the particular objection above with the following answer:

Since nature works for a determinate end under the direction of a higher agent, whatever is done by nature must needs be traced back to God, as to its first cause. So also whatever is done voluntarily must also be traced back to some higher cause other than human reason or will, since these can change or fail; for all things that are changeable and capable of defect must be traced back to an immovable and self-necessary first principle, as was shown in the body of the Article.

Rather than argue for the necessity of a god, sometheistsbase their belief upon grounds independent of, or prior to, reason, making Occam's razor irrelevant. This was the stance ofSøren Kierkegaard,who viewed belief in God as aleap of faiththat sometimes directly opposed reason.[66]This is also the doctrine ofGordon Clark'spresuppositional apologetics,with the exception that Clark never thought the leap of faith was contrary to reason (see alsoFideism).

Variousarguments in favor of Godestablish God as a useful or even necessary assumption. Contrastingly some anti-theists hold firmly to the belief that assuming the existence of God introduces unnecessary complexity (e.g., theUltimate Boeing 747 gambitfrom Dawkins'sThe God Delusion[67]).[68]

Another application of the principle is to be found in the work ofGeorge Berkeley(1685–1753). Berkeley was an idealist who believed that all of reality could be explained in terms of the mind alone. He invoked Occam's razor againstmaterialism,stating that matter was not required by his metaphysics and was thus eliminable. One potential problem with this belief[for whom?]is that it's possible, given Berkeley's position, to findsolipsismitself more in line with the razor than a God-mediated world beyond a single thinker.

Occam's razor may also be recognized in the apocryphal story about an exchange betweenPierre-Simon LaplaceandNapoleon.It is said that in praising Laplace for one of his recent publications, the emperor asked how it was that the name of God, which featured so frequently in the writings ofLagrange,appeared nowhere in Laplace's. At that, he is said to have replied, "It's because I had no need of that hypothesis."[69]Though some points of this story illustrate Laplace'satheism,more careful consideration suggests that he may instead have intended merely to illustrate the power ofmethodological naturalism,or even simply that the fewerlogical premisesone assumes, thestrongeris one's conclusion.

Philosophy of mind

[edit]

In his article "Sensations and Brain Processes" (1959),J. J. C. Smartinvoked Occam's razor with the aim to justify his preference of themind-brain identity theoryoverspirit-body dualism.Dualists state that there are two kinds of substances in the universe: physical (including the body) and spiritual, which is non-physical. In contrast, identity theorists state that everything is physical, including consciousness, and that there is nothing nonphysical. Though it is impossible to appreciate the spiritual when limiting oneself to the physical,[citation needed]Smart maintained that identity theory explains all phenomena by assuming only a physical reality. Subsequently, Smart has been severely criticized for his use (or misuse) of Occam's razor and ultimately retracted his advocacy of it in this context.Paul Churchland(1984) states that by itself Occam's razor is inconclusive regarding duality. In a similar way, Dale Jacquette (1994) stated that Occam's razor has been used in attempts to justify eliminativism and reductionism in the philosophy of mind. Eliminativism is the thesis that the ontology offolk psychologyincluding such entities as "pain", "joy", "desire", "fear", etc., are eliminable in favor of an ontology of a completed neuroscience.

Penal ethics

[edit]

In penal theory and the philosophy of punishment, parsimony refers specifically to taking care in the distribution ofpunishmentin order to avoid excessive punishment. In theutilitarianapproach to the philosophy of punishment,Jeremy Bentham's "parsimony principle" states that any punishment greater than is required to achieve its end is unjust. The concept is related but not identical to the legal concept ofproportionality.Parsimony is a key consideration of the modernrestorative justice,and is a component of utilitarian approaches to punishment, as well as theprison abolition movement.Bentham believed that true parsimony would require punishment to be individualised to take account of thesensibilityof the individual—an individual more sensitive to punishment should be given a proportionately lesser one, since otherwise needless pain would be inflicted. Later utilitarian writers have tended to abandon this idea, in large part due to the impracticality of determining each alleged criminal's relative sensitivity to specific punishments.[70]

Probability theory and statistics

[edit]

Marcus Hutter's universal artificial intelligence builds uponSolomonoff's mathematical formalization of the razorto calculate the expected value of an action.

There are various papers in scholarly journals deriving formal versions of Occam's razor from probability theory, applying it instatistical inference,and using it to come up with criteria for penalizing complexity in statistical inference. Papers[71][72]have suggested a connection between Occam's razor andKolmogorov complexity.[73]

One of the problems with the original formulation of the razor is that it only applies to models with the same explanatory power (i.e., it only tells us to prefer the simplest of equally good models). A more general form of the razor can be derived from Bayesian model comparison, which is based onBayes factorsand can be used to compare models that do not fit the observations equally well. These methods can sometimes optimally balance the complexity and power of a model. Generally, the exact Occam factor is intractable, but approximations such asAkaike information criterion,Bayesian information criterion,Variational Bayesian methods,false discovery rate,andLaplace's methodare used. Manyartificial intelligenceresearchers are now employing such techniques, for instance through work onOccam Learningor more generally on theFree energy principle.

Statistical versions of Occam's razor have a more rigorous formulation than what philosophical discussions produce. In particular, they must have a specific definition of the termsimplicity,and that definition can vary. For example, in theKolmogorovChaitinminimum description lengthapproach, the subject must pick aTuring machinewhose operations describe the basic operationsbelievedto represent "simplicity" by the subject. However, one could always choose a Turing machine with a simple operation that happened to construct one's entire theory and would hence score highly under the razor. This has led to two opposing camps: one that believes Occam's razor is objective, and one that believes it is subjective.

Objective razor

[edit]

The minimum instruction set of auniversal Turing machinerequires approximately the same length description across different formulations, and is small compared to theKolmogorov complexityof most practical theories.Marcus Hutterhas used this consistency to define a "natural" Turing machine of small size as the proper basis for excluding arbitrarily complex instruction sets in the formulation of razors.[74]Describing the program for the universal program as the "hypothesis", and the representation of the evidence as program data, it has been formally proven underZermelo–Fraenkel set theorythat "the sum of the log universal probability of the model plus the log of the probability of the data given the model should be minimized."[75]Interpreting this as minimising the total length of a two-part message encoding model followed by data given model gives us theminimum message length(MML) principle.[71][72]

One possible conclusion from mixing the concepts of Kolmogorov complexity and Occam's razor is that an ideal data compressor would also be a scientific explanation/formulation generator. Some attempts have been made to re-derive known laws from considerations of simplicity or compressibility.[24][76]

According toJürgen Schmidhuber,the appropriate mathematical theory of Occam's razor already exists, namely,Solomonoff'stheory of optimal inductive inference[77]and its extensions.[78]See discussions in David L. Dowe's "Foreword re C. S. Wallace"[79]for the subtle distinctions between thealgorithmic probabilitywork of Solomonoff and the MML work ofChris Wallace,and see Dowe's "MML, hybrid Bayesian network graphical models, statistical consistency, invariance and uniqueness"[80]both for such discussions and for (in section 4) discussions of MML and Occam's razor. For a specific example of MML as Occam's razor in the problem of decision tree induction, see Dowe and Needham's "Message Length as an Effective Ockham's Razor in Decision Tree Induction".[81]

Mathematical arguments against Occam's razor

[edit]

Theno free lunch(NFL) theorems for inductive inference prove that Occam's razor must rely on ultimately arbitrary assumptions concerning the prior probability distribution found in our world.[82]Specifically, suppose one is given two inductive inference algorithms, A and B, where A is aBayesianprocedure based on the choice of some prior distribution motivated by Occam's razor (e.g., the prior might favor hypotheses with smallerKolmogorov complexity). Suppose that B is the anti-Bayes procedure, which calculates what the Bayesian algorithm A based on Occam's razor will predict – and then predicts the exact opposite. Then there are just as many actual priors (including those different from the Occam's razor prior assumed by A) in which algorithm B outperforms A as priors in which the procedure A based on Occam's razor comes out on top. In particular, the NFL theorems show that the "Occam factors" Bayesian argument for Occam's razor must make ultimately arbitrary modeling assumptions.[83]

Software development

[edit]

In software development, therule of least powerargues the correctprogramming languageto use is the one that is simplest while also solving the targeted software problem. In that form the rule is often credited toTim Berners-Leesince it appeared in his design guidelines for the originalHypertext Transfer Protocol.[84]Complexity in this context is measured either by placing a language into theChomsky hierarchyor by listing idiomatic features of the language and comparing according to some agreed to scale of difficulties between idioms. Many languages once thought to be of lower complexity have evolved or later been discovered to be more complex than originally intended; so, in practice this rule is applied to the relative ease of a programmer to obtain the power of the language, rather than the precise theoretical limits of the language.

Controversial aspects

[edit]

Occam's razor is not an embargo against the positing of any kind of entity, or a recommendation of the simplest theory come what may.[a]Occam's razor is used to adjudicate between theories that have already passed "theoretical scrutiny" tests and are equally well-supported by evidence.[b]Furthermore, it may be used to prioritize empirical testing between two equally plausible but unequally testable hypotheses; thereby minimizing costs and wastes while increasing chances of falsification of the simpler-to-test hypothesis.[citation needed]

Another contentious aspect of the razor is that a theory can become more complex in terms of its structure (orsyntax), while itsontology(orsemantics) becomes simpler, or vice versa.[c]Quine, in a discussion on definition, referred to these two perspectives as "economy of practical expression" and "economy in grammar and vocabulary", respectively.[86]

Galileo Galileilampooned themisuseof Occam's razor in hisDialogue.The principle is represented in the dialogue by Simplicio. The telling point that Galileo presented ironically was that if one really wanted to start from a small number of entities, one could always consider the letters of the alphabet as the fundamental entities, since one could construct the whole of human knowledge out of them.

Instances of using Occam's razor to justify belief in less complex and more simple theories have been criticized as using the razor inappropriately. For instanceFrancis Crickstated that "While Occam's razor is a useful tool in the physical sciences, it can be a very dangerous implement in biology. It is thus very rash to use simplicity and elegance as a guide in biological research."[87]

Anti-razors

[edit]

Occam's razor has met some opposition from people who consider it too extreme or rash.Walter Chatton(c. 1290–1343) was a contemporary of William of Ockham who took exception to Occam's razor and Ockham's use of it. In response he devised his ownanti-razor:"If three things are not enough to verify an affirmative proposition about things, a fourth must be added and so on." Although there have been several philosophers who have formulated similar anti-razors since Chatton's time, no one anti-razor has perpetuated as notably as Chatton's anti-razor, although this could be the case of the Late Renaissance Italian motto of unknown attributionSe non è vero, è ben trovato( "Even if it is not true, it is well conceived" ) when referred to a particularly artful explanation.

Anti-razors have also been created byGottfried Wilhelm Leibniz(1646–1716),Immanuel Kant(1724–1804), andKarl Menger(1902–1985). Leibniz's version took the form of aprinciple of plenitude,asArthur Lovejoyhas called it: the idea being that God created the most varied and populous of possible worlds. Kant felt a need to moderate the effects of Occam's razor and thus created his own counter-razor: "The variety of beings should not rashly be diminished."[88]

Karl Menger found mathematicians to be too parsimonious with regard to variables so he formulated his Law Against Miserliness, which took one of two forms: "Entities must not be reduced to the point of inadequacy" and "It is vain to do with fewer what requires more." A less serious but even more extremist anti-razor is'Pataphysics,the "science of imaginary solutions" developed byAlfred Jarry(1873–1907). Perhaps the ultimate in anti-reductionism, "'Pataphysics seeks no less than to view each event in the universe as completely unique, subject to no laws but its own." Variations on this theme were subsequently explored by the Argentine writerJorge Luis Borgesin his story/mock-essay "Tlön, Uqbar, Orbis Tertius".PhysicistR. V. Jonescontrived Crabtree's Bludgeon, which states that "[n]o set of mutually inconsistent observations can exist for which some human intellect cannot conceive a coherent explanation, however complicated."[89]

Recently, American physicist Igor Mazin argued that because high-profile physics journals prefer publications offering exotic and unusual interpretations, the Occam's razor principle is being replaced by an "Inverse Occam's razor", implying that the simplest possible explanation is usually rejected.[90]

Other

[edit]

Since 2012The Skepticmagazine annually awards the Ockham Awards, or simply the Ockhams, named after Occam's razor, atQED.[91]The Ockhams were introduced by editor-in-chiefDeborah Hydeto "recognise the effort and time that have gone into the community’s favourite skeptical blogs, skeptical podcasts, skeptical campaigns and outstanding contributors to the skeptical cause."[92]Thetrophies,designed by Neil Davies and Karl Derrick, carry the upper text "Ockham's"and the lower text"The Skeptic. Shaving away unnecessary assumptions since 1285."Between the texts, there is an image of a double-edgedsafety razorblade,and both lower corners feature an image of William of Ockham's face.[92]

See also

[edit]

Notes

[edit]
  1. ^"Ockham's razor does not say that the more simple a hypothesis, the better."[85]
  2. ^"Today, we think of the principle of parsimony as a heuristic device. We don't assume that the simpler theory is correct and the more complex one false. We know from experience that more often than not the theory that requires more complicated machinations is wrong. Until proved otherwise, the more complex theory competing with a simpler explanation should be put on the back burner, but not thrown onto the trash heap of history until proven false."[85]
  3. ^"While these two facets of simplicity are frequently conflated, it is important to treat them as distinct. One reason for doing so is that considerations of parsimony and of elegance typically pull in different directions. Postulating extra entities may allow a theory to be formulated more simply, while reducing the ontology of a theory may only be possible at the price of making it syntactically more complex."[53]

References

[edit]
  1. ^Barry, C. M. (27 May 2014)."Who sharpened Occam's Razor?".Irish Philosophy.Archivedfrom the original on 5 October 2022.Retrieved5 August2022.
  2. ^Schaffer, Jonathan (2015)."What Not to Multiply Without Necessity"(PDF).Australasian Journal of Philosophy.93(4): 644–664.doi:10.1080/00048402.2014.992447.S2CID16923735.Archived(PDF)from the original on 9 September 2020.Retrieved8 August2019.
  3. ^Duignan, Brian."Occam's Razor".Encyclopedia Britannica.Archivedfrom the original on 25 September 2023.Retrieved11 May2021.
  4. ^Ball, Philip (11 August 2016)."The Tyranny of Simple Explanations".The Atlantic.Archivedfrom the original on 2 February 2023.Retrieved2 February2023.
  5. ^abcHugh G. Gauch,Scientific Method in Practice, Cambridge University Press,2003,ISBN0-521-01708-4,ISBN978-0-521-01708-4.
  6. ^abcdefHoffman, Roald; Minkin, Vladimir I.; Carpenter, Barry K. (1997)."Ockham's Razor and Chemistry".Hyle: International Journal for Philosophy of Chemistry.3:3–28.Archivedfrom the original on 14 July 2018.Retrieved30 May2004.
  7. ^abSober, Elliott (2015).Ockam's Razor: A User's Manual.Cambridge University Press. p. 4.ISBN978-1107692534.
  8. ^Roger Ariew, Ockham's Razor: A Historical and Philosophical Analysis of Ockham's Principle of Parsimony, 1976
  9. ^abJohannes Poncius's commentary on John Duns Scotus'sOpus Oxoniense,book III, dist. 34, q. 1. in John Duns ScotusOpera Omnia,vol.15, Ed. Luke Wadding, Louvain (1639), reprinted Paris: Vives, (1894) p.483a
  10. ^Aristotle,Physics189a15,On the Heavens271a33. See also Franklin,op cit.note 44 to chap. 9.
  11. ^Charlesworth, M. J. (1956). "Aristotle's Razor".Philosophical Studies.6:105–112.doi:10.5840/philstudies1956606.
  12. ^abcFranklin, James(2001).The Science of Conjecture: Evidence and Probability before Pascal.The Johns Hopkins University Press.Chap 9. p. 241.
  13. ^Alistair Cameron Crombie,Robert Grosseteste and the Origins of Experimental Science 1100–1700(1953) pp. 85–86
  14. ^"SUMMA THEOLOGICA: The existence of God (Prima Pars, Q. 2)".Newadvent.org.Archivedfrom the original on 28 April 2013.Retrieved26 March2013.
  15. ^Vallee, Jacques (11 February 2013)."What Ockham really said".Boing Boing.Archivedfrom the original on 31 March 2013.Retrieved26 March2013.
  16. ^Bauer, Laurie (2007).The linguistics Student's Handbook.Edinburgh: Edinburgh University Press.p. 155.
  17. ^Flew, Antony(1979).A Dictionary of Philosophy.London: Pan Books.p. 253.
  18. ^Crombie, Alistair Cameron(1959),Medieval and Early Modern Philosophy,Cambridge, MA: Harvard, Vol. 2, p. 30.
  19. ^"Ockham's razor".Encyclopædia Britannica.Encyclopædia Britannica Online. 2010.Archivedfrom the original on 23 August 2010.Retrieved12 June2010.
  20. ^Hawking, Stephen(2003).On the Shoulders of Giants.Running Press. p. 731.ISBN978-0-7624-1698-1.Retrieved24 February2016.[permanent dead link]
  21. ^Primary source:Newton (2011,p. 387) wrote the following two "philosophizing rules" at the beginning of part 3 of thePrincipia1726 edition.
    Regula I. Causas rerum naturalium non-plures admitti debere, quam quæ & veræ sint & earum phænomenis explicandis sufficient.
    Regula II. Ideoque effectuum naturalium ejusdem generis eædem assignandæ sunt causæ, quatenus fieri potest.
  22. ^Logical Constructions.Metaphysics Research Lab, Stanford University. 2016.Archivedfrom the original on 26 January 2021.Retrieved29 March2011.
  23. ^Induction: From Kolmogorov and Solomonoff to De Finetti and Back to Kolmogorov JJ McCall – Metroeconomica, 2004 – Wiley Online Library.
  24. ^abSoklakov, A. N. (2002). "Occam's Razor as a formal basis for a physical theory".Foundations of Physics Letters.15(2): 107–135.arXiv:math-ph/0009007.Bibcode:2000math.ph...9007S.doi:10.1023/A:1020994407185.S2CID14940740.
  25. ^Rathmanner, Samuel;Hutter, Marcus(2011)."A philosophical treatise of universal induction".Entropy.13(6): 1076–1136.arXiv:1105.5721.Bibcode:2011Entrp..13.1076R.doi:10.3390/e13061076.S2CID2499910.
  26. ^Baker, Alan (25 February 2010)."Simplicity".In Zalta, Edward N. (ed.).The Stanford Encyclopedia of Philosophy (Summer 2011 Edition).Archivedfrom the original on 24 February 2021.Retrieved6 April2013.
  27. ^abc"What is Occam's Razor?".math.ucr.edu.Archivedfrom the original on 6 July 2017.
  28. ^Stormy Dawn (17 July 2017).Everywhere The Soles of Your Feet Shall Tread.Archway.ISBN9781480838024.Archivedfrom the original on 28 October 2023.Retrieved22 May2017.
  29. ^Sotos, John G. (2006) [1991].Zebra Cards: An Aid to Obscure Diagnoses.Mt. Vernon, VA: Mt. Vernon Book Systems.ISBN978-0-9818193-0-3.
  30. ^Becher, Erich (1905). "The Philosophical Views of Ernst Mach".The Philosophical Review.14(5): 535–562.doi:10.2307/2177489.JSTOR2177489.
  31. ^Pegis 1945.
  32. ^Stanovich, Keith E. (2007).How to Think Straight About Psychology.Boston: Pearson Education, pp. 19–33.
  33. ^"ad hoc hypothesis - The Skeptic's Dictionary - Skepdic.com".skepdic.com.Archived fromthe originalon 27 April 2009.
  34. ^Swinburne 1997 and Williams, Gareth T, 2008.
  35. ^Green, K. C.; Armstrong, J. S. (2015)."Simple versus complex forecasting: The evidence".Journal of Business Research.68(8): 1678–1685.doi:10.1016/j.jbusres.2015.03.026.Archivedfrom the original on 8 June 2020.Retrieved22 January2019.(subscription required)
  36. ^MacKay, David J. C. (2003).Information Theory, Inference, and Learning Algorithms(PDF).Bibcode:2003itil.book.....M.Archived(PDF)from the original on 15 September 2012.
  37. ^abJefferys, William H.; Berger, James O. (1991)."Ockham's Razor and Bayesian Statistics"(PDF).American Scientist.80(1): 64–72.JSTOR29774559.Archived(PDF)from the original on 4 March 2005.(preprint available as "Sharpening Occam's Razor on a Bayesian Strop" ).
  38. ^James, Gareth; et al. (2013).An Introduction to Statistical Learning.springer. pp. 105, 203–204.ISBN9781461471370.
  39. ^Popper, Karl(1992) [1934].Logik der Forschung[The Logic of Scientific Discovery] (2nd ed.). London: Routledge. pp. 121–132.ISBN978-84-309-0711-3.
  40. ^Sober, Elliott(1975).Simplicity.Oxford:Clarendon Press.ISBN978-0-19-824407-3.
  41. ^Sober, Elliott (2004)."What is the Problem of Simplicity?".In Zellner, Arnold;Keuzenkamp, Hugo A.;McAleer, Michael (eds.).Simplicity, Inference and Modeling: Keeping it Sophisticatedly Simple.Cambridge, U.K.: Cambridge University Press. pp. 13–31.ISBN978-0-521-80361-8.Archivedfrom the original on 28 October 2023.Retrieved4 August2012.Paper as PDF.
  42. ^Swinburne, Richard (1997). Simplicity as Evidence for Truth. Milwaukee, Wisconsin: Marquette University Press.ISBN978-0-87462-164-8.
  43. ^Einstein, Albert(1905)."Does the Inertia of a Body Depend Upon Its Energy Content?".Annalen der Physik(in German).323(18): 639–41.Bibcode:1905AnP...323..639E.doi:10.1002/andp.19053231314.Archivedfrom the original on 21 October 2019.Retrieved21 October2019.
  44. ^L. Nash, The Nature of the Natural Sciences, Boston: Little, Brown (1963).
  45. ^de Maupertuis, P. L. M. (1744).Mémoires de l'Académie Royale(in French). p. 423.
  46. ^de Broglie, L. (1925).Annales de Physique(in French). pp. 22–128.
  47. ^RA Jackson, Mechanism: An Introduction to the Study of Organic Reactions, Clarendon, Oxford, 1972.
  48. ^Carpenter, B. K. (1984).Determination of Organic Reaction Mechanism,New York: Wiley-Interscience.
  49. ^Einstein, Albert (1934)."On the Method of Theoretical Physics".Philosophy of Science.1(2): 165 [163–169].doi:10.1086/286316.JSTOR184387.S2CID44787169.Archivedfrom the original on 22 January 2023.Retrieved22 January2023.
  50. ^Mettenheim, Christoph von (1998).Popper Versus Einstein: On the Philosophical Foundations of Physics.Mohr Siebeck. p. 34.ISBN978-3-16-146910-7.Archivedfrom the original on 22 January 2023.Retrieved22 January2023.
  51. ^Geis, Gilbert; Geis, Professor Emeritus of Criminology Law and & Society Gilbert; Bienen, Leigh B. (1998).Crimes of the Century: From Leopold and Loeb to O.J. Simpson.UPNE. p. 39.ISBN978-1-55553-360-1.Archivedfrom the original on 5 April 2023.Retrieved10 February2023.
  52. ^"Everything Should Be Made as Simple as Possible, But Not Simpler".13 May 2011.Archivedfrom the original on 29 May 2012.
  53. ^abcAlan Baker (2010) [2004]."Simplicity".Stanford Encyclopedia of Philosophy.California: Stanford University.Archivedfrom the original on 26 March 2014.Retrieved22 January2005.
  54. ^abcdeCourtney, A.; Courtney, M. (2008). "Comments Regarding 'On the Nature of Science'".Physics in Canada.64(3): 7–8.arXiv:0812.4932.Bibcode:2008arXiv0812.4932C.
  55. ^abcSober, Elliott (1994). "Let's Razor Occam's Razor". In Knowles, Dudley (ed.).Explanation and Its Limits.Cambridge University Press. pp. 73–93.
  56. ^Rabinowitz, Matthew; Myers, Lance; Banjevic, Milena; Chan, Albert; Sweetkind-Singer, Joshua; Haberer, Jessica; McCann, Kelly; Wolkowicz, Roland (1 March 2006)."Accurate prediction of HIV-1 drug response from the reverse transcriptase and protease amino acid sequences using sparse models created by convex optimization".Bioinformatics.22(5): 541–549.doi:10.1093/bioinformatics/btk011.PMID16368772.
  57. ^Paul Pojman (2009)."Ernst Mach".The Stanford Encyclopedia of Philosophy.California: Stanford University.Archivedfrom the original on 11 November 2020.Retrieved4 October2009.
  58. ^Sober, Elliot (1998).Reconstructing the Past: Parsimony, Evolution, and Inference(2nd ed.). Massachusetts Institute of Technology: The MIT Press. p. 7.ISBN978-0-262-69144-4.
  59. ^Wiley, Edward O. (2011).Phylogenetics: the theory and practice of phylogenetic systematics(2nd ed.). Wiley-Blackwell.ISBN978-0-470-90596-8.
  60. ^Brower, AVZ (2017)."Statistical consistency and phylogenetic inference: a brief review".Cladistics.34(5): 562–567.doi:10.1111/cla.12216.PMID34649374.
  61. ^Brower &, Schuh (2021).Biological Systematics: Principles and Applications (3rd edn.).Cornell University Press.
  62. ^Crick 1988, p. 146.
  63. ^"William Ockham".Encyclopedia of Philosophy.Stanford.Archivedfrom the original on 7 October 2019.Retrieved24 February2016.
  64. ^Dale T Irvin & Scott W Sunquist.History of World Christian Movement Volume, I: Earliest Christianity to 1453,p. 434.ISBN9781570753961.
  65. ^"SUMMA THEOLOGICA: The existence of God (Prima Pars, Q. 2)".Newadvent.org.Archivedfrom the original on 28 April 2013.Retrieved26 March2013.
  66. ^McDonald 2005.
  67. ^Dawkins, Richard (1 January 2007).The God delusion.London: Black Swan. pp. 157–158.ISBN978-0-552-77331-7.
  68. ^Schmitt, Carl; Schwab, George; Strong, Tracy B. (2005).Political Theology.University of Chicago Press.doi:10.7208/chicago/9780226738901.001.0001.ISBN978-0-226-73889-5.
  69. ^p. 282,Mémoires du docteur F. Antommarchi, ou les derniers momens de NapoléonArchived14 May 2016 at theWayback Machine,vol. 1, 1825, Paris: Barrois L'Ainé
  70. ^Tonry, Michael (2005)."Obsolescence and Immanence in Penal Theory and Policy"(PDF).Columbia Law Review.105:1233–1275. Archived fromthe original(PDF)on 23 June 2006.
  71. ^abWallace, C. S.; Boulton, D. M. (1 August 1968)."An Information Measure for Classification".The Computer Journal.11(2): 185–194.doi:10.1093/comjnl/11.2.185.
  72. ^abWallace, C. S. (1 April 1999)."Minimum Message Length and Kolmogorov Complexity"(PDF).The Computer Journal.42(4): 270–283.doi:10.1093/comjnl/42.4.270.
  73. ^Nannen, Volker."A short introduction to Model Selection, Kolmogorov Complexity and Minimum Description Length"(PDF).Archived(PDF)from the original on 2 June 2010.Retrieved3 July2010.
  74. ^"Algorithmic Information Theory".Archivedfrom the original on 24 December 2007.
  75. ^Vitanyi, P.M.B.; Ming Li (March 2000)."Minimum description length induction, Bayesianism, and Kolmogorov complexity".IEEE Transactions on Information Theory.46(2): 446–464.arXiv:cs/9901014.doi:10.1109/18.825807.
  76. ^Standish, Russell K (2000). "Why Occam's Razor".Foundations of Physics Letters.17(3): 255–266.arXiv:physics/0001020.Bibcode:2004FoPhL..17..255S.doi:10.1023/B:FOPL.0000032475.18334.0e.S2CID17143230.
  77. ^Solomonoff, Ray(1964)."A formal theory of inductive inference. Part I."Information and Control.7(1–22): 1964.doi:10.1016/s0019-9958(64)90223-2.
  78. ^Schmidhuber, J.(2006). "The New AI: General & Sound & Relevant for Physics". In Goertzel, B.; Pennachin, C. (eds.).Artificial General Intelligence.pp. 177–200.arXiv:cs.AI/0302012.
  79. ^Dowe, David L. (2008). "Foreword re C. S. Wallace".Computer Journal.51(5): 523–560.doi:10.1093/comjnl/bxm117.S2CID5387092.
  80. ^David L. Dowe (2010): "MML, hybrid Bayesian network graphical models, statistical consistency, invariance and uniqueness. A formal theory of inductive inference."Handbook of the Philosophy of Science– (HPS Volume 7) Philosophy of Statistics, Elsevier 2010 Page(s):901–982.https://web.archive.org/web/20140204001435/http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.185.709&rep=rep1&type=pdf
  81. ^Scott Needham and David L. Dowe (2001): "Message Length as an Effective Ockham's Razor in Decision Tree Induction." Proc. 8th International Workshop on Artificial Intelligence and Statistics (AI+STATS 2001), Key West, Florida, U.S.A., January 2001 Page(s): 253–260"2001 Ockham.pdf"(PDF).Archived(PDF)from the original on 23 September 2015.Retrieved2 September2015.
  82. ^Adam, S., and Pardalos, P. (2019),No-free lunch Theorem: A review,in "Approximation and Optimization", Springer, 57-82
  83. ^Wolpert, D.H (1995), On the Bayesian "Occam Factors" Argument for Occam's Razor, in "Computational Learning Theory and Natural Learning Systems: Selecting Good Models", MIT Press
  84. ^Berners-Lee, Tim(4 March 2013)."Principles of Design".World Wide Web Consortium.Archivedfrom the original on 15 June 2022.Retrieved5 June2022.
  85. ^abRobert T. Carroll(12 September 2014)."Occam's Razor".The Skeptic's Dictionary.Archivedfrom the original on 1 March 2016.Retrieved24 February2016.
  86. ^Quine, W V O(1961). "Two dogmas of empiricism".From a logical point of view.Cambridge: Harvard University Press. pp. 20–46.ISBN978-0-674-32351-3.
  87. ^Gross, Fridolin (December 2019)."Occam's Razor in Molecular and Systems Biology".Philosophy of Science.86(5): 1134–1145.doi:10.1086/705474.
  88. ^Immanuel Kant(1929). Norman Kemp-Smith transl (ed.).The Critique of Pure Reason.Palgrave Macmillan. p. 92.Archivedfrom the original on 16 May 2012.Retrieved27 October2012.Entium varietates non-temere esse minuendas
  89. ^Gordon Woo (20 June 2011).Calculating Catastrophe.World Scientific. pp. 303–.ISBN978-1-84816-893-0.Archivedfrom the original on 28 October 2023.Retrieved10 August2021.
  90. ^Mazin, Igor (April 2022)."Inverse Occam's razor".Nature Physics.18(4): 367–368.arXiv:2204.08284.doi:10.1038/s41567-022-01575-2.S2CID247832936.Archivedfrom the original on 9 July 2023.Retrieved9 July2023.
  91. ^Korteweg, Leon (2 December 2016)."QED 2016 – verslag van een lang weekend tussen skeptici".Skepter(in Dutch).29(4).Stichting Skepsis:45–46.Archivedfrom the original on 18 October 2017.Retrieved21 October2017.
  92. ^abHyde, Deborah(2012)."The Skeptic Magazine Awards 2011: Winners".The Skeptic.Vol. 23, no. 4.Archivedfrom the original on 23 October 2017.

Further reading

[edit]
[edit]
  • Ockham's Razor,BBC Radio 4 discussion with Sir Anthony Kenny, Marilyn Adams & Richard Cross (In Our Time,31 May 2007)