Imre Lakatos; November 9, 1922 – February 2, 1974) was a Hungarian philosopher of mathematics and science, known for his thesis of the fallibility of mathematics and its ‘methodology of proofs and refutations’ in its pre-axiomatic stages of development, and also for introducing the concept of the ‘research programme’ in his methodology of scientific research programmes.
The transcript of his talk is below.
Man’s respect for knowledge is one of his most peculiar characteristics. Knowledge in Latin is scientia, and science came to be the name of the most respectable kind of knowledge. But what distinguishes knowledge from superstition, ideology or pseudoscience? The Catholic Church excommunicated Copernicans, the Communist Party persecuted Mendelians on the ground that their doctrines were pseudoscientific. But then the problem of the demarcation between science and pseudoscience is not merely a problem of armchair philosophy: it is of vital social and political relevance.
Many philosophers have tried to solve the problem of demarcation in the following terms: a statement constitutes knowledge if sufficiently many people believe it sufficiently strongly. But the history of thought shows us that many people were totally committed to absurd beliefs. If the strengths of beliefs were a hallmark of knowledge, we should have to rank some tales about demons, angels, devils, and of heaven and hell as knowledge. Scientists, on the other hand, are very sceptical even of their best theories. Newton’s is the most powerful theory science has yet produced, but Newton himself never believed that bodies attract each other at a distance. So no degree of commitment to beliefs makes them knowledge. Indeed, the hallmark of scientific behaviour is a certain scepticism even towards one’s most cherished theories. Blind commitment to a theory is not an intellectual virtue: it is an intellectual crime.
Thus a statement may be pseudoscientific even if it is eminently ‘plausible’ and everybody believes in it, and it may be scientifically valuable even if it is unbelievable and nobody believes in it. A theory may even be of supreme scientific value even if no one understands it, let alone believes in it.
The cognitive value of a theory has nothing to do with its psychological influence on people’s minds. Belief, commitment, understanding are states of the human mind. But the objective, scientific value of a theory is independent of the human mind which creates it or understands it. Its scientific value depends only on what objective support these conjectures have in facts. As Hume said:
If we take in our hand any volume; of divinity, or school metaphysics, for instance; let us ask, does it contain any abstract reasoning concerning quantity or number? No. Does it contain any experimental reasoning concerning matter of fact and existence? No. Commit it then to the flames. For it can contain nothing but sophistry and illusion.
But what exactly is ‘experimental’ reasoning? [If we look at the vast seventeenth-century literature on witchcraft, it is full of reports of careful observations and sworn evidence – even of experiments. Glanvill, the house philosopher of the early Royal Society, regarded witchcraft as the paradigm of experimental reasoning. We have to define experimental reasoning before we start Humean book burning.]
In scientific reasoning, theories are confronted with facts; and one of the central conditions of scientific reasoning is that theories must be supported by facts. Now how exactly can facts support theory?
Several different answers have been proposed. Newton himself thought that he proved his laws from facts. [He was proud of not uttering mere hypotheses: he only published theories proven from facts. In particular,] He claimed that he deduced his laws from the ‘phenomena’ provided by Kepler. But his boast was nonsense, since according to Kepler, planets move in ellipses, but according to Newton’s theory, planets would move in ellipses only if the planets did not disturb each other in their motion. But they do. This is why Newton had to devise a perturbation theory from which it follows that no planet moves in an ellipse.
One can today easily demonstrate that there can be no valid derivation of a law of nature from any finite number of facts; but we still keep reading about scientific theories being proved from facts. Why this stubborn resistance to elementary logic?
There is a very plausible explanation. Scientists want to make their theories respectable, deserving of the title ‘science’, that is, genuine knowledge. Now the most relevant knowledge in the seventeenth century, when science was born, concerned God, the Devil, Heaven and Hell. If one got one’s conjectures about matters of divinity wrong, the consequence of one’s mistake was no less than eternal damnation. Theological knowledge cannot be fallible: it must be beyond doubt. Now the Enlightenment thought that we were fallible and ignorant about matters theological. There is no scientific theology and, therefore, no theological knowledge. Knowledge can only be about Nature, but this new type of knowledge had to be judged by the standards they took over straight from theology: it had to be proven beyond doubt. Science had to achieve the very certainty which had escaped theology. A scientist, worthy of the name, was not allowed to guess: he had to prove each sentence he uttered from facts. This was the criterion of scientific honesty. Theories unproven from facts were regarded as sinful pseudoscience, heresy in the scientific community.
It was only the downfall of Newtonian theory in this century which made scientists realize that their standards of honesty had been utopian. [Before Einstein most scientists thought that Newton had deciphered God’s ultimate laws by proving them from the facts. Ampère, in the early nineteenth century, felt he had to call his book on his speculations concerning electromagnetism: Mathematical Theory of Electrodynamic Phenomena Unequivocally Deduced from Experiment. But at the end of the volume he casually confesses that some of the experiments were never performed and even that the necessary instruments had not been constructed! ] If all scientific theories are equally unprovable, what distinguishes scientific knowledge from ignorance, science from pseudoscience?
One answer to this question was provided in the twentieth century by ‘inductive logicians’. Inductive logic set out to define the probabilities of different theories according to the available total evidence. If the mathematical probability of a theory is high, it qualifies as scientific; if it is low or even zero, it is not scientific. Thus the hallmark of scientific honesty would be never to say anything that is not at least highly probable. [ Probabilism has an attractive feature: instead of simply providing a black-and-white distinction between science and pseudoscience, it provides a continuous scale from poor theories with low probability to good theories with high probability.]
But, in 1934, Karl Popper, one of the most influential philosophers of our time, argued that the mathematical probability of all theories, scientific or pseudoscientific, given any amount of evidence is zero. If Popper is right, scientific theories are not only equally unprovable but also equally improbable.
A new demarcation criterion was needed and Popper proposed a rather stunning one.[ A theory may be scientific even if there is not a shred of evidence in its favour, and it may be pseudoscientific even if all the available evidence is in its favour. That is, the scientific or non-scientific character of a theory can be determined independently of the facts.] A theory is ‘scientific’ if one is prepared to specify in advance a crucial experiment (or observation) which can falsify it, and it is pseudoscientific if one refuses to specify such a ‘potential falsifier’. But if so, we do not demarcate scientific theories from pseudoscientific ones, but rather scientific methods from non-scientific method. [Marxism, for a Popperian, is scientific if the Marxists are prepared to specify facts which , if observed, make them give up Marxism. If they refuse to do so, Marxism becomes a pseudoscience. It is always interesting to ask a Marxist, what conceivable event would make him abandon his Marxism. If he is committed to Marxism, he is bound to find it immoral to specify a state of affairs which can falsify it.] Thus a proposition may petrify into pseudo-scientific dogma or become genuine knowledge, depending on whether we are prepared to state observable conditions which would refute it.
Is, then, Popper’s falsifiability criterion the solution to the problem of demarcating science from pseudoscience? No. For Popper’s criterion ignores the remarkable tenacity of scientific theories. Scientists have thick skins. They do not abandon a theory [merely] because facts contradict it. They normally either invent some rescue hypothesis to explain what they then call a mere anomaly and if they cannot explain the anomaly, they ignore it, and direct their attention to other problems. Note that scientists talk about anomalies, [recalcitrant instances,] and not refutations. History of science, of course, is full of accounts of how crucial experiments allegedly killed theories. But all such accounts are fabricated long after the theory has been abandoned. [Had Popper ever asked a Newtonian scientist under what experimental conditions he would abandon Newtonian theory, some Newtonian scientists would have been exactly as nonplussed as are some Marxists.]
What, then, is the hallmark of science? Do we have to capitulate and agree that a scientific revolution is just an irrational change in commitment, that it is a religious conversion? Tom Kuhn, a distinguished American philosopher of science, arrived at this conclusion after discovering the naivety of Popper’s falsificationism. But if Kuhn is right, then there is no explicit demarcation between science and pseudoscience, no distinction between scientific progress and intellectual decay, there is no objective standard of honesty. But what criteria can he then offer to demarcate scientific progress from intellectual degeneration ?
In the last few years I have been advocating a methodology of scientific research programmes, which solves some of the problems which both Popper and Kuhn failed to solve.
First, I claim that the typical descriptive unit of great scientific achievements is not an isolated hypothesis but rather a research programme. [Science is not simply trial and error, a series of conjectures and refutations.] ‘All swans are white’ may be falsified by the discovery of one black swan. But such trivial trial and error does not rank as science. Newtonian science, for instance, is not simply a set of four conjectures – the three laws of mechanics and the law of gravitation. These four laws constitute only the ‘hard core’ of the Newtonian programme. But this hard core is tenaciously protected from refutation by a vast ‘protective belt’ of auxiliary hypotheses. And, even more importantly, the research programme also has a ‘heuristic’, that is, a powerful problem-solving machinery, which, with the help of sophisticated mathematical techniques, digests anomalies and even turns them into positive evidence. For instance, if a planet does not move exactly as it should, the Newtonian scientist checks his conjectures concerning atmospheric refraction, concerning propagation of light in magnetic storms, and hundreds of other conjectures which are all part of the programme. He may even invent a hitherto unknown planet and calculate its position, mass and velocity in order to explain the anomaly.
Now, Newton’s theory of gravitation, Einstein’s relativity theory, quantum mechanics, Marxism, Freudism, are all research programmes, each with a characteristic hard core stubbornly defended, each with its more flexible protective belt and each with its elaborate problem-solving machinery. Each of them, at any stage of its development, has unsolved problems and undigested anomalies. All theories, in this sense, are born refuted and die refuted. But are they equally good? Until now I have been describing what research programmes are like. But how can one distinguish a scientific or progressive programme from a pseudoscientific or degenerating one?
Contrary to Popper, the difference cannot be that some are still unrefuted, while others are already refuted. [When Newton published his Principia, it was common knowledge that it could not properly explain even the motion of the moon; in fact, lunar motion refuted Newton.] Kaufmann, a distinguished physicist, refuted Einstein’s relativity theory in the very year it was published. But all the research programmes I admire have one characteristic in common. They all predict novel facts, facts which had been either undreamt of, or have indeed been contradicted by previous or rival programmes. In 1686, when Newton published his theory of gravitation, there were, for instance, two current theories concerning comets. The more popular one regarded comets as a signal from an angry God warning that He will strike and bring disaster. A little known theory of Kepler’s held that comets were celestial bodies moving along straight lines. Now according to Newtonian theory, some of them moved in hyperbolas or parabolas never to return; others moved in ordinary ellipses. Halley, working in Newton’s programme, calculated on the basis of observing a brief stretch of a comet’s path that it would return in seventy-two year’s time; he calculated to the minute when it would be seen again at a well-defined point of the sky. This was incredible. But seventy-two years later, [when both Newton and Halley were long dead,] Halley’s comet returned exactly as Halley predicted. Similarly, Newtonian scientists predicted the existence and exact motion of small planets which had never been observed before. [Or let us take Einstein’s programme. This programme made the stunning prediction that if one measures the distance between two stars in the night and if one measure the distance between them during the day (when they are visible during an eclipse of the sun), the two measurements will be different. Nobody had thought to make such an observation before Einstein’s programme.] Thus, in a progressive research programme, theory leads to the discovery of hitherto unknown novel facts.
In degenerating programmes, however, theories are fabricated only in order to accommodate known facts. Has, for instance, Marxism ever predicted a stunning novel fact successfully? Never! It has some famous unsuccessful predictions. It predicted the absolute impoverishment of the working class. It predicted that the first socialist revolution would take place in the industrially most developed society. It predicted that socialist societies would be free of revolutions. It predicted that there will be no conflict of interests between socialist countries. Thus the early predictions of Marxism were bold and stunning, but they failed.
Marxism ‘explained’ all its failures. It ‘explained’ the rising living standards of the working class by devising a theory of imperialism; it ‘explained’ even why the first socialist revolution occurred in industrially backward Russia. It ‘explained’ Berlin 1953, Budapest 1956, Prague 1968. It ‘explained’ the Russian-Chinese conflict. But their auxiliary hypotheses were all cooked up after the event to protect Marxian theory from the facts. The Newtonian programme led to novel facts; the Marxian programme lagged behind the facts and has been running fast to catch up with them.
To sum up: [The hallmark of empirical progress is not trivial verifications: Popper is right that there are millions of them. It is no success for Newtonian theory that stones, when dropped, fall towards the earth, no matter how often this is repeated. But, ] so-called ‘refutations’ are not the hallmark of empirical failure, as Popper has preached, since all programmes grow in a permanent ocean of anomalies. What really counts are dramatic, unexpected, stunning predictions: a few of them are enough to tilt the balance; where theory lags behind the facts, we are dealing with miserable degenerating research programmes.
Now, how do scientific revolutions come about? If we have two rival research programmes, and one is progressing while the other is degenerating, scientists tend to join the progressive programme. This is the rationale of scientific revolutions. But while it is a matter of intellectual honesty to keep the record public, it is not dishonest to stick to a degenerating programme and try to turn it into a progressive one.
As opposed to Popper the methodology of scientific research programmes does not offer instant rationality. One must treat budding programmes leniently: programmes may take decades before they get off the ground and become empirically progressive. Criticism is not a Popperian quick kill, by refutation. Important criticism is always constructive: there is no refutation without a better theory. Kuhn is wrong in thinking that scientific revolutions are sudden, irrational changes in vision. [The history of science refutes both Popper and Kuhn: ] On close inspection both Popperian crucial experiments and Kuhnian revolutions turn out to be myths: what normally happens is that progressive research programmes replace degenerating ones.
The problem of demarcation between science and pseudoscience has grave implications also for the institutionalization of criticism. Copernicus’s theory was banned by the Catholic Church in 1616 because it was said to be pseudoscientific. It was taken off the index in 1820 because by that time the Church deemed that facts had proved it and therefore it became scientific. The Central Committee of the Soviet Communist Party in 1949 declared Mendelian genetics pseudoscientific and had its advocates, like Academician Vavilov, killed in concentration camps; after Vavilov’s murder Mendelian genetics was rehabilitated; but the Party’s right to decide what is science and publishable and what is pseudoscience and punishable was upheld. The new liberal Establishment of the West also exercises the right to deny freedom of speech to what it regards as pseudoscience, as we have seen in the case of the debate concerning race and intelligence. All these judgments were inevitably based on some sort of demarcation criterion. And this is why the problem of demarcation between science and pseudoscience is not a pseudo-problem of armchair philosophers: it has grave ethical and political implications.
Since 11 April 2023: 253 total views, 3 views today