Читайте также: |
|
Several phases can be distinguished in the development of the SP. Some of these
are briefly described in what follows.
The first phase starts with Galileo and is the application of experimentation
and speculation to investigate the physical world under certain methodological
restrictions. This phase reaches its first climax with Newton’s theory of
gravitation and the establishment of mechanics as the primary model for natural
phenomena. During this phase the notion of determinism is prevalent. Mechanics
develops by physicists like Leibniz, Maupertuis, Lagrange, Hamilton and
Laplace in the form of classical mechanics, with celestial mechanics as its crown
achievement. The mathematical formulation goes under the name of analytical
mechanics.
A second phase begins with the realisation that there are non-deterministic
aspects that need to be taken into account for the description of phenomena in
nature. Maxwell and Darwin publish the same year, 1859, scientific works that
use chance and randomness as tools for modelling processes in nature. This
quickly develops into statistical mechanics by Boltzmann, Gibbs and others and
in biology into neo-Darwinism after the 1950s.
A third phase begins with relativity theory and quantum mechanics in the first
decades of the 20th century. During this phase, in which the atomic structure of
matter is established, one is forced to give up classical realism. In the theory of
relativity the notions of space and time being some independent background, as
conceived by Newton, is totally changed into a four-dimensional space-time
continuum where time and space mix, implying a change in the notion of
simultaneity, and in general relativity both space and time are dynamical
variables that interact with matter.
A fourth phase starts after World War II with the development of biology as a
form of micro mechanics. The discovery of the DNA alpha-helix structure and
the structure of proteins and their use for the description of the theory of
evolution marks its entry on the scene.
A fifth phase starts with Hubble’s discovery in 1925 of the expansion of the
universe, the development of elementary particle physics and modern cosmology
leading to the attempt at formulating a theory of everything (TOE).
In this greatly simplified scheme, in the face of which many physicists might
feel unhappy not to find themselves clearly represented (like solid state physics
which is included in quantum physics/mechanics) there is a clear trend from
determinism, that is mechanical and without purpose, to the addition of
randomness, that is also without visible purpose, to the implementation of the
mechanistic scheme to describe all of nature - classical mechanics, celestial
mechanics, solid mechanics, fluid mechanics, continuum mechanics, quantum
mechanics etc. - including modern biology. There are of course good reasons for
this this course of events. The reasons are to be found in the new notion of
knowledge that is introduced in the SP and that will be elucidated below.
III. What is needed to establish a scheme of knowledge?
Many scientists are sceptical to metaphysics and of its possible relevance to
science. This is in my opinion a mistake, since in any serious self-reflection on
ones practice in science such aspects enter. To put up a scheme of knowledge we
need at least three elements:
1.An aim for the knowledge,
2.A set of criteria to validate a knowledge claim according to the aim,
3.A common practice to obtain the same, or at least as similar as possible,
experiences for us to reach consensus on what the basic observations or so called
facts should be.
The aim determines the validation criteria and will influence the practice. In the
SP the archetype of the common practice is the work in the laboratory or the
observatory. This includes the understanding of the operation and use of the
instruments that extend our senses into the micro or the macro world.
a. An aim for the knowledge
Some people would argue that there is no need for an aim with knowledge. We
seek knowledge for its own sake and more knowledge is better than less. In fact
they would probably argue that they simply seek the truth about nature. The
problem is not only that even this is an aim, but that we have not the slightest
clue as to what could represent the truth. Our senses play tricks with us all the
time. It is in order to diminish this problem that both Galileo and after him
Descartes insist on that science should occupy itself with only the so called primary qualities of matter, like length, time, mass, velocity etc. The secondary
qualities such as smell, colour, taste etc. are understood to depend upon our
senses and should be avoided. Even so it is not possible to formulate criteria for
a knowledge claim that pretends to aim for truth, since we don’t know how truth
would manifest itself for us, even in the laboratory or the observatory, and not
even if we stick to the primary qualities of matter. The history of physical
theories, and how they are turned over, presents a clear illustration of this. As
soon as we formulate a subordinate aim for the knowledge we seek, we can set
up a set of criteria for it. The result is then a series of maps of nature in
accordance with the aim.
Several schemes of knowledge have been set up in the past. Not all of them
were complete in all three aspects. However Aristotle devised a rather complete
scheme. His aim was to find and establish the foreordained hierarchical order of
everything in the universe. To this end he required information of four “causes”:
the material, formal, efficient and final cause. Only then do we, according to
Aristotle, have knowledge and understanding of a thing in relation to the
universe. For him basically all human experiences could be used as common
practice although even the Greeks understood the problem with the subjectivity
of the senses. It seems to be the lack of a common practice that to some extent
stops the Greeks from advancing in a systematic way into science as we
understand it today. We therefore should denote the Greek science as the proto scientific
phase.
Francis Bacon suggested that the aim of the Scienza Nova, the new science, the
new knowledge of nature, should be to obtain power over nature. He wanted
science “to bring relief of man's estate“ and to use science to develop utilities for
man. He then realised that brooding into possible aims with nature stood in the
way of this cause. “Inquiry into final causes is sterile, and like a virgin
consecrated to God, produces nothing” (Bacon 1623) was his incisive
formulation of the predicament we were into. Respecting nature as God’s
creation would stop man from exploiting her. Therefore final aims should be
avoided. Rene Descartes was of the same opinion: final causes are to be
banished from science. Ever since, scientists have adhered to this view, and
decided to study nature as if devoid of intention or meaning. This prevails into
our times.
Modern science has reformulated the aim slightly into a less provocative one:
the aim is to control and obtain power over nature and to predict the future.
Power obviously makes you able to predict the future to some extent. The idea of
predictability then comes into the foreground. The criteria for knowledge
validation are then subject to this aim.
b. A set of criteria to validate a knowledge claim according to the aim
Which processes can be easily predicted? Clearly the repeatable ones! We
therefore insist upon the study of repeatable events and that each experiment
should be possible to repeat several times. We also seek to establish cause-effect
relations in the processes in nature, essentially the efficient causes of Aristotle.
The prototype for this is the relation between the set up of the experiment and its
outcome. The set up is the cause and the outcome the effect of the cause. In this
way each effect could be the cause of a new cause-effect relation in a chain ad
infinitum. Normally it is believed that this chain can be cut at an arbitrary place
and then from this place on the new situation is the cause of the following
events. We do not need to consider any memory of the past. We also limit
ourselves to what is called locality in space and time: all causes should be found
in a limited region in space and time, otherwise the whole universe would be a
potential cause including all its history. This is obviously not possible to have
access to.
In the experimental process there are things that cannot be changed and
altered. These aspects are referred to as universal laws, like the law of
gravitation or Newton’s law, and constants of nature, such as the value of the
electric elementary charge, the gravitational constant, the speed of light in
vacuum, etc. The other circumstances in the cause are referred to as boundary
conditions. It is the experimentalist that must find out how much of the boundary
conditions that can be varied and how much relevant information that we need to
control to perform the experiment in a successful and repeatable way within the
given accuracy of measurement. The theory is then suggested as soon as enough
data are collected. Too few data do not in general make a good starting material
for suggesting a theory. Too many theories can be suggested.
Francis Bacon suggested the inductive method to be used in science: by
measuring more and more accurately we would finally find out the laws of
nature. This strategy has largely been abandoned in favour of the complementary
strategy to guess or hypostatize a law and then subjects it to tests. This
hypothetic-deductive method has been highly successful.
It now becomes clear how the insistence upon determinism enters at the
inception of the project in order to ensure predictability. With determinism we
mean one-to-one mappings. Given a certain situation, another unique and
calculable situation should be reached in the next step in time, within the limits
of experimental uncertainties. In general there are also one-to-many, many-toone
and many-to-many mappings. In quantum mechanics one-to-many situations
are the general case. This diminishes the predictability, and we can only predict
the statistical outcome of many experiments. Still, for technological applications,
this often turns out to be enough.
We also use a principle called Occam's razor: we do not introduce more
hypotheses than necessary to suggest at theory for the phenomena. The least
number of assumptions that give an adequate description of the data are adopted.
The theory is then subjected to further tests of its predictions.
An important element in the scientific process is that we communicate our
findings in an honest manner as to what we have found relevant for the
experiment. It should in principle be possible for another scientist to repeat the
experiment.
Again, final causes are not supposed to be used as elements in the description
or explanation of an experiment. The material causes of Aristotle are of course
there as well as the formal causes. Since nature sometimes, if not always, show
up one-to-many situations, the idea to use chance to model them is in line with
the effort to avoid "final causes". Chance can by definition not have any final
cause or meaning.
In quantum mechanics even the notion of classical realism had to be
abandoned. Quanta cannot be assigned arbitrarily exact locations and velocities
simultaneously according to Heisenberg's uncertainty principle. This implies e.g.
that electrons do not move in orbits around the atomic nucleus. We cannot in a
classical way picture how they move, or rather behave. We can only predict the
possibility to find them at various positions or with various momenta (= mass
times velocity) at a certain time. “It is wrong to think that the task of physics is
to find out how nature is. Physics concerns what we can say about nature” was
Niels Bohr’s conclusion of the situation.
In biology Darwin's theory of evolution is reformulated in modern times by the
notion of DNA, as the carrier of the information of the individual. The model
states that random mutations (of DNA) and the impact of the
physical/biological/geological environment, providing a “natural selection” of
the fittest, are the forces that drive the evolution. From this it follows that
evolution has no direction and no aim or purpose. This is built into the model.
c. A common practice
As to the third element in the structure of knowledge we use rational thinking
and study cause effect relations in the laboratory or observatory. With rational
thinking is here basically meant mathematical modelling. This is important,
since mathematics is the only contradictory free language we know. Once
relations between observable quantities have been formulated in mathematical
terms, we usually agree on these relations. An example is given by quantum
mechanics. Most physicists agree on the mathematical formulation of quantum
mechanics and how to compute various predictions with it. However, as Max
Jammer (1966) has pointed out there are a large number of different ways to
interpret what the mathematical formulation means in ordinary words.
Science has since the time of Galileo been understood as a process of
measurements of the primary qualities of matter. However, as time goes on we
can see that it is rather the experimental situation that characterizes science. Of
course everything is still a kind of measurement: measurement relative to some
standard of the primary qualities of matter. Nevertheless the configuration of the
experimental set up is a decisive aspect of the experimental situation and of the
interpretation of what the measurement says. This becomes even clearer in
quantum mechanics, where the measuring device and the object become
entangled after their interaction in a measuring situation.
In each discipline it takes time to learn to perform experiments and understand
the workings of the instruments or even suggest the construction of new
instruments for various investigations. It is not always easy for the untrained eye
or mind to use scientific instrument and to understand what they show. A well
known example is the astronomical tube of Galileo, which even contemporary
philosophers refused to look into. Another example is the microscope. When
freshmen look on biological specimens in the microscope, they rarely can see
anything useful and clear to being with. Only after learning what to look for and
how to complete the fragments of the picture into an intelligible form can they
discern what the specimen is showing. Similar difficulties are presented with
radiography.
In the theory building we also require that the new theory is consistent, within
the experimental error limits, with older established findings. This most often
means that when the situation is adequate for the old theory to hold, then both
the old and the new theory should give the same predictions within the error
limits.
Above, I have concentrated on the physical side of the SP at the expense of in
particular chemistry. Chemistry has been very important for the development of
the structure of matter, both in the way its structural simplification has been
worked out to base chemistry on the stable basic elements in the periodic table,
but also in the special way the carbon chemistry has been worked out in organic
chemistry. Also the chemistry of gases was instrumental in establishing the
theory of atoms. In quantum chemistry the science of chemistry and physics
meet. In addition to chemistry, I have left out the other sciences like geology,
meteorology, etc. But they can be seen to rely to a large extent on the findings in
the other sciences as applied to various domains of interest.
Дата добавления: 2015-11-14; просмотров: 54 | Нарушение авторских прав
<== предыдущая страница | | | следующая страница ==> |
Role definition | | | IV. Consequences of the SP 1 |