• Welcome to Religious Forums, a friendly forum to discuss all religions in a friendly surrounding.

    Your voice is missing! You will need to register to get access to the following site features:
    • Reply to discussions and create your own threads.
    • Our modern chat room. No add-ons or extensions required, just login and start chatting!
    • Access to private conversations with other members.

    We hope to see you as a part of our community soon!

Mathematics: Formalisms or Physics? Mathematics and the nature of reality

LegionOnomaMoi

Veteran Member
Premium Member
A topic in another thread made me think of how easy it is to miss (without knowing it) so much that is vital for those interested in metaphysical or philosophical perspectives concerning the nature of reality (e.g., materialism, physicalism, dualism, etc.), the nature and philosophy of the sciences and of physics, and related topics. Specifically, the role mathematics plays in “fundamental” physical theories, from the well-established to the highly theoretical, is incredibly understanding reality, and is something that few physicists and far fewer mathematicians ever deal with. Most physicists don’t know much about quantum mechanics let alone the far more difficult quantum field theory (the heart of the standard model of particle physics), they aren’t theoretical physicists or cosmologists trying to create GUTs like string theory or M-theory, and are wont to view the mathematics used in “fundamental” physics as basically playing the same role mathematics did and does in Newtonian Mechanics or classical electromagnetism. Of course, scientists and non-scientists are even more likely to be unaware of the vital role mathematics plays in how we understand the most fundamental “layer(s)” of reality.

For example, in classical physics as in e.g., QM & QFT, we have notions like velocity, momentum, position, mass, etc. In other words, we study systems that have measurable/observable properties we call “observables”. Thus, in classical physics, when we say that a system has a mass of 1 kilogram or speed of 1000 mps, it’s clear how the values “1” or “1000” correspond to the system. True, we could change the values to various equivalents by e.g., expressing speed in terms of fps or mph, but whatever the value, there is a 1-to-1 (direct) relationship between that value and the observable (property) we wish it to represent.

This stops being true long before untested GUTs like M-theory. It is, in a very real sense, at the heart of quantum mechanics and in a certain way it IS QFT/particle physics. It is well-known that classical mechanics breaks down at very small (and large) levels, and also well-known that we can “reduce” classical mechanics to quantum mechanics but the reverse is not true. Unfortunately, when the creators of quantum theory developed quantum mechanics and QED (the earliest quantum field theory) in the early-to-mid 20th century, they retained the language of classical physics but changed the meanings of the terms used. I will use a simple description of an idealized, abstract, and simplified experiment in quantum mechanics to illustrate.

Even those who know very little about mathematics or physics are often aware that things like wave functions and the Schrödinger equation play a large role in QM. They may even know that the wave function is one way of expressing the state of a system in quantum mechanics. But they usually don’t know how physicists derive wave functions in any given experiment. To understand this requires a bare-bones description of such experiments, and the best I know of is from section II of Stapp’s paper The Copenhagen Interpretation (American Journal of Physics Vol. 40), which can also be found in section 3.2 of his Mind, Matter, and Quantum Mechanics (3rd Ed.): “Quantum theory is a procedure by which scientists predict probabilities that measurements of specified kinds will yield results of specified kinds in situations of specified kinds.” Concise? Yes. Informative? Hardly. But luckily there’s more:

“First, [the experimental physicist] transforms his information about the preparation of the system into an initial wave function. Then he applies to it some linear transformation, calculated perhaps from the Schrödinger equation, or perhaps from the S matrix, which converts the initial wave function into a final wave function. This final wave function, which is built on the degrees of freedom of the measured system, is then folded into the wave function corresponding to a possible result…

The essential points are that attention is focused on some system that is prepared in a specified manner and later examined in a specified manner. Quantum theory is a procedure for calculating the predicted probability that the specified type of examination will yield some specified result. This predicted probability is the predicted limit of the relative frequency of occurrence of the specified result, as the number of systems prepared and examined in accordance with the specifications goes to infinity.“

Now to unpack the above a bit. First, notice that this “procedure” which describes how quantum theory is used involves two wave functions (or their mathematical equivalents). It may seem like the “initial” one is similar to some mathematical description you might find in classical physics: you set up your system and use math to describe it (e.g., the mathematical description of the system’s state). But why do we need another wave function in order to describe the “measured system”, why isn’t the “measured system” the same as the system we prepared, and why doesn’t either wave function give us the state of any system? Second, what’s “relative frequency” and what does “goes to infinity” mean? The key to the first question is that, in QM, when we “prepare” a system we fundamentally disturb it-repeatedly. In fact, we keep disturbing it over and over again to give us a kind of “average” value of measurements that we collectively call “preparation”, and then transcribe this into a wave function that doesn’t correspond in any known way to whatever is left after we fundamentally disturbed the “system” by “preparing it. Since we no longer know what our wave function describes in terms of anything physical, we can’t predict anything unless we can relate the wave function to a very specific kind of later “measurement”. So specific, in fact, that we need to construct another “system” which doesn’t exist. Instead, it tells us that, given we disturbed X particular system in Y specified manner, we can use Z specifications to determine the probability that X, Y, and Z will give us the same result (this is the “relative frequency” part: imagine flipping a fair coin. The probability of getting heads given a fair coin flip is not 50% because given any n such tosses we will always get heads 50% of the time but because we would if we let n go to infinity. Likewise, given Y specifications of X with Z measurement specifications n times, we figure out which outcomes occur which percent of the time if we let n go to infinity).

That, in essence, is quantum theory. The problem, though, is that at no time have we used a mathematical description of our physical system(s) that corresponds to anything physical, but rather to wave functions we’ve mathematically derived from taking a “system”, messing it up repeatedly and describing the disturbances as “systems”, and this allows us to plug these mathematical systems into equations that yield the probabilities which makes QM perhaps the most successful theory of all time.

But it gets worse. As in classical physics, we describe “measurements” in terms of “observables”, and we label most of these “observables” the same way (e.g., “momentum”, “position”, etc.). However, in QM our “systems” are mathematical entities. Clearly, we can’t measure/observe the position or momentum of a mathematical function. So “observables” in QM are actually kinds of mathematical functions called operators. Thus the “momentum operator” doesn’t tell us the momentum of anything. It tells us that given a particular system “prepared” in a particular way and “measured” in a particular way, applying the “observable” operator (function) will yield a particular value with a particular probability (i.e., we can’t even just “apply” the operator and get a value we then call momentum, because we don’t get a value).

It gets even worse. We may not know what we are describing with QM, as our “systems” are mathematical and our “observable” properties are mathematical functions, but we at least we derived the mathematical structures experimentally. It’s not as if we started out with the mathematical structure/formalism of QM and then applied it; we started out with classical physics and gradually found through experiments that reality itself appears fundamentally “unsharp” or otherwise “vague” in numerous ways such that the only way to investigate reality at this level was by describing how forcing it to be “non-vague” obeys certain probabilistic tendencies.

This is not true of QFT, and therefore the whole standard model of particle physics. Quantum mechanics is not a relativistic physics, but special relativity has been confirmed experimentally (and Einstein formulated it in 1905 based on experimental findings). QM tells us that the energy of physical systems can fluctuate incredibly wildly. Consider a positron: it is “created” e.g., by a single electron scattering off of something (like a nucleus or proton or whatever) which causes the creation of the positron. Why? Because such an electron is energetic enough for special relativity to require an “antielectron” to conserve the energy of the system, and no Schrödinger equation can describe a system in which “energy” produces (is converted to) “mass” like this. For that we need relativistic quantum mechanics, and QFT is the most complete relativistic quantum theory (hence its status as the foundation of the standard model). But how did we develop the mathematical structures that tell us under what circumstances such a positron is produced? We took the mathematics of QM, took the mathematics of special relativity, and messed around with the mathematical structure of QM until it could incorporate that from special relativity. In other words, it wasn’t like QM stopped working the way classical physics did and we needed something more which we were able to derive experimentally. After all, the so-called “Copenhagen interpretation” dates from 1927, over 20 years after Einstein published his paper on special relativity, and even before developing his famous wave equation Schrödinger tried his hand at a relativistic equation. Everybody involved in the development of QM and its extensions knew this was a problem (and not the only one). The real issue is that we “fixed” the deficiencies of QM mathematically, not through experiments. Thus things like positrons weren’t discovered empirically, but are things we needed to balance equations. This is why we predicted the Higgs Boson and searched for it for years, CERN said they found it in 2012, and the research on what we found continues: we created the standard model mostly via mathematical extensions of a theory (QM) that deals with mathematical, not physical, systems.

So, what is the relationship between physical reality and mathematics? Well, with quantum mechanics we know that the things we call physical systems aren’t physical, but at least we are clear about the procedures we use, allowing a minimalist interpretation (the orthodox position in which QM is deemed irreducibly statistical), no matter how unsatisfactory many find this to be. Most of modern “fundamental” physics consists of extensions to QM (in which almost none of the particles of the standard model can exist), which we construct mathematically, not empirically. Basically, most of modern physics that involves the nature of reality treats reality as mathematical constructs. So much so, in fact, that information theory has been increasingly used and adopted as the only suitable language, context, and “ontological” framework for modern physics, because information theory is statistical, non-physical, and perfectly suited to describing “reality” as depicted in modern physics: non-physical.
 

beenherebeforeagain

Rogue Animist
Premium Member
A topic in another thread made me think of how easy it is to miss (without knowing it) so much that is vital for those interested in metaphysical or philosophical perspectives concerning the nature of reality (e.g., materialism, physicalism, dualism, etc.), the nature and philosophy of the sciences and of physics, and related topics. Specifically, the role mathematics plays in “fundamental” physical theories, from the well-established to the highly theoretical, is incredibly understanding reality, and is something that few physicists and far fewer mathematicians ever deal with. Most physicists don’t know much about quantum mechanics let alone the far more difficult quantum field theory (the heart of the standard model of particle physics), they aren’t theoretical physicists or cosmologists trying to create GUTs like string theory or M-theory, and are wont to view the mathematics used in “fundamental” physics as basically playing the same role mathematics did and does in Newtonian Mechanics or classical electromagnetism. Of course, scientists and non-scientists are even more likely to be unaware of the vital role mathematics plays in how we understand the most fundamental “layer(s)” of reality.

For example, in classical physics as in e.g., QM & QFT, we have notions like velocity, momentum, position, mass, etc. In other words, we study systems that have measurable/observable properties we call “observables”. Thus, in classical physics, when we say that a system has a mass of 1 kilogram or speed of 1000 mps, it’s clear how the values “1” or “1000” correspond to the system. True, we could change the values to various equivalents by e.g., expressing speed in terms of fps or mph, but whatever the value, there is a 1-to-1 (direct) relationship between that value and the observable (property) we wish it to represent.

This stops being true long before untested GUTs like M-theory. It is, in a very real sense, at the heart of quantum mechanics and in a certain way it IS QFT/particle physics. It is well-known that classical mechanics breaks down at very small (and large) levels, and also well-known that we can “reduce” classical mechanics to quantum mechanics but the reverse is not true. Unfortunately, when the creators of quantum theory developed quantum mechanics and QED (the earliest quantum field theory) in the early-to-mid 20th century, they retained the language of classical physics but changed the meanings of the terms used. I will use a simple description of an idealized, abstract, and simplified experiment in quantum mechanics to illustrate.

Even those who know very little about mathematics or physics are often aware that things like wave functions and the Schrödinger equation play a large role in QM. They may even know that the wave function is one way of expressing the state of a system in quantum mechanics. But they usually don’t know how physicists derive wave functions in any given experiment. To understand this requires a bare-bones description of such experiments, and the best I know of is from section II of Stapp’s paper The Copenhagen Interpretation (American Journal of Physics Vol. 40), which can also be found in section 3.2 of his Mind, Matter, and Quantum Mechanics (3rd Ed.): “Quantum theory is a procedure by which scientists predict probabilities that measurements of specified kinds will yield results of specified kinds in situations of specified kinds.” Concise? Yes. Informative? Hardly. But luckily there’s more:

“First, [the experimental physicist] transforms his information about the preparation of the system into an initial wave function. Then he applies to it some linear transformation, calculated perhaps from the Schrödinger equation, or perhaps from the S matrix, which converts the initial wave function into a final wave function. This final wave function, which is built on the degrees of freedom of the measured system, is then folded into the wave function corresponding to a possible result…

The essential points are that attention is focused on some system that is prepared in a specified manner and later examined in a specified manner. Quantum theory is a procedure for calculating the predicted probability that the specified type of examination will yield some specified result. This predicted probability is the predicted limit of the relative frequency of occurrence of the specified result, as the number of systems prepared and examined in accordance with the specifications goes to infinity.“

Now to unpack the above a bit. First, notice that this “procedure” which describes how quantum theory is used involves two wave functions (or their mathematical equivalents). It may seem like the “initial” one is similar to some mathematical description you might find in classical physics: you set up your system and use math to describe it (e.g., the mathematical description of the system’s state). But why do we need another wave function in order to describe the “measured system”, why isn’t the “measured system” the same as the system we prepared, and why doesn’t either wave function give us the state of any system? Second, what’s “relative frequency” and what does “goes to infinity” mean? The key to the first question is that, in QM, when we “prepare” a system we fundamentally disturb it-repeatedly. In fact, we keep disturbing it over and over again to give us a kind of “average” value of measurements that we collectively call “preparation”, and then transcribe this into a wave function that doesn’t correspond in any known way to whatever is left after we fundamentally disturbed the “system” by “preparing it. Since we no longer know what our wave function describes in terms of anything physical, we can’t predict anything unless we can relate the wave function to a very specific kind of later “measurement”. So specific, in fact, that we need to construct another “system” which doesn’t exist. Instead, it tells us that, given we disturbed X particular system in Y specified manner, we can use Z specifications to determine the probability that X, Y, and Z will give us the same result (this is the “relative frequency” part: imagine flipping a fair coin. The probability of getting heads given a fair coin flip is not 50% because given any n such tosses we will always get heads 50% of the time but because we would if we let n go to infinity. Likewise, given Y specifications of X with Z measurement specifications n times, we figure out which outcomes occur which percent of the time if we let n go to infinity).

That, in essence, is quantum theory. The problem, though, is that at no time have we used a mathematical description of our physical system(s) that corresponds to anything physical, but rather to wave functions we’ve mathematically derived from taking a “system”, messing it up repeatedly and describing the disturbances as “systems”, and this allows us to plug these mathematical systems into equations that yield the probabilities which makes QM perhaps the most successful theory of all time.

But it gets worse. As in classical physics, we describe “measurements” in terms of “observables”, and we label most of these “observables” the same way (e.g., “momentum”, “position”, etc.). However, in QM our “systems” are mathematical entities. Clearly, we can’t measure/observe the position or momentum of a mathematical function. So “observables” in QM are actually kinds of mathematical functions called operators. Thus the “momentum operator” doesn’t tell us the momentum of anything. It tells us that given a particular system “prepared” in a particular way and “measured” in a particular way, applying the “observable” operator (function) will yield a particular value with a particular probability (i.e., we can’t even just “apply” the operator and get a value we then call momentum, because we don’t get a value).

It gets even worse. We may not know what we are describing with QM, as our “systems” are mathematical and our “observable” properties are mathematical functions, but we at least we derived the mathematical structures experimentally. It’s not as if we started out with the mathematical structure/formalism of QM and then applied it; we started out with classical physics and gradually found through experiments that reality itself appears fundamentally “unsharp” or otherwise “vague” in numerous ways such that the only way to investigate reality at this level was by describing how forcing it to be “non-vague” obeys certain probabilistic tendencies.

This is not true of QFT, and therefore the whole standard model of particle physics. Quantum mechanics is not a relativistic physics, but special relativity has been confirmed experimentally (and Einstein formulated it in 1905 based on experimental findings). QM tells us that the energy of physical systems can fluctuate incredibly wildly. Consider a positron: it is “created” e.g., by a single electron scattering off of something (like a nucleus or proton or whatever) which causes the creation of the positron. Why? Because such an electron is energetic enough for special relativity to require an “antielectron” to conserve the energy of the system, and no Schrödinger equation can describe a system in which “energy” produces (is converted to) “mass” like this. For that we need relativistic quantum mechanics, and QFT is the most complete relativistic quantum theory (hence its status as the foundation of the standard model). But how did we develop the mathematical structures that tell us under what circumstances such a positron is produced? We took the mathematics of QM, took the mathematics of special relativity, and messed around with the mathematical structure of QM until it could incorporate that from special relativity. In other words, it wasn’t like QM stopped working the way classical physics did and we needed something more which we were able to derive experimentally. After all, the so-called “Copenhagen interpretation” dates from 1927, over 20 years after Einstein published his paper on special relativity, and even before developing his famous wave equation Schrödinger tried his hand at a relativistic equation. Everybody involved in the development of QM and its extensions knew this was a problem (and not the only one). The real issue is that we “fixed” the deficiencies of QM mathematically, not through experiments. Thus things like positrons weren’t discovered empirically, but are things we needed to balance equations. This is why we predicted the Higgs Boson and searched for it for years, CERN said they found it in 2012, and the research on what we found continues: we created the standard model mostly via mathematical extensions of a theory (QM) that deals with mathematical, not physical, systems.

So, what is the relationship between physical reality and mathematics? Well, with quantum mechanics we know that the things we call physical systems aren’t physical, but at least we are clear about the procedures we use, allowing a minimalist interpretation (the orthodox position in which QM is deemed irreducibly statistical), no matter how unsatisfactory many find this to be. Most of modern “fundamental” physics consists of extensions to QM (in which almost none of the particles of the standard model can exist), which we construct mathematically, not empirically. Basically, most of modern physics that involves the nature of reality treats reality as mathematical constructs. So much so, in fact, that information theory has been increasingly used and adopted as the only suitable language, context, and “ontological” framework for modern physics, because information theory is statistical, non-physical, and perfectly suited to describing “reality” as depicted in modern physics: non-physical.
Thanks for the interesting primer!:cool:
 

Brickjectivity

Veteran Member
Staff member
Premium Member
So, as I understand what you are saying, Quantum Field Theory is fairly robust compared with Quantum Mechanics? It seems you also said QFT was part of QM, however. Then QM is mostly a Mathematical structure to explain certain observations and predict particles, while QFT is a more solid nugget of theory, which everything in QM has to conform with?

There is something else you said that I wonder about. You said that Classical Mechanics can be explained in terms of Quantum Mechanics but not the other way around. There was an article I read here (archived link ) which I wanted to ask about, because it says one of the intriguing quantum effects (particle duality) has been modeled with silk droplets. So I am wondering if that relates to what you were saying, and if it means that the physical reality behind Quantum Mechanics might be some variant of classical physics lying beneath it and could be modeled that way? I like that the models are driven by observations, what little can be gotten. I feel like our ability to observe quanta are very poor, what with having to spend billions to build accelerators like CERN. We have to so do much just to smash things together, so can we conclude definitely that we cannot model quantum mechanics using classical or is just that the data we have don't confirm it? Maybe there is just a limit to what we can find out?
 

Yerda

Veteran Member
Although I'll have to read it again, I found that illuminating. I'll be back with questions in the next few days.
 

Laika

Well-Known Member
Premium Member
Many thanks for this. I will have to read it again to grasp it.

As a materialist and non-scientist, alot of my ideas clash with scientific theories, but I simply don't have the background to know whether to trust my own philosophy or defer to the superior knowledge of scientists. it's quite intimidating and I'm left feeling dirty as some kind of "creationist" sticking to scripture or dogma when I really love science and genuinely want answers.

Please, Please, Please, do more of these! The slow kids at the back of the class really appreciate it! :D
 

LegionOnomaMoi

Veteran Member
Premium Member
So, as I understand what you are saying, Quantum Field Theory is fairly robust compared with Quantum Mechanics?
I suppose it all depends upon what you mean by robust. QM presents less of a break between reality and mathematics than QFT or quantum chromodynamics or the standard particle or M-theory and so forth. The problem is (and this is the sense in which QFT is, as you put it, more robust) we know that QM is not the whole story. It cannot account for experimental findings that predate QM which Einstein first coherently and consistently formulated in 1905 and which we refer to as special relativity.
It seems you also said QFT was part of QM, however.
I would say its the other way around. QM is incorporated into QFT in order to (among other things) allow for a relativistic "version" of QM.

Then QM is mostly a Mathematical structure to explain certain observations and predict particles, while QFT is a more solid nugget of theory, which everything in QM has to conform with?
For simplicity, I'm leaving out aspects of the standard model and QFT that QM can't "deal" with beyond special relativity, but there are aspects of the standard model/quantum field theories such as quantum chromodynamics which QM can't "deal" with but the standard model/QFT can. In that sense the standard model (and by extension QFT, upon which it is founded) is a "more solid nugget of theory".

The problem is that we didn't arrive at the more "complete" or "powerful" theories that allow QFT to "deal" with things that QM can't through experiments and empirical findings (at least in general). Instead, we fiddled with equations to make disparate theories "fit". For example, as the Schrödinger equation or its equivalents in QM, no matter how many variables we plug into it or whatever algebraic manipulations we employ, can't ever describe a system that special relativity tells us should behave in a particular way (in particular, conserve energy by the creation or annihilation of "particles"). We need something that does what the Schrödinger equation does but that CAN "describe" processes demanded thanks to special relativity. Normally, when a theory in the sciences can't account for phenomena that those in relevant fields agree it must, we fiddle around with experiments in order to come up with a better theory. In modern physics, this didn't happen. We fiddled around with the math (e.g., making QM relativistic by changing its mathematical structure(s), equations, etc.), and called the resulting mathematical manipulations physical reality. Then we took this mathematical framework we derived formally by combining distinct theories (individually empirically supported but not collectively) and we used it to design experiments and experimental designs in order to use the mathematical framework to interpret the results of the experiments we only performed because of that framework. More simply (and overly simplistically), we went looking for things like positrons because the math said we needed them, but as the math is as close as we can get to "physical reality", any experiments depend upon interpreting the mathematical framework to have shown what we created it to be able to show in the first place.

You said that Classical Mechanics can be explained in terms of Quantum Mechanics but not the other way around.
I did. That's actually another simplification. It's true that QM doesn't breakdown the way that classical mechanics does in that at a certain level all classical mechanics fails (the simplest way to see this is to recognize that according to classical mechanics the gravitational pull between the nucleus and electron(s) of every atom in the universe predicts that every electron in the universe would plunge almost instantaneously into the nucleus that electron is supposed to "orbit", meaning that no electrons or nuclei could exist in classical mechanics). However, it's also true that not only is there no clear divide between classical and quantum physics and some of the deficiencies of QM due to the fact that classical physics demands a theory that can account for phenomena such as classical electrodynamics in ways that even QED (quantum electrodynamics) can't . I chose the following paper as an example because not only is it available for full in free but I need only quote from the easily found abstract and it is sufficient: "Quantum electrodynamics presents intrinsic limitations in the description of physical processes that make it impossible to recover from it the type of description we have in classical electrodynamics. Hence one cannot consider classical electrodynamics as reducing to quantum electrodynamics and being recovered from it by some sort of limiting procedure. Quantum electrodynamics has to be seen not as a more fundamental theory, but as an upgrade of classical electrodynamics, which permits an extension of classical theory to the description of phenomena that, while being related to the conceptual framework of the classical theory, cannot be addressed from the classical theory."
Valente, M. B. (2011). The Relation between Classical and Quantum Electrodynamics. THEORIA. An International Journal for Theory, History and Foundations of Science, 26(1), 51-68.

So I am wondering if that relates to what you were saying
Actually yes, although not in the ways one might immediately think of (including me, as I've known about Couder's work for a while and I think I may have even mentioned or responded to something about it here). I'm deeply interested in complex systems, in particular living systems (and in particular the brain). Systems biology and related approaches to biology (e.g., relational biology) sometimes rely on models of living systems that aren't compatible with either quantum or classical physics. For example, in addition to the very controversial quantum consciousness theories proposed by those like Stapp and Penrose, there are theories/models of quantum-like consciousness. These are models which use the mathematics of quantum mechanics to describe the dynamics of the brain (as well as finance, see e.g., Khrennikov, A. Y. (2014). Ubiquitous quantum structure. Springer.) without claiming that actual quantum processes take place. The founder of relational biology, mathematical biologist Robert Rosen, formally demonstrated that living systems cannot be described using physics as it exists period because they defy reductionism at the macroscale. Thus there are many phenomena that we are currently investigating in various sciences that seem to act in ways that quantum systems due but that quantum physics doesn't "allow" for, or in ways that NO physics currently allows for. In addition, the use of the mathematical structure to describe "classical" systems is another instance of the ways in which "physical systems" are increasingly reduced to mathematics, not constituent parts and their interactions.

Maybe there is just a limit to what we can find out?
I would say there must be, and not just because of the uncertainty principle. As someone who DOESN'T think that reality reduces to mathematics, though, I would add that there seems to be no limit to what we are willing to posit we can find out or even HAVE found out.
 
Last edited:

Thief

Rogue Theologian
A topic in another thread made me think of how easy it is to miss (without knowing it) so much that is vital for those interested in metaphysical or philosophical perspectives concerning the nature of reality (e.g., materialism, physicalism, dualism, etc.), the nature and philosophy of the sciences and of physics, and related topics. Specifically, the role mathematics plays in “fundamental” physical theories, from the well-established to the highly theoretical, is incredibly understanding reality, and is something that few physicists and far fewer mathematicians ever deal with. Most physicists don’t know much about quantum mechanics let alone the far more difficult quantum field theory (the heart of the standard model of particle physics), they aren’t theoretical physicists or cosmologists trying to create GUTs like string theory or M-theory, and are wont to view the mathematics used in “fundamental” physics as basically playing the same role mathematics did and does in Newtonian Mechanics or classical electromagnetism. Of course, scientists and non-scientists are even more likely to be unaware of the vital role mathematics plays in how we understand the most fundamental “layer(s)” of reality.

For example, in classical physics as in e.g., QM & QFT, we have notions like velocity, momentum, position, mass, etc. In other words, we study systems that have measurable/observable properties we call “observables”. Thus, in classical physics, when we say that a system has a mass of 1 kilogram or speed of 1000 mps, it’s clear how the values “1” or “1000” correspond to the system. True, we could change the values to various equivalents by e.g., expressing speed in terms of fps or mph, but whatever the value, there is a 1-to-1 (direct) relationship between that value and the observable (property) we wish it to represent.

This stops being true long before untested GUTs like M-theory. It is, in a very real sense, at the heart of quantum mechanics and in a certain way it IS QFT/particle physics. It is well-known that classical mechanics breaks down at very small (and large) levels, and also well-known that we can “reduce” classical mechanics to quantum mechanics but the reverse is not true. Unfortunately, when the creators of quantum theory developed quantum mechanics and QED (the earliest quantum field theory) in the early-to-mid 20th century, they retained the language of classical physics but changed the meanings of the terms used. I will use a simple description of an idealized, abstract, and simplified experiment in quantum mechanics to illustrate.

Even those who know very little about mathematics or physics are often aware that things like wave functions and the Schrödinger equation play a large role in QM. They may even know that the wave function is one way of expressing the state of a system in quantum mechanics. But they usually don’t know how physicists derive wave functions in any given experiment. To understand this requires a bare-bones description of such experiments, and the best I know of is from section II of Stapp’s paper The Copenhagen Interpretation (American Journal of Physics Vol. 40), which can also be found in section 3.2 of his Mind, Matter, and Quantum Mechanics (3rd Ed.): “Quantum theory is a procedure by which scientists predict probabilities that measurements of specified kinds will yield results of specified kinds in situations of specified kinds.” Concise? Yes. Informative? Hardly. But luckily there’s more:

“First, [the experimental physicist] transforms his information about the preparation of the system into an initial wave function. Then he applies to it some linear transformation, calculated perhaps from the Schrödinger equation, or perhaps from the S matrix, which converts the initial wave function into a final wave function. This final wave function, which is built on the degrees of freedom of the measured system, is then folded into the wave function corresponding to a possible result…

The essential points are that attention is focused on some system that is prepared in a specified manner and later examined in a specified manner. Quantum theory is a procedure for calculating the predicted probability that the specified type of examination will yield some specified result. This predicted probability is the predicted limit of the relative frequency of occurrence of the specified result, as the number of systems prepared and examined in accordance with the specifications goes to infinity.“

Now to unpack the above a bit. First, notice that this “procedure” which describes how quantum theory is used involves two wave functions (or their mathematical equivalents). It may seem like the “initial” one is similar to some mathematical description you might find in classical physics: you set up your system and use math to describe it (e.g., the mathematical description of the system’s state). But why do we need another wave function in order to describe the “measured system”, why isn’t the “measured system” the same as the system we prepared, and why doesn’t either wave function give us the state of any system? Second, what’s “relative frequency” and what does “goes to infinity” mean? The key to the first question is that, in QM, when we “prepare” a system we fundamentally disturb it-repeatedly. In fact, we keep disturbing it over and over again to give us a kind of “average” value of measurements that we collectively call “preparation”, and then transcribe this into a wave function that doesn’t correspond in any known way to whatever is left after we fundamentally disturbed the “system” by “preparing it. Since we no longer know what our wave function describes in terms of anything physical, we can’t predict anything unless we can relate the wave function to a very specific kind of later “measurement”. So specific, in fact, that we need to construct another “system” which doesn’t exist. Instead, it tells us that, given we disturbed X particular system in Y specified manner, we can use Z specifications to determine the probability that X, Y, and Z will give us the same result (this is the “relative frequency” part: imagine flipping a fair coin. The probability of getting heads given a fair coin flip is not 50% because given any n such tosses we will always get heads 50% of the time but because we would if we let n go to infinity. Likewise, given Y specifications of X with Z measurement specifications n times, we figure out which outcomes occur which percent of the time if we let n go to infinity).

That, in essence, is quantum theory. The problem, though, is that at no time have we used a mathematical description of our physical system(s) that corresponds to anything physical, but rather to wave functions we’ve mathematically derived from taking a “system”, messing it up repeatedly and describing the disturbances as “systems”, and this allows us to plug these mathematical systems into equations that yield the probabilities which makes QM perhaps the most successful theory of all time.

But it gets worse. As in classical physics, we describe “measurements” in terms of “observables”, and we label most of these “observables” the same way (e.g., “momentum”, “position”, etc.). However, in QM our “systems” are mathematical entities. Clearly, we can’t measure/observe the position or momentum of a mathematical function. So “observables” in QM are actually kinds of mathematical functions called operators. Thus the “momentum operator” doesn’t tell us the momentum of anything. It tells us that given a particular system “prepared” in a particular way and “measured” in a particular way, applying the “observable” operator (function) will yield a particular value with a particular probability (i.e., we can’t even just “apply” the operator and get a value we then call momentum, because we don’t get a value).

It gets even worse. We may not know what we are describing with QM, as our “systems” are mathematical and our “observable” properties are mathematical functions, but we at least we derived the mathematical structures experimentally. It’s not as if we started out with the mathematical structure/formalism of QM and then applied it; we started out with classical physics and gradually found through experiments that reality itself appears fundamentally “unsharp” or otherwise “vague” in numerous ways such that the only way to investigate reality at this level was by describing how forcing it to be “non-vague” obeys certain probabilistic tendencies.

This is not true of QFT, and therefore the whole standard model of particle physics. Quantum mechanics is not a relativistic physics, but special relativity has been confirmed experimentally (and Einstein formulated it in 1905 based on experimental findings). QM tells us that the energy of physical systems can fluctuate incredibly wildly. Consider a positron: it is “created” e.g., by a single electron scattering off of something (like a nucleus or proton or whatever) which causes the creation of the positron. Why? Because such an electron is energetic enough for special relativity to require an “antielectron” to conserve the energy of the system, and no Schrödinger equation can describe a system in which “energy” produces (is converted to) “mass” like this. For that we need relativistic quantum mechanics, and QFT is the most complete relativistic quantum theory (hence its status as the foundation of the standard model). But how did we develop the mathematical structures that tell us under what circumstances such a positron is produced? We took the mathematics of QM, took the mathematics of special relativity, and messed around with the mathematical structure of QM until it could incorporate that from special relativity. In other words, it wasn’t like QM stopped working the way classical physics did and we needed something more which we were able to derive experimentally. After all, the so-called “Copenhagen interpretation” dates from 1927, over 20 years after Einstein published his paper on special relativity, and even before developing his famous wave equation Schrödinger tried his hand at a relativistic equation. Everybody involved in the development of QM and its extensions knew this was a problem (and not the only one). The real issue is that we “fixed” the deficiencies of QM mathematically, not through experiments. Thus things like positrons weren’t discovered empirically, but are things we needed to balance equations. This is why we predicted the Higgs Boson and searched for it for years, CERN said they found it in 2012, and the research on what we found continues: we created the standard model mostly via mathematical extensions of a theory (QM) that deals with mathematical, not physical, systems.

So, what is the relationship between physical reality and mathematics? Well, with quantum mechanics we know that the things we call physical systems aren’t physical, but at least we are clear about the procedures we use, allowing a minimalist interpretation (the orthodox position in which QM is deemed irreducibly statistical), no matter how unsatisfactory many find this to be. Most of modern “fundamental” physics consists of extensions to QM (in which almost none of the particles of the standard model can exist), which we construct mathematically, not empirically. Basically, most of modern physics that involves the nature of reality treats reality as mathematical constructs. So much so, in fact, that information theory has been increasingly used and adopted as the only suitable language, context, and “ontological” framework for modern physics, because information theory is statistical, non-physical, and perfectly suited to describing “reality” as depicted in modern physics: non-physical.
kinda like saying.... I prefer the realm of numbers over the sidewalk beneath my feet.
 
Top