Did we "make up" the geometry of the electron gun? The probabilities?
Yes. It began by resolving the paradoxical results of two conflicting interpretations by incorporating the deterministic mathematics of classical physics into some hybrid which could explain the results of an experiment. Then came Feynman, who said that our results (the pattern) wasn't just a matter of what we were detecting based on "where" electrons or photons or whatever went, but was simply wrong. We were "observing" by making slits in the first place, which determined what we would find by changing the state of the system before we observed it. The "path integral" means that every pattern we detect is determined by changing the states of the system in a way not reflected in the wavefunctions, the initial specifications, the initial "state", or the final state. The final state of the actual trajectories are all possible routes through the screen and all points on the detector. And Wheeler introduced the idea that even if we tried to "undetermine" the paths by "observing" after the electron or whatever actually took some path, we would never find the "superposition state" or anything "quantum". And with new sophisticated set-ups, we can actually implement Wheeler's "delayed-choice" experiment. He have. Over and over again, in different waves. And each time we find the same thing: we can detect confliciting "states" at the same time, and there is no theoretical limit to how many different conflicting states we could detect at any given time.
we already know there is wave behaviour.
Which we determine. And there is "particle" behavior. Which we also determine. And we can find both, at the same time, by determining them. Which means that every experiment means an unknown margin of error between initial and final states, because each are determined by every part of the experiment, and all we have our inferences of this indirect treatment of an indeterministic system we treat deterministically.
I also don't see how we've "altered" the system by firing an electron down an electron gun.
Because we never describe this. You quoted wikipedia before about the "states" of the system. How our specifications used to transcribe the "system" into a wavefunction such that we have anything to say about whatever it is we "fire"?
You don't know how, because you're denying the realism of quantum mechanics. We end up with the probabilities we do because the wave function dictates we do.
What "realism"? According to the "many-worlds" interpretation, there are no probabilities. We always get every possible result, but we only end up with the single state in our universe. Of course, because that state is determined in advance by the specifications used transcribing the wavefunction, how did we get it out of some infinite possible universes in which we somehow made any "meaningful" measurement which allows us to say anything at all?
We know lots about the initial state - for instance, the particle's "direction" (as much as that makes sense in this context) approximate energy, and rest mass. Do we need anything else?
No. But it we be nice to have any of the above. Because we don't.
3D probability wavefunctions don't take defined paths, so isn't that the expected result?
A probability function describes the probability of something happening. If the wave function describes the actual system, it isn't a probability function. It's a nonlocal system in multiple places at once taking multiple trajectories. The trajectories we say it takes correspond to the way in which we transcribed the "system" to get an initial state and the way in which we determined the trajectories ahead of time, contradicting quantum theory itself.
A macroscopic quantum result is not a counterexample to deriving classical mechanics from quantum. A counterexample would be a situation where classical and quantum mechanics disagree, and classical is correct. Would you like to cite such a situation?
Sure. As soon as you tell me what either means. "Classical" mechanics is the name we give to an approach which we found stopped working at some "fuzzy" point, and not only didn't work, but was completely contrary to our entire framework. "Quantum" mechanics is the framework we use to describe experiments which contradict the actual theory itself. And when I cited a peer-reviewed journal article in which the author stated this, you said he was wrong because wikipedia talked about quantum states.
This is the measurement paradox: the process of measurement cannot be described by standard quantum dynamics."
Ellis, G. F. (2012). On the limits of quantum theory: Contextuality and the quantumclassical cut. Annals of Physics.
You ask for citations, but then you ignore them, so why ask?
Except for the probabilistic element of which value you end up with from your measurement.
Which is classical determinism. There's always a "margin for error', and we determine these in quantum systems by pretending we have classical systems.
Inference leads us to believe, as shown by Feynman, Wheeler, and everyone since Bohm, Heisenberg, & Bohr, that we quantum systems cannot be observed. Any observation determines the result. Which means we cannot start with known quantum state, because the only way to know it is to be have some way of observing it. But doing so destroys the quantum system. So we invent the state, and use the same inference which informs us that quantum reality is fundamentally, ontologically, and "really" indeterministic, to develop mathematical methods which allow us to ignore our own theory and perform experiments which violate the theoretical framework of quantum theory.
The universe is closed by definition, so no.
The indeterminism appears when your quantum system is not closed - unlike the universe - and includes an interface to a classical, measuring object. It is only in the ontology of this classical, 4D view of reality that quantum theory is indeterministic.
This is backwards. Quantum mechanics treats systems as closed, deterministic systems. However, decoherence depends on quantum systems being open:
"over the past three or so decades it has been slowly realized that the isolated-system assumptionwhich, as we have described above, had proved so fruitful in classical physics and had simply been taken over to quantum physicshad in fact been the crucial obstacle to an understanding of the quantum-to-classical transition. It was recognized that the openness of quantum systems, i.e., their interaction with the environment, is
essential to explaining how quantum systems (and thus, assuming that quantum mechanics is a universal physical theory, all systems) become effectively classical: How their quantumness seems to slip out of view as we go to larger scales, finally arriving in the world of our experience where cats are either alive or dead but are never observed to be in a superposition of such two classically distinct properties." (
Decoherence and the Quantum-to-Classical Transition; Springer, 2007).
The problem with the above is that while it is now generally accepted to be the correct approach (somehow), it doesn't resolve anything. Because "to observe interference effects between components of a superposition state, these components must have not been measured, i.e., which-state information must not be available." (ibid).
What? It means that quantum systems are indeterministic in a known, predictable way.
Which is not indeterminism, and is
absolutely not countefactual indefiniteness. Countefactual indefiniteness literally means "if I didn't observe what I just did, what I observed wouldn't be there". It's why Einstein asked "is the moon there when you don't look at it?"
By definition, all observations are correct
Can the following describe "correct" observations?
"Yesterday, upon the stair,
I met a man who wasnt there
He wasnt there again today
I wish, I wish hed go away..."
Someone observed something which wasn't there. That's counterfactual indefiniteness.
...? And no scientist in the world has noticed fabrication of data? (Because your error margin is data as much as your actual values are.)
No, every single physicist knows this. And every single scientist interested in quantum theory learns it.
Excuse me a moment; I must snark.
Hey, if it's done well (and 'twas) no problem.
However, I did choose that particular example explicitly because GTR tells us that velocity is relative.
No, it tells us that the observation is relative. And we know with respect to what. Everett et al. took the determinisim of quantum formalisms, and said that all possible "states" existed, but the one we ended up with in our universe is the result of our measurement. However, although Everett intended to "take it like it is", the problem is that quantum formalisms are deterministic: we start with an initial state and a function which allows us to "know" at the very least a far smaller probability range than the "many-world" approach dictates is involved in every measurement. Which means we can't get the states we do using this interpretation.
Everett's interpreation is designed around no "selecting" going on.
What are you basing your understanding of Everett on?