Locality Is Dead! Long Live Locality!

Several decades of theory and experiment into EPR correlations have led to the widely held belief that reality is non-local, in spite of the fact that this violates special relativity. To date, no experiment has shown a violation of special relativity, and EPR experiments do not demonstrate the existence of superluminal information exchange, merely correlations which violate certain inequalities. Every “loophole” in these hidden variable theories has been thought plugged. However, there is much confusion in the literature due to conflation of the terms “locality,” “realism,” “hidden variables,” “non-contextuality.” The presence of local hidden variables is thought to necessarily lead to a Kolmogorov probability structure (hence non-contextuality), but this is an assumption, one which is not true in general once context effects are taken into account. Treated as an observational theory, several authors have shown no incompatibility between quantum mechanics and locality, and that the Bell scenario is actually about whether reality is contextual. This paper proposes a descriptive theory by assuming a generated reality (following Whitehead's Process Theory) which can violate the principle of continuity and possess non-Kolmogorov probability structure, and reproduce the results of non-relativistic quantum mechanics, while allowing only causally local information exchange without hidden variables. A generated reality is thus compatible with both quantum mechanics and special relativity, reproducing all of the results expected from quantum mechanics while still maintaining causally local realism. This process model thus appears to be an ideal candidate for developing theories for the unification of quantum mechanics and general relativity.


INTRODUCTION
The debate as to whether reality at its fundamental level conforms to the tenets of local realism has been decided decisively in the negative. Non-locality has won. Locality is dead! Or is it? In spite of an ever more sophisticated series of hidden variable theorems and dramatic experimental results, has the issue of local realism truly been laid to rest? In the years following the seminal paper of Einstein, Podolsky and Rosen (EPR) [1], the concept of local realism has become equated with the concept of (deterministic) local, non-contextual, hidden variables (LNHV). The assumption of LNHV leads to inequalities on measurement correlations, which experiments have shown are violated. The conclusion is that LNHV do not exist.
The restriction of local hidden variables to deterministic LNHV was unnecessary, likewise, the restriction of local realism to LNHV is also excessive. Arguments in support of this will be provided and an explicit model of non-relativistic quantum mechanics (NRQM), the Process Algebra model, will be demonstrated. The Process Algebra model is a local, realist, generative, contextual, and discrete model of NRQM without hidden variables in which NRQM appears as an effective theory in the continuum limit as Planck length and time are taken to zero [2][3][4][5][6]. It has been argued that the Process Algebra model provides a true completion of NRQM [4]. The Process Algebra provides a nuanced framework for representing interactions between fundamental entities which the standard Hilbert space formulation lacks. It suggests that the quantum paradoxes and conundrums are due to a failure of the usual Hilbert space formalism to correctly represent particle interactions.

THE ARGUMENTS AGAINST LOCAL HIDDEN VARIABLES
The story against local realism begins with the 1935 EPR paper [1], which tackled two questions, that of the completeness of quantum mechanics as a physical theory and that of the nature of reality. As to completeness, they wrote: "Whatever the meaning assigned to the term complete, the following requirement for a complete theory seems to be a necessary one: every element of the physical reality must have a counterpart in the physical theory." They later suggested the following as a plausible, sufficient but not necessary definition of reality: "If, without in any way disturbing a system, we can predict with certainty (i.e., with probability equal to unity) the value of a physical quantity, then there exists an element of physical reality corresponding to this physical quantity (SIC)." Nowadays quantum mechanics is considered to be complete mathematically (or epistemologically), because no addition to quantum mechanics results in a probabilistically better theory [7]. Whether or not it is complete ontologically, thus providing a complete description of physical reality is no longer a criterion [8].
The EPR argument was not about locality per se but against contextuality, the inability to perform simultaneous measurements of incompatible (non-commuting) observables. Locality was inferred from the requirement that the systems should not interact with one another in any manner. Subsequently Bohm [9], Bohm and Aharonov [10], Bell [11], and Clauser, Horne, Shimony, Holt [12] introduced refinements to the EPR argument which brought the scenario (or a version thereof) closer to experimental scrutiny and explicitly addressed the issues of locality and hidden variables. The idea of hidden variables refers to the existence of some unknown parameter space , such that all measurements A and probabilities p associated with an experiment are functions of values, that is, A(x, λ), p(x, λ), where x refers to all of the overt variables associated with the experiment. Early papers focused on deterministic hidden variables, but as Bell pointed out [11], the question of deterministic, non-deterministic, or stochastic is irrelevant; the real question is whether or not the assumption of hidden variables can explain the observations. The Bell scenario involves two quantum systems, I, II, which interact to form an entangled state, ensuring that the states of the two systems are correlated. The systems then propagate to spacelike separated locations, X, Y, which, if special relativity (and therefore locality) holds, should ensure that they are unable to interact with one another in any manner. Next two independent observers, conventionally Alice and Bob, are allowed to carry out measurements, Alice of system I, Bob of system II, of (usually non-commuting) observables A,B, respectively, each parameterized by a,b, respectively. After collecting their data, Alice and Bob then determine various correlations among their measurements and then test these results against a specific inequality, namely −2 ≤ E(a ′ , b ′ ) + E(a ′ , b ′′ ) + E(a ′′ , b ′ ) − E(a ′′ , b ′′ ) ≤ 2 where E(x, y) is the expectation value of the product of the outcomes of measurements of the two systems when Alice's observable setting is x and Bob's observable setting is y ( [13], chapter 8).
Bell [11], Jarrett [14], and Shimony [13] emphasized that one of the key components of the argument leading to the inequality is that the probability distribution given by the hidden variables must satisfy a factorizability condition. Following Shimony [13] and Jarrett [14], let p i (x/k, a, b) denote the probability of observer i measuring outcome x given complete state k, Alice's setting a and Bob's setting b. p(m, n/k, a, b) is the joint probability when Alice obtains measurement m and Bob obtains measurement n. p i (x/k, a, b, y) is the conditional probability when the second observer obtains measurement y. Jarrett defined two independence conditions: Parameter Independence p i (m/k, a, b) = p 1 (m/k, a) p 2 (n/k, a, b) = p 2 (n/k, b) Outcome Independence p 1 (m/k, a, b, n) = p 1 (m/k, a, b) p 2 (n/k, a, b, m) = p 2 (n/k, a, b) Jarrett showed that the assumption of both independence conditions leads to the factorizability condition: p(m, n/k, a, b) = p 1 (m/k, a)p 2 (n/k, b). This condition is an essential component of most hidden variable arguments [10,11,[15][16][17][18]. When true, the hidden variables are non-contextual with a Kolmogorov probability structure, unlike quantum mechanics which has a non-Kolmogorov probability structure by virtue of the Born rule. All hidden variable arguments assume factorizability. Assuming factorizability, Fine [19,20] showed that the presence of deterministic LNHV implies the existence of a joint probability distribution for even non-commuting observables, violating the predictions of quantum mechanics. Kolmogorov [21] himself emphasized that probability theory was fundamentally a contextual theory. Probability distributions were context dependent. Fine [20] developed some criteria for when a joint probability distribution exists but years earlier Vorob'ev [22] had presented a complete set of criteria for the existence of joint distributions in the general case and several examples where his criteria failed to be satisfied. This clearly showed that the factorizability condition is an assumption, not a necessity. The fundamental question is whether hidden variables are non-contextual (factorizable) or contextual.
Another approach to the EPR scenario has focused upon contextuality directly. EPR [1] and others [23] define realism to mean that every element of reality possesses, a priori, a definite value for every possible observable. Contextuality asserts that not all observables can have pre-existing values. Realism is then equated with non-contextuality. A succession of ever more powerful results (von Neumann [24], Gleason [25], Mackey [26], Kochen-Specker [27], Mermin [28]) have shown quite conclusively that quantum mechanics, in its Hilbert space formulation, is a contextual theory. Dispersion-free measures do not exist, so that it is impossible through acts of measurement to assign definite values to all possible observables to single physical entities. If realism is equivalent to non-contextuality, then quantum mechanics shows that realism does not exist. One is left with Wheeler's famous dictum that "no phenomenon is a phenomenon until it is an observed phenomenon" [29], so that there is no external reality; the actions of an observer cause reality to manifest. The realist perspective attempts to avoid falling down this particular philosophical rabbit hole.
It took nearly 50 years to develop the technology to allow testing of these theories. Most experiments have involved entangled photons although a few have used entangled electrons [30][31][32][33][34][35][36][37][38]. Since the first experiments of Aspect and Grangier [30] showed that the Bell inequality was indeed violated, a number of possible "loopholes" have been proposed related to experimental factors such as detector accuracy, propagation losses, detector distance, superdeterminism, all of which have been eliminated by subsequent experiments [31][32][33][34][35][36]. Even free will have been challenged [39]. Bell's inequality has been violated in all of these experiments to a statistical level of at least 11 standard deviations, and any presumed non-local influence must propagate with a speed of at least 50,000c. Bell type experiments have now become school demonstrations [37]. There is now an experiment which visualizes correlations referred to in Bell's theorem [38].
For those who accept Bell's argument and its variations, the issue would appear to be put to rest. LNHV do not exist and reality, if it even exists, is non-local.

THE ARGUMENTS AGAINST THE ARGUMENTS AGAINST LOCAL HIDDEN VARIABLES
Or is it? As in every walk of life, things are not as simple as they first appear.
Khrennikov [40], drawing on work of Hertz and Boltzmann, divides theories into two general kinds: descriptive, and observational. Descriptive (ontological) theories attempt to describe the entities (causes) that give rise to observed phenomena. Observational (epistemological) theories attempt to merely provide a predictive framework for these same phenomena. Quantum mechanics is, mostly, an observational theory. The arguments against local hidden variable described above are framed within an observational framework.
The Bell scenario involves three distinct stages: interaction, propagation, measurement. The literature has focused primarily upon the measurement stage, simply assuming the first two stages as given. Nevertheless, the derivation of the inequalities depends upon assumptions made regarding these initial stages. Quantum mechanics is not involved in this derivation. Experiments have been performed which show that the inequalities are violated. Logically then, there must be errors in the assumptions leading to the inequalities. This is not necessarily a vindication of quantum mechanics.
From an observational/epistemological perspective the key problem is the assumption of factorizability of the probability associated with whatever variables are assumed to be present in the description of the measurement situation. A factorizable probability leads inevitably to a joint distribution for the measurements, regardless of whether they are compatible or complementary. Is the assumption of factorizability necessary for any model of local realism?
Khrennikov [40], following earlier work of Landau [41], constructed a quantum mechanical analog of the CHSH inequality. Given 4 observables, A 1 , A 2 for system 1 and B 1 , B 2 for system 2, he considered: . If either the A or B operators are compatible, then | < B > | ≤ 1. He then showed that there exists a quantum state such that || B|| 2 ≥ (1 + µ) > 1 so that quantum mechanics violates the above inequality. Khrennikov argued that the quantum analog of the CHSH inequality measures the degree of incompatibility among the observables being measured on each system and is not a reflection of nonlocality. Similarly, Cabello [42] has demonstrated formally that the generalized Bell inequality and Kochen-Specker contextuality are equivalent in quantum mechanics. Nieuwenhuizen [43] also examined the CHSH inequality and showed that the probability measures required to calculate the various correlation functions were subject to contextuality effects, so that no joint probability distribution, required to make meaningful the resulting inequality, exists. He referred to this as the contextuality loophole, and also argued for its universality. Kupczynski [44] advocates for an purely epistemological, ensemble interpretation of quantum mechanics, and also argues that the Bell argument is invalid because it fails to take into account the contextual nature of the probability distributions associated with these ensembles.
Khrennikov has developed an extension of Kolmogorov probability theory called contextual probability theory [45]. Contextual probability theory does for probability theory what non-Euclidean geometry did for geometry. The point of departure from Kolmogorov probability is the sum rule, which takes the form p b λ is the probabilistic measure of interference and the p terms are various conditional probabilities over contexts (C) and observables a,b. In Kolmogorov probability, λ = 0, otherwise λ can be a trigonometric function or a hyperbolic function. Contextual probability has been applied to a number of classical level phenomena in biology, psychology, and economics [45][46][47][48][49][50][51] and to the Bell situation [52,53]. The appearance of non-Kolmogorov probability at a classical level demonstrates empirically that the assumption that any probability associated with local hidden variables must be Kolmogorov is prima facie false. Hidden variables may be non-contextual or contextual at both classical and quantum levels. This does not necessitate non-locality or a failure of realism.
Dzhafarov et al. [54], Dzhafarov and Kon [55], and Dzhafarov and Kujala [56] has presented an alternative approach to that of Khrennikov termed Contextuality by Default. Following the notation in [57], each random variable is associated with the quantity q being measured and the context a within which the measurement is made, and denoted, R a q . Consider two measurements, q,q' and two contexts a,b. For a fixed context a, the pair R a q , R a q ′ is termed bunch, representing the collection of measurements associated to a specific context. It is reasonable to believe that such a pair is jointly distributed. For a fixed measurement q, the pair R a q , R b q is termed a connection for q. The most basic form of contextuality occurs when no joint distribution can be found for a connection. In such a case they are said to be inconsistently connected. This is the situation of contextuality by default. Dzhafarov considers this to be the most trivial form of contextuality since it is so ubiquitous. Dzhafarov has developed a more restricted notion of contextuality in line with contextuality in physics. He considers couplings between bunches. For example, given two bunches R a q , R a q ′ and R b q , R b q ′ , a coupling is a set of jointly distributed random variables (A,B,X,Y), subject to certain constraints, such that (A,B) is distributed as R a q , R a q ′ and (X,Y) is distributed as The constraints involve A,X and B,Y which correspond to measurements of q and q', respectively. A measurement q is considered to be context independent if among all couplings (A,B,X,Y), we have Pr(A = X) = 0. It can be shown that such a coupling may not exist even if the system is consistently connected. Now consider all couplings (A,X) for just the connection R a q , R b q ′ and find the minimal value m' for Pr(A = X). Then consider the global coupling (A,B,X,Y) and again find the minimal value m for Pr(A = X). If m = m' the system is noncontextual and if m>m' then the system is contextual. This form of contextuality is analogous to that found in physics and gives rise to similar types of inequalities.
Contextuality by default has been observed experimentally [57]. Moreover, two recent studies [58,59] have demonstrated the strong form of contextuality in a social psychological setting [58] and in individuals [59]. Contextuality in the form observed in quantum mechanical settings is thus not unique to the quantum domain but can occur in classical settings as well. Dzhafarov and Kon [55] have analyzed the Bell scenario within the contextuality-by-default model, and showed that it can be understood using wholly classical (albeit contextual) probability theory.
Dzharfarov and Kujala [60] applied Contextuality by Default analysis to the double slit experiment. They pointed out that "Contextuality or non-contextuality is a property of a system of random variables representing an empirical situation rather than of the empirical situation itself." They presented a very general model of the two slit situation, using Kolmogorov probability together with the addition of a context parameter, much as Kolmogorov originally argued [21], and obtained the usual statistics. Dzhafarov argued that the Bell argument is not about the nature of physical reality but rather about the failure to take context effects into account when creating a Kolmogorov type probability model of a situation.
The role of context in probability theory has not always been ignored. Kolmogorov [21] and von Mises [61] understood that probability theory was contextual. Pitowski [62,63] analyzed the Bell situation and presented a model based on a form of contextual probability. In his model "The relative frequencies violate Bell's inequality the way they do because the locality principle is true" (SIC) [63]. Later, Pitowski [64] developed a deterministic model of spin statistics using the concept of nonmeasureable sets. He argued [63] that quantum mechanics is essentially a probability theory, which in Khrennikov's language would be viewed as a trigonometric contextual probability theory. Gudder [65] applied a generalized probability theory similar to that of Pitowski to the problem of spin statistics and showed that such a model was compatible with local hidden variables. Gudder [66] had already shown that a hidden variable model of the Bell scenario was possible so long as contextual hidden variables were used, a line of thought supported a few years later by Ballentine [67]. Local contextual hidden variable models have also been developed by Durdevic [68,69]. Recently Griffiths [70], using a coherent histories approach, has reaffirmed that quantum mechanics is a local theory, the inequalities are a consequence of the contextuality of quantum mechanics, and the correlations that are detected in a typical Bell experiment arise due to a common quantum cause [71]. These models seem to have been ignored in the mainstream literature. The belief that the probability theory of the classical world is necessarily Kolmogorov and non-contextual has achieved the status of dogma, and it has proved extremely difficult to disabuse people of this.
A different approach was proposed by Palmer [72] who developed a deterministic model of the spin scenario but where, crucially, there was a non-linear dynamics in place. The consequence of this was the impossibility of forming the correlation functions required for the Bell inequality. Experimentally, correlations can always be calculated but they need not be meaningful [73]. He suggested that the inequalities said nothing about the nature of reality, since the correlation functions involved did not exist.
Remarkably, contextual hidden variable theories have not gained much traction within the foundations community. In spite of their ability to reproduce the quantum mechanical results while preserving locality, they have, for the most part, been ignored in favor of the quasi-mystical notion of non-locality. Shimony ([13], chapter 10) defined two types of contextual hidden variables: environmental, which include experimental conditions, and algebraic, referring to models on quantum logics or lattices. He rejected both types of contextual hidden variables of the environmental type ( [14], chapter 10), arguing that they would still satisfy a factorizability condition, but without proof. Shimony appeared to reject locality, evading special relativity by his "passion at a distance." He argued that a breach of outcome independence did not imply superluminal signaling, but outcome dependence reflects contextuality, not necessarily non-locality.

LOCAL REALISM NEED NOT IMPLY NON-CONTEXTUAL LOCAL HIDDEN VARIABLES
If the observational approach to theory suffices, then it appears clear that measurement is a contextual act. Non-disturbing, noncontextual, "objective" measurements do not exist universally. The classical notion of objectivity does not hold true. This does not, however, imply that an act of measurement creates reality, merely that reality may be altered by such an act. Reality appears to be interactive, and thus characterized by a weak form of subjectivity. This appears easier to accept than non-locality, since evidence for contextuality is all around us, while evidence for non-locality is profoundly lacking. Griffiths writes [70] "To be sure, those who claim that instantaneous non-local influences are present in the quantum world will generally admit that they cannot be used to transmit information; this is known as the 'no-signaling' principle, widely assumed in quantum information theory. This means that such influences (including wave function collapse) cannot be directly detected in any experiment. The simplest explanation for their lack of influence is that such influences do not exist. " Pusey, Barrett and Rudolph [74] have argued that the wave function is ontological, and experiments visualizing Belltype non-local behavior [38], quantum jumps [75], quantum measurement processes [76], quantum trajectories [77], quantum wave functions [78], and single photons [79] would seem to support this. Evidence that quantum jumps [75] and quantum measurements [76] evolve over a period of time suggests that there is an actual "something" out there corresponding to such behavior. This suggests that a purely observational theory is inadequate and a descriptive theory is also needed to explain contextuality and the presence of long range correlations.
The dominant viewpoint, however, is that the wave function is merely epistemological, and that quantum mechanics deals only with the statistics and behaviors of ensembles. The ability to carry out single photon and single particle experiments demonstrates that this is not a necessity [80,81]. The ensemble approach focuses on the density matrix ρ, defined as ρ = where the |φ i > are pure states [82]. Moreover, the density matrix involves two different kinds of probability: an explicit classical probability in the form of the real valued p i and an implicit non-Kolmogorov probability in the form of the squared modulus of the complex valued amplitudes of the pure states. These two considerations suggest that ensembles should be treated ontologically as supervening on pure states. Thus, questions of ontology should reference pure states.
The challenge in finding an ontological model of pure states in quantum mechanics lies not with the measurement problem (for which a detailed model within the usual quantum mechanical framework has been proposed [83,84]) but rather in providing an ontological understanding of superposition. A naïve attribution of measurements to a single particle in a superposition state leads to confusion and paradox, causing many to abandon an ontological interpretation of the wave function and sometimes reality itself.
Norsen [85] has pointed out that dispensing with the idea of realism, broadly considered, results in the end of scientific inquiry, because without some notion of a "reality, " what is it that scientists have to talk about? But must realism be identified with classical objectivity? Zeilinger et al. write "objects have physical properties independent of measurement (the assumption of realism)"( [34], p. 250401-1). But this is just the definition of non-contextuality. This definition of realism seems to beg the question.
Rosen [86] suggested that physics' focus on inanimate matter has resulted in an unnecessarily limited world view. Experience with emergent systems in biology, psychology, and economics [87,88] has demonstrated that many naturally occurring systems and phenomena are transient, open, multiscale, emergent, generated and generative, contextual, and subjective. It is doubtful that anyone seriously considers these systems to be "real" only as a result of "observation." It is not realism that needs to be abandoned-rather concepts such as ideal object, ideal non-disturbing measurement, and non-contextuality must go.
A metaphysics with subjective elements was proposed by Whitehead [89] nearly a century ago. It is a process model of reality which emphasizes its transient, generated, generative, emergent, and contextual features. Several authors have proposed process models of physics [2-6, 90-99] and even Shimony [13] wrote about Whitehead's idea of process.
Consider an alternative definition of realism. To begin, why is it necessary that an element of reality possess a priori all of the properties that can be measured on it? Measurement is an act, and always involves an interaction between a system and a measuring apparatus. It is generated in the moment as the interaction takes place. There is no need to assume that something corresponding to this measurement exists in the system prior to its interaction with the measurement apparatus. It is only necessary that the system possesses the potential to determine such a measurement when it interacts with a measurement apparatus. It is equally unreasonable to assert that nothing exists prior to the interaction with the measurement apparatus. The system must exist or what exactly does the measurement apparatus interact with? Moreover, the interaction with the system results in a systematic difference which ensures that only particular measured values are returned with particular frequencies. If the measurement apparatus creates the measurement, then why just these values and no others? The system must possess a potentiality which becomes realized in any interaction with a measurement apparatus.
Why too is it necessary that for something to be real it must be knowable to a human observer. Quantum mechanics appears obsessed with the idea of measurement, yet events occur in nature without any obvious "measurement" taking place and without any "observer." A theory of natural processes should, reasonably, describe the evolution of such processes as they occur "in vivo, " and not merely "in vitro, " in a laboratory. Moreover, such a theory should be able to deal with single entities, not merely ensembles.
At a bare minimum, any entity must possess some propensity to determine a difference in the state or future history of at least one of the entities that it interacts with. For, if something is thought to exist, but in any interaction with anything whatsoever, in any manner and for all time, it never determines any difference whatsoever, then it might just as well not exist, since its existence will never be noted. Of course an entity may never have an opportunity to interact with another entity, so it is not essential that a difference be realized, only that the entity possesses a propensity to determine a difference, should a suitable interaction occur. These two considerations suggest the following minimalist definition of realism: An entity is real if it exhibits a propensity to determine a systematic difference. An entity or phenomenon can determine a difference in only a single interaction for a single time, or across many interactions over multiple times. It may determine the same difference at multiple times, or it may determine different differences on different occasions, but these should be systematic in some manner. With respect to the same initial conditions, the difference may be deterministic (fixed single value), non-deterministic (fixed set of values), or stochastic (fixed probability distribution). There is no need for an observer, particularly a human observer. Previous comments aside, if this potentiality is never realized then its reality is rather moot. Thus, there is a need for an entity with which it can interact so as to realize this potentiality and thus register its reality. It may be that this additional entity is itself, and that selfinteraction might provide the most basic interaction realizing a potentiality, perhaps that of bare existence. The realization of other potentialities requires interactions with wholly separate entities. The minimum requirement for such entities is that they can register a difference in either state or history, a concept referred to as salience [100]. There is no need for consciousness or agency.
The determination of a difference requires interaction; it is relational. Some differences may be private, specific to an individual entity (for example quarks and gluons) or public, accessible to many entities (for example photons). The idea of reality being a propensity to determine a difference has much in common with the pragmatism of John Dewey [101] and the process view of Whitehead [89]. Propensity drives home the point that it is the capacity to determine a difference which matters, not which difference it is which is determined.
A quantum system determines a potentiality to obtain certain measured values through an act of measurement. These values are only realized through an interaction with a measurement apparatus. It is not necessary that the quantum system possess these measured values, merely that it possess the propensity to determine them, if only in a statistical manner. It is a set of dispositions. Since only select measurements are made possible, it makes a difference, and by the definition proposed here, it is real. It might be associated with a particular measurement on one occasion, but this need not be the same on a subsequent occasion. That will depend upon the intervening interaction history. Properties can be real, they can be contextual, they simply need not be eternal.
Measurements are not specific properties of the system but propensities, one of which may be realized following a measurement act. The system does not possess these properties but rather, together with a suitable measurement apparatus, acts as a generator of properties. Thus, quantum systems should not be thought of as "objects" but rather as "processes." Since processes make a difference, somewhat, somewhere, sometime, processes are held to be elements of reality. They are ontological entities but with characteristics such as transience, emergence, generativity, agency, contextuality, and locality.
Some quantum properties have a universal character, such as whether their wave function is scalar, spinor, vector, tensor, or their charge, rest mass, and so on. These can be attributed to the system itself. Many other properties, however, are contextual in character and thus should be treated as generative propensities. For example, whether a quantum system is to be considered wave-like or particle-like is contextual. Indeed, Ionicioiu et al. [18] showed that the wave-particle distinction is not compatible with a non-contextual hidden variable representation. The tracks of fundamental particles are also contextual in that they do not occur in the absence of a detector. Mott [102] showed that the formation of particle tracks in a bubble chamber was an emergent feature of the interaction between the particle and the atoms in the bubble chamber. There are no tracks without the bubble chamber. The formation of tracks is a propensity of the particle, not a property of the particle.
Consider again the Bell correlations. Reichenbach [103] formulated the principle of the common cause, which states that if an improbable coincidence has occurred, there must exist a common cause. In the Bell scenario, the common cause is presumed to occur at the time at which measurements on the two systems take place. However, consider the following Gedanken experiment. Consider a source which produces a pair of spinentangled particles I, II, with opposite momenta p, -p, which are allowed to move to locations a distance d apart, where there are placed detectors for Alice and Bob. The detectors of Alice and Bob are space-like separated. The outcome of each measurement is sent to a common location at a distance h from each detector where there is a recording device which measures the outcome from each detector simultaneously. A common trigger is established at a distance r from each detector. When a signal from the trigger is sent, each detector makes a measurement, thus ensuring simultaneity. In the interval between trigger signals, Alice and Bob are free to alter the specific measurement being made by their detector.
The output from the source has a wave function of the form The question is: at what point in this situation is the supposed passion at distance to play out? The unusual notation for the wave function is intentional. In part it follows Mott's [102] argument in which he emphasized that the wave function for the interaction between particle and bubble chamber fluid possessed a complex, non-separable dependence on both the particle and the atom with which it is interacting. It is also meant to emphasis the algebraic aspect of Hilbert space expressed in the bra-ket notation and the representation of this algebra by means of the wave function. This will be important later in the discussion of the Process Algebra. Note that knowledge of this wave function comes from outside Alice and Bob. From Alice's point of view, there is a single particle whose wave function is | 1 (r 1 ) >= 1 √ 2 (|φ 1 (r 1 ) > +|φ 2 (r 1 ) >), while Bob sees a single particle with wave function | 2 (r 2 ) >= If they were to pool their descriptions they would assume that the combined wave function is | (r 1 , r 2 ) >= | 1 (r 1 ) > | 2 (r 2 ) >, which is clearly inconsistent with the actual case and also with what is observed by the recorder. Why then are we so quick to assume that the entangled wave function Alice and Bob will make their measurements simultaneously, at exactly the same proper times for each particle. If Alice were to carry out her measurement slightly earlier than Bob, Alice would presumably collapse the wave function of particle I and cause the wave function of particle II to collapse to the corresponding entangled state, and conversely if Bob measures first. Here, however, the measurements are carried out simultaneously. So what will happen? If nothing exists prior to the simultaneous measurements of Alice and Bob, then it is not clear at all how this is to be resolved. Alice and Bob believe that they are working with free particles yet the recorder will obtain correlated measurements. A definite difference will be observed.
Thus, according to the definition of realism being considered here, a definite something must exist prior to the measurements taking place.
The entangled system exhibits a propensity to determine two simultaneous correlated measurements which make a difference, and thus it is reasonable to consider that the entangled system represents a single element of reality. The error lies in assuming that one has two quantum systems which are somehow correlated. That seems possible only through the passage of some signal between them, but in the case of simultaneous measurements, what could such a signal convey? Such a signal not only must be instantaneous, but it must also effect a choice. This seems implausible.
A simpler explanation is that there is only one system, but it produces two measurements. This could occur in one of two ways. First of all, when the entangled state is created, entangled particles are emitted in either of the two entangled states, |φ 1 : ρ 1 (r 1 , r 2 ) > or |φ 2 : ρ 2 (r 1 , r 2 ) >, and then propagated. In that case the wave function is merely epistemological, describing a statistical ensemble of entangled particles. The second possibility is more interesting. At the moment of measurement the entangled system has the propensity to manifest either state |φ 1 : ρ 1 (r 1 , r 2 ) > or |φ 2 : ρ 2 (r 1 , r 2 ) >, never both, with 50-50 frequency. In either of these cases there is no need for quasimystical instantaneous signals to be passing back and forth between measurement devices. Now suppose that the experimental situation is rescaled, so that all distances diminish by a proportion p. The above argument can be repeated and again, the entangled system exhibits the propensity to manifest either state |φ 1 : ρ 1 (r 1 , r 2 ) > or |φ 2 : ρ 2 (r 1 , r 2 ) >, never both. It follows that this must hold true for every moment of time. Since, at every such moment of time, the entangled system has the propensity to determine a difference, there must be an element of reality present at each moment of time. The entangled system is thus the generators of these momentary propensities.
Since the wave function appears to be at least partially ontological, the simplest explanation for this propensity is that at each moment of time, the entangled particle system actually manifests either state |φ 1 : ρ 1 (r 1 , r 2 ) > or |φ 2 : ρ 2 (r 1 , r 2 ) >, never both, and that the state may change from moment to moment.
If one believes in the principle of continuity then this is quite problematic and perplexing. If time is continuous then the evolution of the entangled system must be neither smooth, nor continuous. For if H is the Hamiltonian and time evolution is given by the usual operator, U(t) = exp[−itH/h], then assume that there is some time interval [t 1 , t 2 ] on which the entangled state is constant. Then U(t)=I for every t ≤ t 2 − t 1 , and hence U(t)=I for all t. It appears that one must abandon the principle of continuity. Bancal et al. [17] and Gisin [104] and colleagues addressed this problem in the context of a Bell scenario. They studied the case of 4 quantum observers and, by assuming the principle of continuity and a constant, finite but unspecified superluminal speed v (c < v < ∞) of propagation of any hidden signals, were able to find an inequality involving various correlated measurements, as well as a quantum state which violated the inequality. They concluded that either the principle of continuity must be violated, or superluminal signaling must be possible. Gisin wrote "Note that the finding of such a speed would falsify both quantum theory and relativity, a result not many physicists are willing to envisage" [104, pg 10] thus favoring abandoning the principle of continuity. Bancal et al. wrote "This gives further weight to the idea that quantum correlations somehow arise from outside spacetime, in the sense that no story in space and time can describe how they occur" ( [17], pg. 4).
The theory of special relativity has survived multiple experimental tests and has yet to be violated. The principle of continuity, however, is frequently violated at smaller scales, so why should it not be violated at the smallest scale? This would appear to provide far less of a shock to our conceptual system than the assumption of instantaneous transmission of undetectable signals. Indeed a recent paper argues that there is an upper limit to the frequency of any physical process, including any "clock, " with the shortest temporal interval on the order of 10 −33 s [105].
Local realism, according to the new definition, appears to be perfectly tenable provided one accepts contextuality and abandons the principle of continuity. The remainder of this paper is devoted to describing just such a locally realist model of non-relativistic quantum mechanics (NRQM) without hidden variables.

PROCESS AND THE PROCESS ALGEBRA MODEL
One way to implement a descriptive theory together with the new definition of realism is through the concept of process. This has its origins in the writings of Heraclitus and Siddhartha Gautama and its modern version in Whitehead's process theory. A propensity to determine differences can be accommodated if those differences are generated. A process is viewed as a generator of primitive events called actual occasions, the base elements of reality. By the new definition of realism, if a process is responsible for determining differences, manifesting as actual occasions, then a process must be accorded the same ontological status, that is, the same degree of reality, as those occasions.
Whitehead considered a process to be a sequence of events having a coherent temporal structure in which relations between the events are more fundamental than the events themselves. Whitehead viewed process as being ontologically prior to substance and becoming to be a fundamental aspect of being. Becoming is fundamental to process, and fundamental to becoming is transience. In process theory events have a transient existence, coming into being, manifesting briefly, then fading away. Each actual occasion exists only long enough to prehend the realities of the previous events and to form a response to them, thereby immediately passing out of existence and becoming data for subsequent events. Actual occasions, the basic elements of reality, are held to be inseparable occasions-in-connection, giving reality a holistic aspect. The act of prehension underscores that information plays a fundamental role in the unfolding of reality, and that in particular, it is meaning that is necessary to give rise to coherence among events [90]. Reality is emergent, arising out of a lower level of actual occasions as are the fundamental physical entities, which are viewed as emergent configurations of actual occasions [13,89].
Actual occasions do not occur in space-time, nor do they move in space-time. Instead, actual occasions form space-time. They are the primitive "events" upon which natural entities supervene and from which they emerge. Processes, being generators of actual occasions, are thus generators of space-time itself. Logically, processes do not exist within space-time, they stand outside of it, thus fulfilling the suggestion of Bancal et al. [17].
In keeping with the descriptive perspective, actual occasions constitute the signs by means of which processes implement their propensity to determine differences. An actual occasion marks a specific expression of this propensity, determining one of whatever many differences the process may determine. Actual occasions thus mark whichever difference is being determined, in the moment, and this difference may vary from moment to moment.
The idea of process depends crucially upon the idea of becoming, and that in turn requires a transient now. Such a concept is thought to be incompatible with special relativity, but this is a misunderstanding of what special relativity implies. As Wigner pointed out [106], what special relativity demonstrated is the non-existence of global frames of reference. All global frames of reference are mathematical fictions. Simultaneity, and thus a transient now, can exist, but co-moving observers within the universe will not agree about this. Denying simultaneity is another example of misplaced omniscience. Reality may unfold according to a transient now even if human observers cannot detect it. Several authors have argued that it is not the block universe which is a necessity but rather some form of presentism [107][108][109][110].
Similar to the actual occasions that they generate, processes shift between periods of activity and inactivity. While active, they express a propensity to determine differences, manifesting in distinct attributes and functionality. Processes interact with one another according to their attributes and functionalities and the actual occasions that they manifest, and these interactions are triggered by the manifesting of particular actual occasions.
Process ideas can be seen in Finkelstein's quantum relativity [92,95], Noyes's bit-string physics [94], Bastin and Kilmister's combinatorial physics [96], Hiley's process physics [93], Cahill's process physics [97]. Emergent models of physics include Nelson [111], Adler [112], Levin and Wen [113], and two time models such as stochastic quantization [114] and Bars' two time physics [115].The process algebra model has many roots: Sorkin's causal sets [116] (whose basic elements could be reinterpreted as actual occasions), Lee's discrete time dynamics [117], Kempf 's interpolation model of QFT [118] (which suggested that NRQM could be emergent from a discrete space). Related models include the cellular automata models of 't Hooft [119] and Elze [120], which appear to be special cases of process algebra models. Trofimova [98,99] has proposed several process algebra based formalisms for describing the principles of transience which govern processes in functional constructivism. Her approach to process algebra uses several functional differentiation classes, a concept of "performance" and several universal process-trends. It applies particularly to complex, adaptive, multiscale systems. Her work has provided much inspiration for the author.
The next few pages describe the process algebra model in terse detail. A more leisurely discussion can be found in [2,4]. The process algebra model considered here views quantum mechanics as an (incomplete) effective theory, being the asymptotic limit as spatiotemporal scales become infinitesimal. The Hilbert space formalism is considered to be mathematically coarse, blurring the distinction between ontological and epistemological, and leading to a great deal of unnecessary confusion.
Since von Neumann, the use of the language of Hilbert spaces for formulating NRQM has become dogma [24]. The Process Algebra model starts from the realization that the Hilbert space of NRQM is a reproducing kernel Hilbert space [121]. Given a reproducing kernel Hilbert space H(X) with base space X, one can find a discrete subspace Y of X (sampling subspace), and a Hilbert space H(Y) on Y, such that each function in H(Y) can be lifted to a function in H(X) via interpolation. Interpolation means that if (z) is a function in H(X), then for each y ∈ Y there exists an interpolation function y (z) on H(X) such that (z) = y∈Y (y) y (z). In general there are usually an infinite number of these sampling subspaces. The interpolations functions are not unique. They are usually chosen by reason of goodness of fit. In the case that the subspace Y has the form of a regular lattice the interpolation functions may be taken to be sinc functions (sin x/x) [90]. If the subspace has an irregular structure with density matching the Beurling density [122], Fechtinger-Gröchenik interpolation theory may be used instead [121,123]. Interpolation does not reproduce all functions on H(X) but rather a more limited set of band-limited functions, that is, functions whose Fourier transform is limited to a bounded set, ensuring the existence of a natural ultraviolet cutoff.
In interpolation theory, H(X) is considered to be fundamental while H(Y) is derived, a result of a sampling procedure. The Process Algebra model reverses this relationship. The discrete subsets Y are considered to be fundamental, their elements representing the actual occasions of Whitehead's process theory. The elements of H(Y) are the ontological wave functions, and the elements of H(X) are derived (emergent) through an (arbitrary) interpolation procedure. The elements of Y are considered to be generated by process, P, and the value (y) assigned to a point y in Y is also generated by P by causally propagating specific information from prior actual occasions to nascent actual occasions by means of a causal propagator, K. The resulting wave function (z) = y∈Y (y) y (z) is thus emergent. The discrete subsets are called causal tapestries and their individual points are called informons. The triad of prior causal tapestry, process, nascent causal tapestry forms a compound present.
It is essential to understand that all of the physics takes place on the causal tapestry Y. The space X is treated as emergent. Interpolation may be used to recover all of the physics on the emergent space X. The informons represent the fundamental elements of reality. Information among the elements is propagated only in a locally causal manner. Note that these informons do not constitute hidden variables. There is no additional parameter space associated with these informons. They constitute the fundamental elements of space-time, and their causal relationships are space-time. Moreover, they are the wave function. The wave function (z) is not a function of these informons, it is these informons. The process model thus possesses local realism without hidden variables. This is an important distinguishing feature from other contextual hidden variable models. Moreover, the wave function in the process algebra framework is both ontological and epistemological. More about this will be discussed later. As informons are the formal representation of actual occasions, it is important to note that informons do not move in space-time. Information propagates, informons do not. They merely come into existence and then fade away. Informons may be generated in a discontinuous manner. This does not violate special relativity since no information is transferred between the informons that comprise a causal tapestry, only causally from prior to nascent causal tapestry.
A simple visual analogy might help. Think of space-time like an LED display, with each active LED element representing an actual occasion. These LED elements are lit at random, but on an ultrafast, imperceptible time scale. The resulting image represents our observable reality. Processes are represented by the signal which determines which elements are lit.
Each informon takes the form: [n]<p n ;m n :φ n (z); Ŵ n >{G n } where 1) n is a heuristic mathematical label, 2) p n is a structured set of intrinsic properties, 3) m n : φ n (z) is a pair of extrinsic properties, 4) Ŵ n is the local coupling effectiveness, 5) G n is a causally ordered collection of informons, with causal metric ρ, called the content (based on an idea of Markoupoulou [124]). The union of content sets over all informons in the causal tapestry must itself form a causal set [116,124,125]. The causal distance is related to the depth of the causal structure, and the delay in formation flow (important in the case of non-zero rest mass).
The brackets [,],<,>,{,} are simply delimiters. The local process strength at an informon n is given as Ŵ * n Ŵ n . The information residing in the informons of the content is utilized by the generating process to create the informon. The intrinsic properties p n are attributed to the generating process P and imparted to each informon generated by P. The extrinsic properties are unique to each informon but are frame dependent. Each informon n is interpreted as a point m n (causal manifold interpretation or embedding) in some causal manifold M. Its content set G n causally embeds into M. Each causal tapestry forms a causal antichain in M, and thus represents a discrete sampling of a spacelike hypersurface in M. Each informon n is associated with a local Hilbert space interpretation of the form φ n (r) = Ŵ n f n (r, m m ), the Hilbert space H(M) being that over the causal manifold M. Each causal tapestry I is associated with two different maps: a tapestry realization (or allowing a slight misnomer, a tapestry "wave function") of the form (n) = Ŵ n , and a global Hilbert space interpretation over the causal manifold of the form (r) = n∈I Ŵ n f n (r, m n ).
When the informons of a causal tapestry embed into the causal manifold as a discrete lattice, it is possible to replace each f n (r, m n ) by a spatial translation (T m n f (r) = f (r − m n )) of a single generic sinc function g(σ , z) = sin σ z/σ z, so that (r) = n∈I Ŵ n T m n g(σ , r). The lattice spacing must be consistent with the Beurling density [122]. Maymon and Oppenheim [126] have shown that non-uniform embeddings still provide a highly accurate approximation using sinc interpolation so long as the spatiotemporal density is large enough. A more realistic model requires the use of non-uniform embeddings and more sophisticated interpolation techniques, such as Fechtinger-Gröchenik theory [121]. A tapestry realization is analogous to a space representation of a wave function. There is a dual causal tapestry which can be formed using the duals of the content sets and which gives rise to an analog of the momentum representation, but this construction will not be needed here.
A process generates individual informon in a series of short rounds, collectively forming a round, in which information is propagated. A causal tapestry is generated in a series of rounds, forming a complete generation. Processes possess three additional intrinsic characteristics: 1) r, the number of prior informons whose information is incorporated into an informon n. It is also the cardinality of G n , and the number of short rounds needed to form n. 2) N, the number of informons in each generation, and thus the number of rounds and the cardinality of the causal tapestry I. 3) R, the number of informons generated per round. A primitive process has R=1. Otherwise the process is compound.
Process properties include invariants such as charge, rest mass, mathematical type (scalar, spinor, vector, tensor, real, complex, quaternion etc.) as well as conserved quantities such as energy, momentum, angular momentum. Conserved quantities are not considered fundamental but rather result from symmetries of the causal propagator [90,127]. A process can possess a well-defined energy or momentum but there is no dispersion free measure because the informons which are generated by the process are dispersed in space-time. The Heisenberg Uncertainty relations still hold. The action of a process involves: 1) The assignment of a new informon label 2) The assignment of property set p n 3) The assignment of causal relations and distances to prior informons 4) The assignment of a content set G n 5) The propagation of information from prior informons. 6) Determination of local coupling effectiveness by propagating the local coupling effectiveness from each informon in G n forward to n according to the rule Ŵ n = m∈I K(n, m)Ŵ m where the propagator K will depend upon the causal distance ρ(n, m). The propagator will be determined by particle and interacting potentials.
The dispersion of informons and subsequent causal diffusion of their information is consistent with the interpretation of the Schrödinger equation as describing a diffusion process [128]. One particular version of the process algebra model can be shown to be equivalent to Feynman path integrals [2][3][4]129], but without the interpretation of motion along all possible paths. Moreover, it can be shown that if the propagator is relativistically invariant, then the generation of informons is also relativistically invariant [2][3][4][5][6].
Interactions between processes are conjectured as being triggered by the generation of informons according to the compatibility between the processes. Compatibility between interacting complex systems is an idea first proposed by Trofimova [130]. In the current context it can be thought of as a generalization of the idea of coupling factors. Compatibility (P,M) is conjectured to be a function of fixed factors such as mass, charge, coupling constants, and of the local compatibilities. The probability of an interaction taking place (P,M) is in turn a function of the compatibility, (P,M) = χ( (P,M). The precise form of these functions depends upon the particular case. The Born rule is expected to arise from these interactions and from the compatibility, but a precise derivation is not yet in hand. If one naively applies the Born rule, then probability will be proportional to the local process strength. If so then it will be non-Kolmogorov by virtue of the presence of interaction terms The global Hilbert space interpretation is an ontological wave function, in that it describes the informons generated during one complete action of a process, and so one possible history of a quantum system. To carry out calculations, however, it is necessary to consider all possible histories. To do so requires the use of the process graph defined in the next section. It essentially is a combinatorial tool which keeps track of every possible history of the system as it evolves under a process from a fixed prior causal tapestry. Each possible history yields a distinct global Hilbert space interpretation. The Process Covering Map P(I) gathers together these interpretations into a single set valued map. From this one can form a combinatorial interpretation which can be used for calculations involving single systems. When multiple systems are involved, the process graph must be extended into a configuration space graph together with its associated configuration space covering map [2][3][4]. The details can be found elsewhere and are not needed for the arguments to follow. It can be shown [2][3][4] that for a primitive process P and prior causal tapestry I, in the asymptotic limit as Planck length and Planck time tend to 0; r, N → ∞, P(I) { (r)}, tends to a single function. Thus, in the case of a primitive process, in the asymptotic limit, the process generates only a single wave function which corresponds to the usual NRQM wave function. For a primitive process the wave function becomes both ontological and computational. This is not true for compound processes, so that the ontological wave function (global Hilbert space interpretation) which describes a single instance of reality, and the computational wave function which is used for making predictions, are no longer the same [2][3][4]. This failure to distinguish between these cases may be the source of much confusion about the interpretation of the wave function.
The process covering map gives rise to a correspondence between processes and (set-valued) operators on the space of global Hilbert space interpretations. The standard operator formalism is thus an emergent feature of the Process Algebra model arising in the asymptotic limit of infinite information and infinitesimal scale [2,4,90].
An important concept is that of epistemological equivalence. Epistemological equivalence of two processes P and Q means that their global Hilbert space interpretations, P (r), Q (r), respectively, are equal as functions over the causal manifold. In other words, If two processes are epistemologically equivalent then the specifics of informon generation do not matter in so far as NRQM is concerned. They generate the same emergent wave functions and therefore will yield the same NRQM predictions. This is useful because processes can be modeled heuristically based upon mathematical convenience just so long as they are epistemologically equivalent to any real processes. In particular one can use processes based upon combinatorial games which have particularly valuable characteristics [131][132][133]. Epistemological equivalence may also possess ontological implications in that it might be impossible on principle for macroscopic observers to be able to access information about this most fundamental level. To use a computer analogy, it is generally inadvisable for a computer program to be able to access and change its own code. Perhaps that is the case for nature as well.

THE PROCESS ALGEBRA
The various paradoxes and conundrums posed by NRQM can be addressed through the Process Algebra. Processes can interact in a myriad of ways and the Process Algebra provides the formal language for describing these interactions. The power of epistemological equivalence is that it allows for many different representations of process to be considered based on heuristic, computational, or conceptual reasons, and it ensures that the results of calculations will still agree with one another. In this it is akin to the concept of gauge invariance. The most useful such representation to date is based upon combinatorial game theory. These games have been used for decades for generating mathematical structures [130][131][132] and are used heuristically as a model for how processes generate informons.
Processes may influence one another in two different ways. The first (coupling) involves the generation of individual informons, their relative timing as well as the sources of information which enters into their generation. Coupling results in epistemologically equivalent processes, so properties are unaltered. The second (interaction) involves the activation or inactivation of individual processes and the creation of new processes. Epistemological equivalence is broken and properties are altered.
Two processes P 1 ,P 2 may be independent, meaning that the neither constrains the actions of the other in any way. This relationship is denoted simply by the comma ", ". Compound processes (R >1) can be formed from primitive processes (R = 1) by various coupling operations. A coupling affects timing and information flow. Two processes may generate informons concurrently (products) during each round, or sequentially (sums), with only one process generating informons during a given round. Information from either or both processes may enter into the generation of a given informon (free) or information incorporated into an informon by a process may only come from informons previously generated by that process (exclusive). This leads to four possible operators: 1. Free sequential (free sum): P 1⊕ P 2 2. Exclusive sequential (exclusive sum): P 1 ⊕ P 2 3. Free concurrent (free product): P 1⊗ P 2 4. Exclusive concurrent (exclusive product): P 1 ⊗ P 2 The operation of concatenation is used to denote processes that act in successive generation cycles. Thus, P 1 · P 2 (or simply P 1 P 2 ) indicates that P 1 acts during the first generation cycle, while P 2 acts during the second generation cycle.
Interactions break epistemological equivalence and can do so in myriad ways. Interactions between processes may activate an inactive process or inactivate an active process. In addition, an interaction among processes P 1 ,P 2 ,..,P n may generate a new process, P, which can be described in functional form as F(P 1 ,P 2 ,..,P n ) = P. If (P 1 ,P 2 ,..,P n ) describes a coupling among P 1 ,P 2 ,..,P n then the functional relation may be described using the operation of concatenation, as (P 1 ,P 2 ,..,P n ) P.
Since there are potentially so many different types of interactions, a set of generic operators are used to indicate the presence of an interaction with the specifics to be spelled out if known. Thus, there are 1. Free sequential (free interactive sum): P 1⊞ P 2 2. Exclusive sequential (exclusive interactive sum): P 1 ⊞ P 2 3. Free concurrent (free interactive product): P 1⊠ P 2 4. Exclusive concurrent (exclusive product): P 1 ⊠ P 2 Independence, sums and products are commutative, associative and distributive operations. Concatenation is non-commutative and non-associative in general. The zero process, O, is the process that does nothing.
An important and special form of interaction is the coupling interaction. Such interactions respect epistemological equivalence and thus are potentially reversible through a subsequent coupling interaction. An example is a rotation to a different eigenbasis as a result of an engagement with a measurement apparatus.
If the propagator is spatio-temporal invariant, so is the associated process. Since processes are independent of spacetime, their actions too are independent of any extrinsic causal manifold interpretation. They will act in the same manner regardless of where the embeddings into the causal manifold occur. Thus, if the propagator is invariant under space and time translations, so is the associated process.
Another point worth mentioning is that due to the noncommutativity of concatenation generally, there is an intrinsic temporal asymmetry within the process algebra model. Temporal evolution according to the process algebra model is not time reversible. It is quite permissible for two processes P, Q to be time reversible individually, but yet their concatenation is not time reversible. Assume that QP = PQ. If T is the time reversal operator (which means that if P assigns an informon the causal manifold interpretation (t, z), then T[P] assigns it the interpretation (−t, z)

), then T[P] = P and T[Q] = Q but T[PQ] = T[Q]T[P] = QP = PQ.
Unlike NRQM where multi-particle systems require tensor products, and QFT uses the Fock space, the multi-particle representation of the Process Algebra requires the use of a categorical co-product space. This space consists of formal, rather than arithmetic, sums of global Hilbert space interpretations. Thus, process algebra sums correspond to arithmetic sums while products correspond to co-product (formal) sums. Tensor products appear in the configuration space covering map, which again highlights the difference between the ontological global Hilbert space interpretation and the epistemological process and configuration space covering maps.
The impact of these different operations is best demonstrated using a process graph. The process graph G(P) of a process P is defined as follows: rounds 0 to N are laid out in order. At round 0 one places the informons of the prior causal tapestry. At round k, place each informon n that was generated during round k and draw a directed line from each prior informon in its content set G n to n and label it withthe causal distance between the two informons. Note that no lines link informons of the prior causal tapestry to one another or nascent informons to one another since no information passes among them. Let G(P)n = {n} ∪ G n , the subgraph of G(P) consisting of n and its content set. The process graph is used to determine the causal manifold interpretation of the nascent causal tapestry and the global Hilbert space interpretation. If a process acts on the same prior causal tapestry it may produce a different process graph, thus a different history. The process covering map gathers together the global Hilbert space interpretations of these different process graphs, thus all of the possible histories required for a sum over histories calculation. A configuration space graph and configuration space covering map can be defined for products of processes.
Let |P| denote the total number of informons generated during the current generation cycle. For any two processes P,Q we have This highlights some of the subtle differences between these operations.
The basic rules for applying these operations in combining processes are the following: 1 The free sum is only used for single systems and combining states which possess identical property sets (pure states). 2 The exclusive sum is used for single systems and combining states which possess distinct property sets (mixed states). 3 The free product is used for multiple systems which possess distinct character (scalar, spinorial, vectorial, and so on) such as coupling a boson and a fermion. It is unclear whether two bosons might couple via a free product. 4 The exclusive product is used for multiple systems which possess the same character such as coupling two bosons or two fermions.
The Process Algebra can be represented in many different ways as an algebra of processes, as an algebra of combinatorial games in one model of process, as an algebra of causal tapestry realizations, as a Hilbert space of global Hilbert space representations. Note that the latter representation is not faithful, that is it does not possess all of the structure of the Process Algebra. As stated previously, this results in a loss of causally meaningful information. To emphasize these different representations, one can describe Process Algebra elements as |P>, |G P >, | P (n) >, | P (r) >.

CALCULATIONS IN THE PROCESS ALGEBRA MODEL
To illustrate the difference in Hilbert space and Process Algebra approaches, consider first how the process algebra approach deals with superpositions. The linearity of the Schrödinger equation allows for two solutions 1 (r), 2 (r) to be summed together to yield a new solution (r) = α 1 1 (r) + α 2 2 (r) of the same equation, and therefore a possible state. In the Hilbert space of NRQM note that every space-time point r possesses wave function contributions from both states 1 (r), 2 (r) and so if interpreted ontologically this means that the system at the point r is manifesting both states simultaneously, regardless of what those states entail. This is not a problem if the wave function is interpreted epistemologically (statistically) since in that case it is merely a tool to calculate the probability of being observed in either of the two states at the point r. This works for an ensemble of particles, but what of a single particle? How does one explain fixed probabilities if, unlike in classical probability, the object being observed possesses no definite attributes until after being observed? How is it that every observer determines the same attributes and probabilities? By contrast consider the process algebra approach. Each component state 1 (r), 2 (r) is generated by its own process P 1 , P 2 , respectively. The superposition process is represented in the Process Algebra by the exclusive sum, P = α 1 P 1 ⊕ α 2 P 2 so that each sub-process generates a unique, distinct causal tapestry, I 1 , I 2 (I 1 ∩ I 2 = ∅), respectively, and no information from one process enters into the generation of any informon of the other process. The two causal tapestries embed into disjoint regions of the causal manifold. The causal tapestries are thus physically and informationally isolated from one another. The causal tapestry for P is I = I 1 ∪ I 2 and on this causal tapestry the causal tapestry wave function (n) for P takes the form 1 (n) + 2 (n), where each i (n) is extended to I by setting i (n) = 0 for n ∈ I/I i , Note that although the algebraic sum is used there is no ontological confusion since for any n, (n) receives contributions from either 1 (n) or 2 (n) but never both. Note though that the co-product free sum could also be used. This causal tapestry wave function describes a single system which manifests in either of the two states, which remain distinct, yet whose wave functions are intertwined. The informons that support these causal tapestry wave functions are generated sequentially, never concurrently, and so at any given moment only one of the two states is manifesting, so it only ever appears in a single state. The intertwining of the wave functions creates the impression of a mixed reality state at the macro-level, yet that is never the case at the micro-level.
In a superposition of processes, P = α 1 P 1 ⊕ α 2 P 2 the effect of each modifier α i is to modify the value of the local coupling effectiveness (Ŵ n → α i Ŵ n for example) so that (n) = α 1 1 (n) + α 2 2 (n) and the global Hilbert space interpretation is formed in the usual manner, (r) = α 1 1 (r) + α 2 2 (r).
The combined global Hilbert space interpretation thus takes the form This is a map in the Hilbert space H(M) over the causal manifold, the same Hilbert space where the NRQM wave function resides, and it constitutes the Process Algebra approximation to the NRQM wave function. Note that in moving to the Hilbert space, causal ontological information is lost since now the wave function is over space-time locations, not informons, and contributions from the sub-processes are now summed, not intertwined. This information could be preserved if the algebraic sum above were replaced by a co-product free sum, but that is not a Hilbert space property. However, it would not be possible to carry out the usual NRQM calculations in that case.
The concept of informon being new, there is currently no direct evidence to suggest a particular model ofr their generation. Epistemological equivalence, however, allows one to side step that for the moment, and any strategy resulting in an epistemologically equivalent model suffices. Several such models have been presented in the literature [2][3][4][5][6]90]. Work is underway to model informon generation as a diffusion process as has been suggested in the literature for the Schrödinger equation [111,114,128]. The discussion here is meant as an in-principle demonstration of the Process Algebra framework and not a final theory.
The causal tapestry wave function is ontological, representing a single complete action of a process in generating informons (and thus a region of space-time events). The causal tapestry wave function represents the outcome of interactions with other processes only with respect to that single process. Statistical calculations require the use of the process and configuration space graphs and covering maps. The compatibility between processes is conjectured to be a function of the local coupling effectiveness Ŵ n , which in turn reflects the effect of myriad local interactions. If the effect of these interactions is summarized in a potential, then it seems reasonable, as an initial approximation, to assume that the local coupling effectiveness will depend in some manner upon the Lagrangian. The probability of an interaction, triggered by the generation of informons, must be positive real valued and is conjectured to depend on the compatibility. It seems reasonable, therefore, to conjecture that the compatibility (hence probability) should depend on the local process strength, Ŵ * n Ŵ n which is both positive real and relativistically invariant. Therefore, making the simplest assumption, the local probability is assumed to be given by the Born rule, P n = Ŵ * n Ŵ n . If an interaction depends upon the presence of several informons A = {n i }, then the probability depends upon P A = m∈A Ŵ * n Ŵ n , the local process strength over A. As in NRQM these local coupling effectiveness values can be normalized relative to the global process strength P I = n∈I Ŵ * n Ŵ n = | (n)| 2 . Recall from a previous section that interaction effects are already encoded within the Ŵ n . By analogy with Dirac's bra-ket formalism, one can introduce a scalar product on the causal tapestry wave function of the form < (n)| ′ (m) >= n,m∈I * (n) ′ (m)δ(n, m) where δ(n, m) is akin to a differential. The percentage of contribution to the strength by the process generating (n) to the process generating ′ (m) is given by | < (n)| ′ (m) > | 2 . The global strength is given as | (n)| 2 = n∈I m∈I * (n) (m)δ(n, m) = n∈I Ŵ * n Ŵ n . For example if (n) = α 1 1 (n) + α 2 2 (n) then < 2 (n)| (n)> = n∈I 2 m∈I 1 ∪I 2 * 2 (n) (n)δ(n, m) = α 2 n∈I 2 Ŵ * n Ŵ n = α 2 | 2 (n)| 2 = α 2 assuming suitable normalization of the process strength These calculations can also be carried out using the global Hilbert space interpretation with the usual Hilbert space scalar product < (r)| ′ (r) >= M * (r) ′ (r)dr.
Conisder a regionÂ of the causal manifold containing causal manifold interpretations of informons in the set A. Then we may define PÂ = Â * (r) (r)dr = Â n,n ′ ∈I Ŵ * n Ŵ n ′ f n (r, m n )f n ′ (r, m n ′ )dr = Â n,n ′ ∈A Ŵ * n Ŵ n ′ f n (r, m n )f n ′ (r, m n ′ )dr since the f n decrease in value rapidly away from A. Similarly, Thus Â n,n ′ ∈I Ŵ * n Ŵ n ′ f n (r, m n )f n ′ (r, m n ′ )dr ≈ M n,n ′ ∈A Ŵ * n Ŵ n ′ f n (r, m n )f n ′ (r, m n ′ )dr = n∈A Ŵ * n Ŵ n again since the f n are orthogonal to one another. Likewise Thus, the basic calculations can all be carried out on the causal tapestry itself and so as stated previously, all of the relevant physics occurs on the causal tapestry. The global Hilbert space interpretation provides an observer dependent link to the NRQM Hilbert space which facilitates some calculations and comparisons to NRQM but is not necessary for the physics.
Each causal tapestry wave function (n) provides an ontological representation of the action of a single process, and thus the history of one occurrence of that process. This is insufficient to carry out computations because only N informons will be generated out of a possible infinitude. The causal tapestry wave function may suffice for primitive processes under certain asymptotic conditions, but it fails for compound processes as it generates only N informon tuples whereas at least N R must be determined. For computations one must resort to the process covering graph. There all possible causal tapestry histories are gathered (by union) into a single causal graph and the causal tapestry wave function is generated on this graph, as is the global Hilbert space interpretation. These maps now contain sufficient information about all possible evolutions to make calculations possible. However, the process graph and its wave functions are not ontological; they are merely epistemological structures used to carry out calculations.
The information incorporated into the local coupling effectiveness (and local process strength) takes into account local effects both of and upon the generating process. In the course of generating informons it is presumed that there will be interactions with other processes. When these other processes give rise to relatively persistent macroscopic or classical-like entities it is convenient to summarize these local effects in terms of field notions such as the potential field. The local coupling coefficients should take these effects into account so a natural first choice is to try a Lagrangian approach.
Assume now that a suitable strategy for generating informons has been adopted and that the informons are generated so as to form points on a 4D lattice with lattice spacing (l P ) 4 where l P is the Planck length. For simplicity, assume further that one has a primitive process (R=1) in some energy eigenstate (so we need only consider a single causal tapestry) and that N=r=|c|, the value of c to the nearest integer and stripped of its units. Assume that each generation of an informon occurs in Planck time t P and that each complete action of the process generates |c| informons in time |c|t P . This time interval corresponds to a length ct P =l P. . Thus, each causal tapestry is separated from the next by an interval ct P =l P , hence the choice of lattice spacing. Label each causal tapestry by its generation number n so that the time coordinate for the nth tapestry is n|c|t P and each causal tapestry corresponds to a 3D spatial lattice.
Initial states are generally assumed in NRQM. In the Process Algebra framework, since causal tapestries are generated, so must initial states, and so the question of initial states is actually rather important and subtle within the Process Algebra framework. However, a discussion of this problem would detract from the main focus and an initial causal tapestry I 0 will simply be assumed with causal tapestry wave function 0 (n), global Hilbert space interpretation 0 (r) and corresponding NRQM wave function 0 (r) satisfying 0 (m n ) =ˆ 0 (m n ) on the embedding sites of informons. Each subsequent causal tapestry is labeled by its generation number n and denote the corresponding 3D sublattice as L n .
An effective space-time approach is due to Feynman [129]. However, this is presented here only to provide an "in-principle" demonstration. The Feynman propagator allows for long range transfer of information (long range "paths") and is suitable for NRQM where causal has a different definition but it is not causal in the relativistic sense. Restricting the number of informons being generated serves to truncate the calculation and avoid this problem but then to be precise more definite estimates of the Kernel approximation are needed to determine the size of the resulting amplitude error. These details warrant another paper. If the Lagrangian for a scalar particle of mass m is L = p 2 /2m − Vthen on the causal tapestry one might expect it to take the form L(n, n ′ ) = md(n, n ′ ) 2 /2|c| 2 t 2 P − V(n) where n and n' refer to informons on the nascent and current tapestries, respectively, and d is the causal tapestry distance. Then S[n, n ′ ] = L(n, n ′ )|c|t P so the propagator may be written as P n,n ′ = l 3 P A 3 e i/hS[n,n ′ ] Note that this is defined entirely on the causal tapestry.
The tapestry wave function can be extended from the tapestry I n to the sublattice L n by the convention that m (n) = 0 if n / ∈ I n and that P n,n ′ = 0 if no information propagates from n to n'. Assume that the process has generated tapestries up to generation m and consider generation m+1. Assume that N informons have been generated and consider the causal tapestry wave function m+1 (n). The value of the local coupling effectiveness at the informon n m+1 is Ŵ m+1 n m+1 = n m ∈I m P n m+1 ,n m Ŵ m n m . Expanding back to the initial state one has Ŵ m+1 n m+1 = n m ∈L m · · · n 0 ∈L 0 P n m+1 n m · · · P n 1 n 0 Ŵ 0 n0 = Ŵ m+1 n m+1 = n m ∈L m · · · n 0 ∈L 0 P n m+1 n m · · · P n 1 n 0ˆ 0 (m n 0 ) = Thus, we find that the global Hilbert space interpretation The accuracy of the approximation will depend upon N,r and the value of l p . Parzen's theorem [121] can be used to show that it will become increasingly exact as N, r → ∞ and l P → 0. In the Process Algebra framework, the usual approximation of h→0 does not lead to the classical realm but merely to NRQM. This is the reason that NRQM is considered to be an effective theory in the limit of infinite information N,r→∞ and continuity or infinitesimal scale h→0 Classicality is thought to be a consequence of the complexity process interactions.
The causal tapestry wave function is discrete and finite and so clearly can diverge in values from the corresponding NRQM wave function. These differences can be due to a variety of factors: truncation errors due to the finite number of informon generated, aliasing errors due to the informon density, amplitude errors due to the discrete approaximation to the Kernel integral, time-jitter errors due to non-uniform spacing of informons and information loss if informons are generated non-contiguously [121]. However, if the density of informons exceeds the Beurling density for the NRQM wave function and if the wave function tends to zero at infinity rapidly, then the difference between these two may be surprisingly small. In the case of a primitive process generating a single eigenstate the NRQM wave function can be achieved in one of two ways-either or by the asymptotic procedure described above or by resorting to the process graph and covering map. Both techniques allow for information to divege to infinity, thus capturing all possibly informon generation sequences and therefore, information paths. The causal tapestry itself, however, is always discrete and finite and possesses finite information, which will result in a disagreement between the global Hilbert space interpretation and the NRQM wave function. This error will depend upon the accuracy of the approximation to the integral M K(r j ′ , r j )φ j (r j )dr j , the deviations from uniformity of the informons, the values of N,r and of t P , l P . Determining the error in the general case is probably intractable but there are results for special cases, particularly when the informons occupy contiguous sites in a regular lattice such as in the example above. In one dimension, if the NRQM wave function (t) satisfies |ˆ (t)| ≤ M|t| −γ for 0 < γ ≤ 1, | M K(r j ′ , r j )φ j (r j )dr j − (r j ′ )| ≤ ǫ, the discrepancy between each embedding point and its ideal lattice embedding point is less than δ, and the truncation number r = 2[W 1+1/γ + 1] + 1, then according to a theorem of Butzer [121], the error E satisfies ||E|| ∞ ≤ −K( , γ , ǫ/l P , δ/l P )l P ln l P where K = (1 + 1 γ ) √ 5e ( 14 π + δ/l P + 7 3 √ 5π )|| (1) || ∞ + ò/l P + 6e(M + || || ∞ ) Then ||E|| ∞ ≈ 10 −33 K if l P is the Planck length. Ideally, if the discrete approximation to the Kernel integral and the Kernel integral are equal, the informons lie continguously on a uniform lattice, and the informons occupy a cubic spatial region with N 1/3 informons on one side, then one can use the Yao and Thomas theorem [121] to find a rough estimate for the error E, namely Thus, a value for N of |c| yields an error of ∼10 −10 (m −3/2 ) and if N=|c| 3 then the error is ∼10 −30 (m −3/2 ). Thus, even this relatively simple Process Algebra model can reproduce, to a high degree of accuracy, any NRQM wave function which can be calculated using Feynman path integrals and whose energy is bounded by h/2|c|t P .

THE PROCESS ALGEBRA APPROACH TO THE PARADOXES
The paradoxes and dualities of quantum mechanics appear to arise from one of two main factors, either a failure to utilize a contextual probability model when analyzing an experimental situation, or an attempt to interpret a wave function in both ontological and epistemological terms. The former was pointed out by Dzhafarov [60] in his contextuality by default analysis of the two slit experiment. The difference in outcomes is a consequence of contextuality. It can also be seen in the analysis of wave-particle duality by Ionicioiu et al. [18] who used a Bell type argument to try to show that the wave-particle distinction for a particle cannot be non-contextual.
The Process Algebra approach focuses upon the latter factor, the confusion of ontology and epistemology. Only the causal tapestry realization is truly ontological, although the global Hilbert space interpretation preserves ontology at informon embeddings sites. For epistemology one requires the use of the process and configuration space graphs and the setvalued process and configuration space covering maps. These lead to the usual NRQM wave functions, which are viewed as being epistemological. The resolution of the paradoxes is not to be had through an ever more clever or convoluted elaboration of standard quantum mechanics using the Hilbert space. It is not a problem of computation. The problem lies in attempting to use the Hilbert space formalism to provide an ontological model of quantum mechanics. The Hilbert space is simply too coarse grained to carefully distinguish between distinct ontological states. This coarse graining is adequate for carrying out calculations of quantum mechanical statistics, but not for the purpose of ontology. The problem is that the Hilbert space serves as a quotient space relative to the Process Algebra, which results in a loss of ontological information. The Hilbert space conflates information relative to distinct ontological states leading to confusion when attempts are made to provide an ontological interpretation. This does not occur in the Process Algebra framework. Thus, the problem is not with reality, but rather with the mathematical language used to represent reality.

CONCLUSION
Theoretical and experimental evidence strongly points to the end of local non-contextual hidden variables. This, however, is not the end of local realism. The fundamental result is that reality is contextual. This is entirely compatible with locality. There is no need to introduce non-locality or "spooky action at a distance." There need be no conflict between quantum mechanics and relativity. The cost, however, is to abandon the principle of continuity and accept a process generated reality. A specifc model, the Process Algebra model is presented as an in-principle demonstration of model which is generated, causally local, Lorentz invariant, contextual, without hidden variables and which. can reproduce the results of NRQM to as high level of accuracy. It is suggested that the paradoxes that plague quantum mechanics are due to the inability of the Hilbert space formalism to correctly distinguish between ontology and epistemology. The Process Algebra model corrects this deficiency and promises the possibility of a paradox free quantum mechanics. Hopefully this paper will encourage further research.

DATA AVAILABILITY STATEMENT
The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author/s.

AUTHOR CONTRIBUTIONS
The sole author is responsible for the literature review, conceptualization, analysis, and preparation of the manuscript. Any errors and omissions are their sole responsibility.

ACKNOWLEDGMENTS
WS wishes to thank Irina Trofimova for countless discussions related to process ideas. WS would also like to thank Andrei Khrennikov and Etibar Dzhafarov for many valuable discussions. Further, WS wishes to thank Robb Mann, Rafael Sorkin and Achim Kempf for their support during the gestation of the early version of these ideas. WS also wishes to thank the reviewers who provided very helpful commentary and additional references which helped to improve the quality of the paper.