The Generative Programs Framework
Abstract: Recently there has been significant interest in using causal modelling techniques to understand the structure of physical theories. However, the notion of ‘causation’ is limiting - insisting that a physical theory must involve causal structure already places significant constraints on the form that theory may take. Thus in this paper, we aim to set out a more general structural framework. We argue that any quantitative physical theory can be represented in the form of a generative program, i.e. a list of instructions showing how to generate the empirical data; the information-processing structure associated with this program can be represented by a directed acyclic graph (DAG). We suggest that these graphs can be interpreted as encoding relations of ‘ontological priority,’ and that ontological priority is a suitable generalisation of causation which applies even to theories that don’t have a natural causal structure. We discuss some applications of our framework to philosophical questions about realism, operationalism, free will, locality and fine-tuning.
1 Introduction
Recently the quantum foundations community has been exploring the possibility of using causal models encoded in directed acyclic graphs (DAG) to study the structure of quantum mechanics [1, 2]. This approach is based on the classical causal modelling framework advocated in particular by Pearl [3], Spirtes, Glymour, and Scheines [4], and Woodward [5], which has subsequently been generalized to a quantum causal modelling framework making use of quantum operations and quantum channels [6, 7, 8, 9]. Its proponents sometimes argued that realism, by definition, must take the form of assertions about causal structure [2].
It is not hard to understand the appeal of the causal modelling approach: it is precise and quantitative and offers a way of thinking about quantum mechanics which is clearly ‘realist’ in the sense that it aims to explain rather than merely describe physical phenomena. However, causal modelling was developed for the study of macroscopic causation, not the microscopic world, and there are serious questions about whether it can reasonably be extrapolated to the microscopic world. In particular, fundamental physics shows almost no sign of the asymmetry that is usually taken to characterise the causal relation, and a number of credible accounts of causation suggest that it arises only at the macroscopic level, as a result of the thermodynamic gradient ( [10, 11]), interventions ( [3, 5]), the perspectives of agents ( [12]), or some such feature of reality which plays no role in fundamental physics. So it could be the case that causal DAGs are simply the wrong tool to use to describe fundamental physics. Woodward and Pearl themselves have both sounded a note of caution around the use of causal modelling in quantum mechanics - Pearl considers that the approach should only be applied to deterministic phenomena, and meanwhile Hausman and Woodward write, ‘If the measurement results are not distinct events or do not result from distinct mechanisms, then it must be mistaken to think of the EPR experiment as having a common cause structure or as having a structure in which the measurement results are related as cause and effect’ [13]. Thus we consider that there is a clear need for a more general framework - one which doesn’t presuppose an underlying causal structure, but which still allows us to offer precise, quantitative explanations of physical phenomena. Such a framework should be able to accommodate both causal and non-causal approaches, in order that we can unify all such approaches in a common framework and provide definitions and arguments which apply equally well to all of them.
In this article we set out to provide such a framework, based on the notion of a ‘generative program’ which can be represented as a DAG. Any theory which describes empirical regularities can be written in the form of a generative program, even if it is a purely operational theory. However, like the DAGs used in causal models, these DAGs can also be interpreted as putative descriptions of ontological structures: specifically, we suggest that they can be understood as encoding the relation of ‘ontological priority,’ which has similarities with causation but is significantly more general. Thus we suggest that causal modelling and causal language in the foundations of physics should generically be replaced by accounts in terms of ‘ontological priority,’ with causal terminology being used judiciously only in instances where a causal account is clearly appropriate.
After introducing the framework, we demonstrate several useful applications of it. First, we show how generative programs are interpreted differently by operationalists and scientific realists, thus providing a clear characterisation of scientific realism which does not lean on assumptions about ontology or causal structure. Second, we discuss how notions like ‘locality’ and ‘contextuality’ are represented in the generative program framework, and thus argue that these terms make sense only within a realist approach to physics - theories construed purely operationally cannot be local or non-local, and they cannot be contextual or non-contextual. Third, we discuss how the generative program framework can help clarify long-standing questions about free will and how it relates to physics.
2 Generative Programs
Our aim is that the generative program framework should be capable of accommodating all physical theories and interpretations of physical theories. We will understand a physical theory to be any formalism which generates some empirical data, perhaps conditional on a set of inputs; our framework thus excludes putative ‘physical theories’ which are not capable of specifying any empirical data, but we would contend that if these theories do not eventually become capable of generating some empirical data, they should not be considered physical theories. It may also exclude some putative ‘physical theories’ which are qualitative rather than quantitative, since such theories may be incapable of predicting concrete outcomes for specific scenarios, but since modern physics is largely expressed in quantitative terms this will not be a problem for most physical theories relevant to current science. Finally, we note that one may also think of a physical theory as simply ‘summarising’ rather than ‘generating’ the empirical data, but we will consider summarising data to be simply a special case of generating empirical data, since a summary can be thought of as a single massively conjunctive instruction to generate a set of known empirical data.
Evidently any such physical theory must provide (at least one) set of instructions explaining how the empirical data is to be generated, possibly conditional on a set of inputs. We will refer collectively to the instruction set and inputs as a ‘generative program.’ In many cases a theory may provide several different ways of generating the same empirical data, so the theory can be associated with more than one generative program - for example, classical mechanics can be formulated in either a Hamiltonian or Lagrangian form, which would correspond to different generative programs. We put no restrictions on the form that a generative program could take - in particular, we do not insist that it involves a set of dynamical laws which take in an initial state and perform time evolution to produce later states, and therefore this framework can be used even for ‘theories’ which look very different from paradigmatic examples of physical theories based on forwards-in-time dynamical evolution. Even a physical theory which simply lists the data can still be understood as a generative program, albeit not a very interesting one because such a program does nothing to explain the regularities in the data. Any set of logical instructions that generates the correct empirical data is a viable program within the framework, and because all such programs generate the same empirical data, they are empirically indistinguishable by definition.
A generative program will generally take the form of an ordered series of steps which are performed in the process of generating empirical data, so it can be represented as a directed graph which shows the flow of information from inputs to outputs. These graphs will typically be acyclic, because inconsistencies can arise if we are required to use the output of a later step in order to carry out an earlier step - in particular, we can easily run into logical contradictions similar to the grandfather paradox. In some cases there may be mathematical approaches which can get around these problems (for example, in the study of closed causal loops [14, 15] and process matrices [16, 17, 18, 19, 20] it is sometimes possible to find a unique consistent solution for events happening in such a loop), and we hope to address these possibilities in later work, but for simplicity in this paper we will assume that we are dealing only with programs which can be represented by directed acyclic graphs.
For example, in figures 1,2,3 we show DAGs associated with several different generative programs which could be used to generate the empirical data associated with the Bell correlations - in these figures represents the preparation of the Bell pair and the true ontic state of the pair after the preparation, and represent choices of measurement directions, and represent the actual measurement settings, and and represent the outcomes of the two measurements. Here and throughout the paper we adopt the convention that the ‘direction of time’ as perceived by human observers runs from the bottom of the page towards the top; this representational choice is not part of the definition of the DAG but it makes the graphs easier to read. We use dark arrows to indicate the flow of information prescribed by the generative program, and we include additional dotted arrows showing the causal relations that would appear in standard ‘common sense’ causal accounts of these scenarios, so given our convention for representing time the dotted arrows will always point upwards. It can be seen that generative program DAGs may in some cases have some of the same arrows as the standard causal account, but they can also have arrows going in different directions or connecting different events.
The nodes of the DAG can loosely be understood as ‘variables’ (or sets of variables) in the sense that they can take a range of values. We will take it that a properly-specified generative program specifies clearly what variables are associated with each node and what possible range of values they have: so the ‘constrast class’ for each variable is determined by the program and we do not have to further specify a contrast class when interrogating features of the model. However, our intention is to be quite permissive about what kinds of variables might compose a node - nodes correspond to entities which are postulated to play an active role in generating empirical data, so our framework allows people with different metaphysical views about what kinds of entities there are to make different choices about the entities they place at the nodes of their DAGs.
Commonly when DAGs are used to encode causal models they are based on an interventionist understanding of causation and thus the nodes represent the kinds of observable events on which agents could intervene. But in our framework we will not insist that the nodes must be things on which observers can intervene: a single node could include variables pertaining to events widely distributed across spacetime, or could include variables describing unobservable elements of ontology like fundamental particles, or even variables describing things which are not in spacetime at all, so for example one could imagine a node corresponding to the value of some constant of nature. Indeed, for those who adopt a robustly realist attitude to laws of nature such that laws are understood as playing an active role in producing the content of reality, it might sometimes be appropriate to have a node of the DAG corresponding to a law, thus making explicit the role of the law in the generative process. (Our framework, however, does not make any assumptions about the nature of lawhood: those who do not think of laws in this way can also represent their views in the framework, since they can simply refrain from postulating DAGs which include laws as nodes). As long as the program specifies clearly that these things are variables and tells us what range of values they can take, they are all legitimate nodes in the DAG.
For a given DAG it may not always be possible to distinguish cleanly between nodes corresponding to variables which are observable and nodes corresponding to variables which are not unobservable - as demonstrated by the opponents of logical positivism and constructive empiricist, this distinction cannot usually be drawn in a completely precise way [21, 22]. And indeed, no part of our framework depends on the claim that DAGs can be neatly divided into observable and unobservable parts; in general we will be interested in DAGs which have substructures that can be recognised as corresponding to some of the empirical predictions of quantum mechanics (e.g. nodes corresponding to quantum measurement outcomes) but beyond this we will not need to specify which nodes feature directly in conscious experience and which do not.
There is also a further question about whether conscious experience supervenes only on (some subset of) the nodes of the DAG, or whether observers can be understood as having some empirical access to the structure of the DAG as well - for example, those who espouse what we will refer to as the ‘direct-access’ approach to the flow of time believe that our perception of time flowing is the result of some kind of direct empirical access of the process of ‘temporal becoming’ in which the future is generated from the past [23], so direct-access theorists will presumably want to say that conscious experience supervenes on the arrows as well as the nodes of our DAGs. By contrast, those who espouse a ‘perspectival’ approach to the flow of time believe that our experience of the flow of time is some kind of illusion arising from the nature of our perspective on reality [24, 12], and they will presumably want to say that conscious experiences supervenes only on the nodes. Our framework can be used by either direct-access theorists or proponents of the perspectival approach - the two groups will simply draw different conclusions about which DAGs are empirically equivalent.
We will assume that for each generative program there exists a ‘maximally fine-grained’ version of the corresponding DAG, i.e. the graph contains as many distinct nodes as allowed by the generative program. Specifically, this means that variables should be separated into distinct nodes if the program calculates them in two distinct steps. For example, one might imagine that a generative program for the Bell correlations generates first one measurement result and then the second (possibly conditional on the result of the first), so the measurement results can be treated as two separate nodes; but some generative programs for the Bell correlations might instead generate the two measurement results jointly conditional on the two measurement directions, so the measurement results cannot be treated as two separate nodes. The example DAGs we show in this article will not always be maximally fine-grained, as sometimes it is easier to see the overall graph structure if we combine several nodes into a single composite node, but it is important to keep in mind that the maximally fine-grained version is the one which should be employed when we are making judgements about properties of the program like locality and superdeterminism (see section 4).
2.1 Original Vertices
Usually a generative program will require some inputs which do not depend on anything else within the program, which we will refer to as ‘original vertices.’ For example, in traditional deterministic dynamical theories the initial conditions are ‘original vertices,’ as are the time-evolution laws, and the generative program determines the remaining course of history entirely from those original vertices. Similarly, in operational theories the agent is usually regarded as external to the physical theory and thus in an operational theory the decisions made by agents about actions like state preparations and measurements are ‘original vertices.’ On the other hand, physical theories with a more realist flavour may aspire to model agents within the theory, and therefore they may not treat the decisions made by agents as original vertices: instead an agent may be regarded as a physical system which takes inputs (such as sensory information and memories of the past) and produces outputs according to some internal process which can in principle be predicted within the physical theory.
By treating a vertex as an original vertex we are adopting an official attitude of agnosticism towards its origins, i.e. we are stipulating that its value is simply to be taken as given and there is nothing more to be said about where that value comes from. So for example, one should not imagine that the value of an original vertex is selected in a probabilistic way - that would involve postulating an objective probability distribution and measure outside of the generating program, but by stipulation if such a thing is known it should be included in the generating program, so a variable of this kind would not be original. However, of course we may still assign epistemic probability distributions over the values of the original vertices, provided we are clear that these distributions represent our own uncertainty about their values rather than any objective probabilistic generating process. And since there are by definition no known facts which could play a role in determining the value of , we may invoke the principle of indifference to say that any agent making use of the generating program to arrive at predictions should assign a uniform epistemic probability distribution over the possible values of the original vertex. Here we stipulate that original vertices should always be defined in such a way that a uniform epistemic probability distribution over the values of the variables at the original vertex will induce epistemic probabilities downstream which match the predictions of the theory, ensuring empirical adequacy - in this context ‘uniform’ means simply ‘uniform according to the natural Lebesgue measure over the reals,’ but this is merely a convenient choice, since one could use any other range and measure provided that subsequent processing steps in the generating program are suitably adjusted to ensure the program predicts the expected probabilities.
Of course, there may also be cases where a generative program does postulate an objective probability distribution over the value of some variables, so we can also have probabilistic variables, which are different from original vertices. In this paper we will take it that probabilistic variables can be modelled by an original vertex combined with another vertex representing the probability distribution specified by the generative program - for example, if the program prescribes that a certain variable has value with probability 0.75 and value with probability 0.25, we could model this using an original vertex which can take any value in the range ,and an additional vertex whose value will here be set to , and then we combine the two vertices to give a third vertex according to the prescription: if then , otherwise if then . Because is an original vertex, no objective probability distribution can be postulated over it, but as we have just specified, users of the program will associate it with a subjective probability distribution, which by convention we are taking to be a uniform distribution over the range , and thus this program will reproduce the desired probability distribution over and . Note that usually vertex will have other arrows going into it representing the prior factors which determine the probability distribution (here fully specified by just setting the value ) - for example, if the probabilistic event is a quantum measurement, will have arrows going into it from the quantum state and measurement direction which together determine the probability distribution for the measurement outcomes - see figure 4.
Throughout this article we will largely deal with generative programs which do not represent a whole universe but instead produce a complete physical description of some particular experimental scenario. In such cases we may have background knowledge that the original vertices are not ‘really’ original, since their values are fixed by processes in the world outside of the experimental scenario. But it is still appropriate to regard them as ‘original’ relative to the physical description of the experimental scenario because they depend only on parts of the universe not included in the scenario, so in the context of this specific scenario there is indeed nothing more to be said about where they come from. In the standard terminology of the causal modelling framework, the original vertices are ‘exogenous’ variables which are assumed not to depend on events taking place within the experimental scenario being studied. An experimental scenario or subregion of spacetime which can be given a self-contained description conditional on a set of variables treated as exogenous can essentially be regarded as something like an independent universe, with the original vertices playing the role of a ‘boundary’ for the universe and the dependent vertices making up the bulk.
2.2 Ontological Priority
Although all physical theories must provide some generative program in order to come up with some empirical data, there are different possible views that one might hold on the significance of these programs. For example, operationalists and constructive empiricists [25] do not in general commit to any ontology beyond the observable, so they will presumably not attribute any ontological significance to the generative programs they employ to arrive at empirical data - they will simply regard a generative program as a convenient codification of certain observable regularities. Realists, on the other hand, are committed to the existence of entities or structures which are not observable, and thus they may be inclined to attribute ontological significance to their generative programs. In particular, we suggest that in the most general case realists may think of generative programs as encoding information about ontological priority. ( [26, 27]).
Here we say that ‘A is ontologically prior to B,’ just in case B depends on A for its existence - for example, if one believes that the universe unfolds by a continuous evolution process involving ‘dynamical production,’ then one will naturally think that later states of the universe depend on earlier states of the universe for their existence, so earlier states are ontologically prior to later ones. In principle all sorts of things may depend on other things for their existence - for example, one might think that ‘the fact that A & B’ depends for its existence on ‘the fact that A’ and ‘the fact that B.’ However, in this article we are concerned specifically with ontological priority construed as a physical relation between distinct real entities, so we will not consider applications of it outside of the physical domain.
Ontological priority is often described metaphorically as representing the order in which god would have to create things in order to arrive at the universe - or in less theological terms, one could think of it as a computer program specifying the steps one would have to take in order to arrive at a complete universe [27, 28]. This metaphor makes it clear why it seems natural to interpret ‘generative programs’ in terms of ontological priority. Ontological priority is a reasonably common concept in philosophy, but has not yet made its way into physics in any explicit way, perhaps because physicists tend to be sceptical of notions which sound excessively ‘metaphysical.’ But really ontological priority (at least insofar as that term is understood as pertaining specifically to nomic or modal relations between distinct real entities) is not any more metaphysical than ‘causation,’ so in our view, anyone who is willing to accept ‘causation’ as falling in the domain of physics should also be willing to accept ‘ontological priority.’ After all, causation can be understood as simply a special case of ontological priority: if one thinks of the statement ‘A is the cause of B’ as an assertion that B depends for its existence on A, then this statement means that A is ontologically prior to B, and thus standard causal-dynamical physical theories would typically be associated with a generative program in which the initial condition of the universe is ontologically prior to the rest of history, which is generated from it by time evolution.
On the other hand, as shown by the diagrams in section 2, relations of ontological priority don’t have to coincide with our presuppositions about causal structure - for example, we typically imagine that causation goes forward in time, but we can certainly write down programs in which future events are ontologically prior to past ones. In this article we will use the term ‘strong causation’ to refer to relations which are self-evidently causal according to our common-sense understanding of causation (i.e. mostly relations between macroscopic events where the cause precedes the effect in time); there is probably a case to be made that other relations of ontological priority might also qualify as ‘causal’ according to some generalized notion of causality, but we will not use that terminology in this article. Now, those who espouse a direct-access approach to the flow of time typically believe that our common-sense notion of strong causation reflects our experience of a real structure of temporal becoming, and thus they may argue that DAGs in which relations of ontological priority disagree with intuitions about strong causation are not candidates to represent the actual world, since they don’t get our experiences of temporal becoming right. However, proponents of the perspectival approach to the flow of time in which experience supervenes only on the the nodes and not the arrows of a DAG will presumably take the view that a DAG in which relations of ontological priority disagree with common-sense intuitions about strong causation could be a candidate to represent the actual world.
Furthermore, relations of ontological priority can also apply to more general relata than causal relations, including relata which may not be located in spacetime, because relations of ontological priority may hold between any entities which play an active role in the way in which a generative program produces data, regardless of whether or not the associated theory is usually interpreted as saying that entities of this kind live in spacetime. For example, fundamental constants, mathematical objects like Lagrangians, or even laws of nature could stand in relations of ontological priority, even though these kinds of objects are arguably best understood as non-spatiotemporal.
2.3 Generalizing Causal Modelling
There is a well-recognised connection between causation and ontological priority. Schaffer, for example, has noted that there is a close connection between causation and a philosphical concept known as ‘grounding,’ which is related to ontological priority: ‘the analogies run deep: both (causation and grounding) feel like relations of generation, both look something like partial orders, and both can back explanation’ [29]. Schaffer therefore argues that grounding can be represented by a system of structural equations, akin to the structural equations used to represent causation in causal modelling, and thus the same mathematical formalism can be used to understand them both.111Our approach has much in common with Schaffer’s, but ultimately has different subject matter. In particular, our definition of ontological priority is both narrower and broader than the usual definition of grounding: on the one hand, grounding is usually understood as subsuming all kinds of cases where one entity depends on another for its existence, so for example one might say ‘the fact that A & B’ is grounded by ‘the fact that A’ and ‘the fact that B,’ whereas we have specified that ontological priority applies only to physical relations between distinct real entities. But on the other hand, our definition entails that ontological priority includes causation as a special case, whereas Schaffer at least does not see causation as a special case of grounding. In addition Schaffer’s primary motivation is to better understand the concept of grounding (via a non-reductive analysis), whereas our focus here is on showing how ontological priority can be applied to real scientific problems and arguing that it can be a scientifically useful tool. This means we do not have to deal with concerns like that raised by Jansson [30], who contends that using structural equations to make inferences about grounding relations is problematic because we don’t have access to the kinds of a posteriori knowledge needed to decide which variables the DAG should include or what the appropriate contrast class for each of the variables is - this issue does not arise in our case, because we are not attempting to use structural equations to make inferences about relations of ontological priority, rather we simply take it that the generative program specified within a theory stipulates what the relevant ontological variables are and what possible alternative values they have, so the purpose of the DAG is to clarify the nature of these commitments rather than to figure out what they are.
Here we follow Schaffer in arguing that because ontological priority is naturally represented by a directed acyclic graph, we may apply some mathematical techniques developed in the study of DAGS representing causal models (e.g. see refs [3] and [4]) to our own DAGs encoding ontological priority. For example, as in Schaffer’s grounding case, we can use generative programs to investigate how relations of ontological priority ground counterfactual relations - the generative program prescribes relations of functional dependence between nodes, so we can make quantitative statements about the outcome of making changes to the value of a variable at one node of the graph while leaving other inputs to the program unchanged, and moreover we can do this regardless of whether or not the node in question is one on which realistic agents could actually intervene. For example, suppose we have a generative program where the value of some constant of nature (e.g. the cosmological constant) is an original vertex, and this vertex is used to determine the state of the world at some time . Of course no real agent can intervene to vary the cosmological constant, but nonetheless we can use the generative program to imagine ‘intervening’ by changing the value of the cosmological constant whilst keeping the program and the values of the other original vertices constant: if the program is properly formulated we should be able to calculate a new state for the world at time , so we can check whether it is the case that, according to this particular generative program, the state counterfactually depends on the value of the cosmological constant. The practice of studying the consequences of varying fundamental constants is actually very common in physics (see refs [31, 32]) and our notion of ontological priorty can be understood as formalizing the reasoning behind this practice.
To make mathematical statements about the dependence relations encoded in a generative program, we will make use of a probability distribution over original vertices (although as already noted, this distribution should be understood epistemically rather than ontologically). For example, we may wish to stipulate that part of what it is for a generative program DAG to be a correct representation of a possible generative program is that that program should satisfy the causal Markov condition relative to that DAG - i.e. when we consider the outputs of the generative program given independent random choices of values at the original vertices, we should find that for each vertex , conditional on the set of its ‘parents’ in the graph (i.e. all vertices such that there is an arrow from to ) is statistically independent of all variables which are not its parents or its descendants [3]. Of course in practice the original vertices only take one value so the causal Markov condition can’t be verified using any actual empirical data, but nonetheless we can use this condition to check whether or not a proposed DAG is consistent with the generative program it is supposed to represent - if the program generates variables which fail to exhibit the right independence relations, this suggests that the generative program involves some further relations of ontological priority that we have failed to capture in the DAG.
Similarly, in causal modelling it is common to require that causal graphs should exhibit ‘faithfulness,’ i.e., when we consider independent random choices at the original vertices, two variables in the graph should be statistically independent only if they are ‘d-separated’ in the underlying graph, which is to say that any paths between them are ‘blocked’ by other variables playing a role somewhat like a common cause [3]. So in particular, a graph fails to exhibit faithfulness if there is a causal arrow from A to B but the distribution over values obtained from independent random inputs at the original vertices shows A as being statistically independent of B. And similarly, it seems natural to require that graphs of ontological priority should similarly obey faithfulness, for if there are unexpected independences in the empirical data, this may indicate that there is an opportunity to add something further to the generative program that has been proposed - we discuss this point further in section 4.1.
3 Realism versus operationalism
We now turn to demonstrating some useful applications of the generative program framework. First, we use it to make precise the nature of the commitments associated with ‘operationalism’ and ‘realism,’ Roughly speaking, operationalists are those who regard physical theories as simply providing succinctly-expressed generalizations about empirical data: they do not purport to explain the data or say how it comes about. Whereas ‘realists’ want to do more than this - they aspire to produce theories which do actually explain the empirical data. However, opinions diverge on what exactly that means, and thus there are many different forms of realism. One possibility is what one might refer to as ‘object-oriented realism.’ Realists of this type usually suggest that realism involves explaining empirical data by postulating a set of objects or entities whose behaviour is responsible for the observed effects - for example, the ‘primitive ontology’ approach to quantum mechanics argues that we will not understand quantum mechanics until we have expressed it in terms of ontological building blocks living in three-dimensional space ( [33, 34]). Another possibility is what is commonly referred to as ‘structural realism.’ Realists of this type take it that the essence of realism is explaining empirical data by appeal to some underlying objective structure - and commonly it is assumed that this structure is supposed to be a causal one. For example, proponents of the causal modelling programme for quantum mechanics argue that we must separate out elements of the theory which are merely epistemic from the underlying causal structure, which they identify with the ‘realities of nature’ ( [35, 2]).
However, in our view both of these forms of realism are too narrow. The object-oriented approach is problematic because our access to unobseravble entities or objects is limited and indirect, particularly at the small scales with which many parts of modern physics are concerned. As structural realists have been urging for some time ( [36, 37]), our data tells us most directly about structures and there is no guarantee that our attempts to infer ontology from those structures will succeed - for all we know, the true underlying ontology could be so radically different from anything in our experience that we are not even capable of conceptualizing it. The structural approach avoids this issue by focusing specifically on structure, but when it is couched in terms of causal structure it is also flawed, because this approach takes for granted that reality must have a causal structure at the most fundamental level, and we have seen that there are good reasons to think that may not be the case. This leaves us with a dilemma - what does it mean to be a realist if one does not espouse object-oriented or causal accounts of fundamental physics? If fundamental physics does not have a causal structure, does that mean it is hopeless for us to make any attempt to explain empirical data? Surely not - causation cannot be the only kind of structure which could possibly play an explanatory role. But what kind of structure could replace it?
The literature contains various suggestions for approaches to non-causal explanation which might help fill this gap. However, in many cases the ‘realist’ credentials of these approaches seem somewhat unclear. For example, Felline [38] proposes explaining the Bell correlations by appeal to structures which are not merely higher-level descriptions of more fundamental entities and processes but which are ‘mechanistically fundamental.’ This seems like a promising starting point, but since there are a large number of different structures featuring in scientific theories, more needs to be said about what kinds of structures can be ‘mechanistically fundamental’ and what realist commitments are justified by these structural explanations. The term ‘fundamental’ does not necessarily help here, because ideas around fundamentality in physics are often quite vague and different people have different intuitions about what counts as fundamental.
But the generative program framework offers a useful way of formalising non-causal structures. Specifically, we suggest that it is the structures encoded in generative programs, i.e. structures encoding ontological priority, which should replace causal structure to give a more general version of causal structural realism; structural realists can be committed to the ontological priority structure encoded in a generative program DAG in just the same way as they may be committed to the causal structure encoded in a causal DAG, and thus this framework provides a locus for structural realist commitments even in domains where a purely causal description is inappropriate.
Of course, both operationalists and realists may make use of generative programs, so to clarify the nature of the realist’s commitments we need to understand the ways in which realists and operationalists differ in their use of generative programs. To begin, we note that for most kinds of empirical data there will be many different generative programs which could predict the same empirical data. For example, suppose we observe that whenever an event of type A occurs, then event B follows with certainty, but B does not occur unless A has occurred just prior. It would seem natural to describe this phenomenon by a generative program such that whenever we generate an event of type A we then generate an event of type B at a later time, which would be represented by a DAG with an arrow going from the vertex associated with A to the vertex associated with B. However, one can also define a consistent generative program in which whenever we generate an event of type B we then generate an event of type A at an earlier time, which would be represented by a DAG with an arrow going from the vertex associated with B to the vertex associated with A. The latter possibility involves something that looks a little like ‘retrocausality,’ though the relation going backwards does not qualify as strong causation since our common-sense judgements of causation usually assume causes precede effects in time. Thus this possibility does not accord with the way we would typically narrativize this set of events, but at least for those who espouse a perspectival view of the flow of time it is still a viable program which generates the right empirical data. So we suggest the difference between realist and operationalists lies in the way they will interpret these kinds of empirically equivalent alternative programs: operationalists do not regard physical theories as having representative content or as asserting anything about reality beyond the described regularities, and therefore they cannot say that there is a fact of the matter about which entities stand in relations of ontological priority to each other, so they must take it that empirically equivalent generative programs are simply equivalent representations of the regularities observed by agents. On the other hand, realists about scientific theories typically believe that reality does have some objective structure of its own, and thus they may regard a scientific theory as putting forward a hypothesis to the effect that some particular generative program DAG is a correct representation of the actual structure of our reality. Thus for the realist, generative program DAGs are not merely recipes for arriving at empirical data, but constitute substantive claims about the structures that generate that data - so from the realist point of view generative programs can be right or wrong in what they say about ontological priority, whereas from the operationalist point of view it would be nonsensical to describe an empirically adequate program as right or wrong.
That said, we emphasize that the realist need not be committed to the claim that any particular DAG is the correct one - it is possible to be a realist whilst also maintaining a position of epistemic humility which emphasizes the fallibility of scientific knowledge. So the prudent realist will accept that we cannot know for sure that the generative program specified by a theory matches the true structure of reality, while maintaining that there is nonetheless a fact of the matter about which generative program best captures the structure of reality. We also emphasize that realists need not regard any two superficially distinct programs as being ontologically distinct - for example, most realists would probably consider that two different programs for classical electromagnetism which use different gauge fixing schemes correspond to the same generative program, since different configurations of the potentials related by a gauge transformation are typically taken to represent the same physical situation. So deciding which programs count as distinct and which parts of those programs should be reified will always be a nontrivial problem; but the fact remains that realists must believe that there is some real, ontological structure which explains why we get the data we do, and therefore once we have decided that two programs are distinct it follows that one may be better than the other at capturing this structure.
We can apply our definition to answer the perennial question of whether or not QBism is really a form of realism. QBism is an interpretation of quantum mechanics which asserts that quantum mechanics does not predict or summarize empirical data: rather it is a normative principle which tells rational agents how they should form their beliefs [39, 40]. Thus according to QBism, quantum mechanics is not a physical theory in the sense in which we have used that term, and thus it should not be associated with a generative program or DAG at all since it does not describe empirical data so there is nothing for the program to predict. Evidently if this is the end of the story QBism cannot be a form of realism, since it does not postulate any DAGs at all so it certainly cannot involve a commitment to any particular DAG. However, some QBists do want to identify themselves as realists - they believe in an external world which is responsible for our experiences, they simply don’t think quantum mechanics can be regarded as representing anything about that external world [41, 42]. So this class of QBists would presumably accept that there does exist some generative program and DAG representing the structure of the external world which is responsible for our experiences, while denying that this DAG has anything to do with with quantum mechanics. And this class of QBists do indeed count as realists according to our classification, since we did emphasize that realists need not be committed to the claim that we actually know the DAG which represents the structure of our actual world - though it might be argued that this group of QBists are taking epistemic humility too far by refusing to even contemplate what the DAG responsible for the empirical data could look like.
3.1 Intersubjectivity and Persistence Over Time
We can also apply the generative program framework to bring out some important consequences of taking an operational approach. We have argued that the founding tenet of operationalism is that all ‘empirically equivalent’ physical programs are equally good; evidently this means that the operationalist must be able to specify clearly which parts of a graph of ontological priority are to be regarded as empirical and which are not. As we have already noted, this is in general a non-trivial thing to do. Typically, operationalists seek to make this distinction by postulating an in-principle split between the agent and the rest of the world, so we are to imagine an external agent acting on reality and getting back certain responses, which are the data to be predicted by the theory222See refs [43, 44, 45] for some examples of operational formulations of theories. However, we emphasize that the authors of these papers may not necessarily identify as ‘operationalists’ - it is possible to adopt an operational approach to study the structure of theories without necessarily rejecting the possibility of an underlying realist story.. This allows the operationalist to identify what counts as ‘empirical’ in a somewhat more clean way - in particular, they don’t have to worry about which parts of the internal structure of the agent’s decision-making process should count as empirical. Evidently operationalists of this kind are restricted to choosing generative programs which represent their own choices as original vertices, although they will not be committed to any definite structure for the rest of the graph.
But what happens if we have two or more different agents? In that case, there are two options: either we adopt a first-person operationalism in which each agent aims to construct a generative program to reproduce only their own observations, or we adopt a collective operationalism in which the aim is to construct a generative program to reproduce the observations of all of the agents simultaneously. However, the latter approach seems quite hard to justify within an operationalist picture, because of course ‘the experiences of other agents’ are not directly observable or accessible any more than the other unobservable entities that the operationalist doesn’t want to commit to, so an operationalist who is committed to the reality of the observations made by other agents but not to any other unobservable entities seems to be making a somewhat arbitrary distinction. Thus, in fact operationalism is usually understood in a first-person way (this has been advocated explicitly in the context of pragmatic and operationally-inclined interpretations of quantum mechanics, as in refs [46, 47]), which is in accordance with the operationalist maxim that a scientific theory is just a tool that an agent can employ to predict the measurement results that they will witness.
What this means is that if Alice is an operationalist, her generative program will usually represents her own decisions as original vertices, but it need not represent Bob’s decisions as original vertices: he is just another physical system which she can perform measurements on and which may or may not include some original vertices. Furthermore, in the first-person approach Alice is committed to the view that all programs which predict the same empirical data for her are equally good, and therefore she cannot attach any special significance to programs which represent Bob’s decisions as original vertices: from the operationalist point of the view there is no fact of the matter about whether they are original vertices or not. So Alice necessarily has a description of the world which is different from Bob’s, because both of them give an account of reality which involves quotienting over a set of programs, and only a small subset of Alice’s programs will be in Bob’s set and vice versa. Thus the operationalist view is not compatible with the existence of a single generative program which generates the entire universe: rather operationalists must employ a different set of generative programs for each agent , such that all of the programs in represent the actions of as a set of original vertices but the programs in do not as a general rule represent the actions of other agents as a set of original vertices.
The differences between the sets of programs associated with distinct agents are even more pronounced in operational approaches to quantum mechanics. For operationalists typically hold that an agent should apply unitary quantum mechanics universally to the external physical world, and when this prescription is applied in situations like a Wigner’s friend experiment [48], it has the consequence that different agents can disagree not only on the structure of their generative program DAGs but also on the empirical data generated by the generative programs: Wigner’s friend will use a generative program which specifies a unique outcome for his measurement, whereas Wigner will use a generative program which says that the measuring device remains in a coherent superposition so there is no measurement outcome at all. The Frauchiger-Renner experiment makes this even more explicit by showing that in a certain case the programs used by different agents will predict incompatible outcomes to one and the same measurement [49]. This case, and the related Local Friendliness arguments [50, 51] also highlight that while an agent chooses inputs and receives outputs which become part of a permanent record, they model other agents as having no permanent record of their experiences; so in the quantum context, we end up with a set of disconnected generative programs used by different agents, and though these sets of generative programs will display roughly similar statistical features they nonetheless represent different physical worlds with both different structure and also different empirical content. Thus in general operational theories cannot provide us with any unified account of an intersubjective reality shared by different agents; in addition to being counterintuitive, this also raises serious questions about the scientific method and empirical confirmation in the context of operational theories (see ref [52] for a discussion of the importance of intersubjectivity in empirical confirmation).
These problems with intersubjectivity also extend to intersubjective agreement between different versions of the same observer at different times - for in an operational context it is difficult to model agents persisting over time without undermining the basic premises of operationalism. After all, part of what it is to be an agent persisting over time is to retain memories and act in ways that are partly influenced by one’s past experiences, and this surely requires that the agent is involved in some kind of two-way interaction with the systems described by the theory: the external world must have something to do with the memories and experiences that define the stable character of the persisting agent. One might expect this to give rise to a generative program in which measurement outcomes in the past are used to generate an agent’s future choices, or measurement outcomes and future choices are generated jointly by the program as in an all-at-once model. But using a program of this kind amounts to accepting that the agent is not, after all, separate from the physics: at least some parts of the agent’s decision-making processes are in fact determined within the generative program generating the empirical data. So if we allow some kind of two-way interaction between agents and their surroundings, it becomes difficult to maintain the operationalist’s strict separation between agent and world, thus undermining the operationalist’s identification of what counts as ‘empirical’ and making it unclear exactly which generative programs we are supposed to be quotienting over.
The generative program framework is useful here because it compels the operationalist to make a choice about what their position is really saying about agency: either they must use a completely distinct generative program to predict the result of each individual experimental situation, leaving us with an impoverished and fragmented approach, or they must accept that the agent is, in at least some rudimentary way, modelled as part of the generative program, meaning that we are no longer adhering to a strict operationalist position. An example of such an intermediate possibility is given by ref [53], which is a version of relational quantum mechanics with an additional postulate aiming to secure the possibility of intersubjective agreement between different agents. In this approach, facts are regarded as being relative to ‘observers’ (where every physical system can be an observer) and thus each observer, including different versions of the same observer at different times, has their own perspective on reality such that these perspectives may sometimes contain incompatible empirical facts. But the added postulate of ‘cross-perspective links’ ensures that agents can come to share facts during certain kinds of interactions, meaning that their perspectives come to agree on at least certain kinds of facts. Thus the generative program DAG for this theory would have some disconnected parts but would also contain connections between the DAGs associated with different observers, with those connections generated in the course of appropriate physical interactions. It would be interesting to see this idea made more formal - indeed, trying to come up with a well-defined generative program DAG for this version of RQM could be an important test of its coherence, because this process would require us to resolve some open questions about how exactly the connections in this picture can be constructed.
4 Nonlocality, Contextuality and Superdeterminism
We may now apply the distinction between realism and operationalism to help clarify a somewhat contested issue in physics - what exactly does it mean to give a realist account of the Bell correlations? In ref [2], for example, it is suggested that a realist account must be a causal one, but evidently this answer is not much help if in fact the Bell correlations don’t have a causal structure, as argued in refs [38, 54]. However, the generative program framework provides a straightforward way to resolve this question: any generative program which generates the empirical data exhibited in the Bell experiments can be understood as ‘realist’ provided we are committed to the view that the program corresponds to the true ontological priority structure of reality, or we are at least willing to entertain the possibility that it could correspond to the true ontological priority structure of reality. So programs in and of themselves are not realist or non-realist: rather it is our epistemic attitude towards the program which determines whether we are taking a realist position or not.
This has important consequences for our understanding of nonlocality. There are many different ways of defining the term ‘locality,’ but for our purposes it will be most useful to distinguish two important classes of definitions: some definitions suggest that a theory is ‘local’ only if it explicitly describes local mechanisms by which the empirical data is generated, and other definitions suggest that a theory is ‘local’ provided that it does not postulate any non-local mechanisms (e.g. an instantaneous wavefunction collapse) by which empirical data is generated. While people are free to use words in whatever way they wish, our own view is that the former approach is more useful, because it is neither difficult nor particularly illuminating to come up with a so-called ‘local’ model by simply refusing to give any account of the way in which the data is produced (for example, by being an operationalist or moving to a principle theory description [55]). This also accords with Bell’s own definition: he understands locality as involving the postulation of an underlying mechanism involving only dependence on a variable fixed at the time of the initial state preparation. So we prefer to say that a theory is local if it explicitly describes local mechanisms by which the empirical data is generated, non-local if it explicitly describes non-local mechanisms by which the empirical data is generated, and neither local nor non-local if it refrains from describing any mechanisms by which the empirical data is generated.
Now, many existing attempts to define ‘explicitly local mechanisms,’ assume that the mechanisms in question must be causal ones, but we prefer not do this since our aim is to avoid defining locality in causal terms. But instead we can appeal to the generative program framework, which allows for a natural generalisation of the causal approach: let us say that a theory is local if it is associated with a generative program such that a) in a maximally fine-grained version of the ontological priority graph associated with the program (as defined in section 2), no single node in the graph contains variables which are located at a spacelike separation, and b) no node containing a variable located at some spacetime point is connected by a relation of ontological priority to a node containing a variable located at a spacelike separation from , unless that connection is composed of a series of nodes each containing variables located wholly in spacetime such that each node’s variables are in the past or future lightcone of the previous variable in the sequence. To say that a theory is ‘non-local’ in Bell’s sense is to say that the theory is associated with a generative program which represents Bell-type experiments in some other way.
Requirement a) means that according to our definition of locality, theories with ‘non-separable’ objects or processes spread across space-like separation count as non-local. Requirement b), meanwhile, is written carefully so as to allow certain kinds of retrocausal theories. For example, a retrocausal model where the Bell correlations are mediated by a signal that goes into the future and then back into the past will satisfy requirement b), since the connection between the choice of measurement for one Bell event at spacetime point and the other measurement at a spacetime point spacelike separated from goes via a third node corresponding to an event located in the intersection of the future lightcone of and where the signal arrives at some future point and then begins to go backwards in time. On the other hand, a model in which the connection between the choice of measurement for one Bell event at spacetime point and the other measurement at a spacetime point spacelike separated from goes via a third node which is not located in spacetime (e.g. a modal or nomic connection) does not satisfy requirement b), since the third node is not located wholly in spacetime. Note that this requirement can be relaxed if we wish to define locality in a way that does not allow for retrocausality to be local: then it is enough to require that no node containing a variable located at some spacetime point is connected by a relation of ontological priority to a variable located at a spacelike separation from that point, since there is no way to have two spacelike separated events connected by a series of nodes each containing variables located wholly in spacetime such that each node’s variables are in the future lightcone of the previous variable in the sequence.
Note that clearly there exist both generative programs generating the empirical data exhibited in the Bell experiments which are local in the above sense (e.g. programs derived from a superdeterministic approach) and also generative programs generating the same empirical data which are non-local in the above sense (e.g. programs derived from a spontaneous collapse model). Thus, if one wants to take a position on locality in nature, one must take the philosophical position that there is some fact of the matter about what kind of generative program generates the data, which is ultimately a commitment that goes beyond just describing empirical data. The recent Nobel prize in physics was awarded to experimental tests of Bell’s theorem, so evidently this philosophical question is of great importance to physicists, which is a vindication of the need for philosophical argumentation in modern physics.
In particular, we emphasize that an operationalist approach to the Bell correlations cannot involve a commitment to any particular generative program - the operationalist may find some programs more convenient than others, but ultimately they must say that all empirically equivalent programs are equally correct descriptions of the empirical regularities. Thus, straightforwardly, an operationalist approach of the Bell correlations cannot be local or non-local in the above sense, because the operationalist account cannot be committed to local programs over non-local programs or vice versa. So, insofar as locality is to be understood as a claim about how data is produced, it is not correct for an operationalist to say that their approach to quantum mechanics is ‘local’ just because it does not postulate any explicitly non-local mechanisms - for their approach does not postulate any explicitly local mechanisms either, since they are not committed to any particular generative programs and thus they are not postulating any specific mechanisms at all. From this point of view, nothing can be said one way or another about whether an operational approach is local until we supplement it with a commitment to some particular generative programs or proper subset of the empirically adequate programs - and once such a commitment is made then the approach becomes a realist account, since being committed to a program or subset of programs in this way entails a commitment to the idea that there is some fact of the matter about the objective structure of the world. This indicates that any operationalist who claims to have a local (or nonlocal) account of quantum physics is either a covert realist or is using the term ‘locality’ in some other way that doesn’t tell us anything about how the data was produced.
Similar points can be made about the term ‘contextuality.’ In the language of the generative program framework, to say that a theory is ‘preparation non-contextual’ in Spekkens’ sense [56] is to say that it is associated with a generative program such that the steps employed by the program to arrive at empirical data for a given choice of preparation are the same for all operationally equivalent preparations, i.e. all preparations which are usually regarded as preparing the same quantum state, and to say that a theory is ‘preparation contextual’ in Spekkens’ sense is to say that it is associated with a generative program which does not satisfy this criterion. There exist both preparation contextual and preparation non-contextual generative programs which can generate the empirical data for quantum mechanics (for example, we can arrive at non-contextual programs if we allow the program to violate the assumption about how ontic states are composed which appears in Spekkens’ contextuality proof), and thus since the operationalist cannot be committed to one empirically adequate generative program over another, operationalists cannot be committed either to preparation contextual generative programs or to preparation non-contextual generative programs, so operationalist approaches cannot have either of these properties.
And the same applies to the term ‘superdeterministic.’ Following ref [57], we define No Superdeterminism as the requirement that the response function determining how a system would have responded to measurements other than the ones the observer actually made is entirely consistent with the observer’s past empirical experiences. We formalize this by defining the list of empirically plausible interventions as the set of all interventions the observer believes they can perform on the system, based on their past experiences, and the set of empirically plausible responses as the set of all responses consistent with past empirical data the observer has seen in response to past interventions of various types333To avoid confusion, we stress that a general intervention need not occur at a single event, and a general response can be nondeterministic, as long as it has a well-defined probability distribution.. The response function defined from the program must allow all empirically plausible interventions as inputs, and all outputs in the range of the function must also be empirically plausible. Otherwise the response function would entail something about the future intervention by the observer, and this entailment would be a conspiratorial or superdeterministic pre-arrangement. This definition is the minimum standard to ensure that the description of the system entails nothing about the observer’s future intervention. The underlying idea is that in a superdeterministic program, no observer is led to believe they can make a choice which is in fact physically impossible given that program.
Obviously, this definition rests crucially on a prescription about how to obtain response functions from a generative program. Taking inspiration from the possible-world semantics for counterfactuals, we suggest that response functions are to be obtained as follows: if the observer actually made measurement , then to calculate the output of the response function for a different measurement , we should take the maximally fine-grained ontological priority graph, and change one node in that graph to transform into ; then ignore everything which is ontologically prior to and apply the usual prescriptions of the generative program to calculate the values for all nodes such that is ontologically prior to . So for example, if the program generates events in accordance with the usual temporal order and the result of the measurement node is a function of the choice of measurement and the state of the system, we can simply change to in that function to calculate the output. But for some ‘superdeterministic’ programs, this won’t be possible: there will be empirically plausible choices of measurement such that the generative program just won’t say anything at all about what the outcome is for that choice of measurement, so indeed the observer is led to believe they can make a choice which is in fact physically impossible according to the program. Note that it also won’t be possible to arrive at a sensible response function if the generative program is such that the output of the measurement is actually ontologically prior to the measurement choice: in that case the program doesn’t allow us to calculate how the output would have been different if the input had been different, so again observers don’t have the choices that they imagine themselves to have.
Note it is important that we change entire nodes at once. For example, in an ‘all-at-once’ model where the whole of history is generated atemporally in one go, the whole of history corresponds to a single node, so according to the prescription above we are not allowed to adjust the measurement choice while leaving the rest of history untouched: to assess the response function we have to change to an entirely different course of history in which a different measurement was made. Provided that for every empirically plausible measurement there is some course of history which is possible according to the laws of the all-at-once model in which the observer performed rather than at the relevant point, our response function will have outputs for every empirically plausible measurement so this model won’t be superdeterministic. On the other hand, this model is non-local, since the node containing the whole course of history will include many spacelike separated pairs of events. Of course, if instead of treating the whole course of history as a single node, one instead calculates a course of history and then carefully ‘programs’ each individual particle at the beginning of time to have a trajectory compatible with the desired history, this model is no longer non-local since now the particles can all be treated as separate nodes - but such a model will usually be superdeterministic since the generative program won’t provide a prescription for how to calculate the changes to the behaviour of the other particles if we alter the behaviour of just one particle. Each of the trajectories is fully scripted in harmony in the initial state, and then none of the particles actually intervene upon one another as these scripts play out, even though they appear to interact.
4.1 Fine-Tuning
The generative program framework can also be used to gain clarity on contested issues surrounding ‘fine-tuning.’ Specifically, we will say that a generative program is ‘fine-tuned’ if two or more of the original vertices have values which appear to match in a way that is surprising, given that original vertices are supposed to be independent of one another.
A variety of well-known examples of fine-tuning in physics fit this definition. For example, the flatness problem [58]: the current density of matter and energy in the universe appears to be very close to the critical value required for a flat universe. Furthermore, natural mechanisms ensure that deviations from the critical value tend to increase rapidly over time, and therefore the initial density must be chosen very carefully to match the critical density since otherwise it would be very far from the critical value by now. Thus if we propose a generative program in which the initial density and the critical value are separate original vertices, then this generative program is fine-tuned. However, it is now generally believed that the matching can be explained by the mechanism of inflation [59], which suggests an alternative generative program in which the density is not an original vertex since it is ‘generated’ during a period of inflation which naturally forces it towards the critical density, and therefore we can get rid of the fine-tuning by moving to a different generative program.
Similarly, concerns about ‘naturalness’ in particle physics pertain to the fact that in certain calculations for physical parameters there is supposedly a precise matching between the bare quantities and the vacuum fluctuation contributions such that they cancel up to many orders of magnitude [60]. For example, this occurs in the usual calculation for the cosmological constant - the cosmological constant is very close to zero, and in order to achieve this the contribution due to vacuum fluctuations, which is proportional to the fourth power of the cut-off scale, must very exactly cancel the ‘bare’ value of the cosmological constant [61]. Moreover it seems quite natural to think that both the bare cosmological constant and the cutoff length should be original vertices, as depicted in figure 5, and yet if we propose a generative program in which they are original vertices it will be fine-tuned.
Now, we should clarify what we mean by the claim that original vertices can be expected to be ‘independent,’ for as we noted in section 2, original vertices are not in general to be thought of as probabilistic. Discussions of fine-tuning in physics often presuppose (either implicitly or explicity) some choice of measure relative to which matches between the values of constants or other original vertices can be understood as ‘unlikely’ or ‘not typical,’ but this approach is controversial [62, 63] since there is no possibility of empirically verifying such a measure and it is not straightforward to understand what could possibly ground or justify such a thing. So the term ‘independent’ in our definition shouldn’t be understood in terms of probabilistic independence - it has a less precise meaning, referring simply to the fact that by writing some vertices as original vertices, we are stipulating that there is nothing more to be said about their origins - so in particular, there is no connection between them nor any mechanism which adjusts them to match each other. And thus, since there is ex hypothesi no connection between original vertices, it is reasonable for us to be surprised when we find that that their values are nonetheless very closely matched. The point of all this is that when we find this kind of fine-tuning, we may be inclined to suspect that there is a deeper explanation for the matching of these vertices - there may in fact be some mechanism which does adjust them to match each other, in which case we may wish to expand our generative program to include that mechanism. So fine-tuning of this kind can be regarded as an indication that some part of our model may still benefit from further development: trying to find explanations for the fact that parameters in our models match in surprising ways has historically been a useful way of making scientific progress, so for example, part of the motivation for Einstein’s formulation of GR was the desire to explain the apparent coincidence of the matching values of the gravitational and inertial mass [64]. Thus looking for ‘fine-tuning’ is potentially a valuable way to decide where to focus our research efforts.
However, we emphasize that in the absence of any objectively correct measure over the values of original vertices, fine-tuning must ultimately be understood as a subjective heuristic rather than an objective flaw of the model. For without a measure there is no way to quantify how close two original vertices are or how surprised we should be by their matching - we cannot make claims to the effect that their matching is ‘highly improbable,’ or ‘atypical’ because those terms assume a well-defined probability distribution over the values of the original vertices. Moreover, even if the vertices are suspiciously close, there is always a possibility that their apparent matching is really just a coincidence. So the presence of what appears to be ‘fine-tuning’ doesn’t guarantee that there will be a deeper explanation to be found - the only way to know for sure is to actually look for an explanation and see if that effort leads to scientific progress.
It is also helpful to consider the relation between fine-tuning in this sense and fine-tuning as it is used in the causal modelling paradigm. Causal graphs which fail to obey ‘faithfulness’ as described in section 2.3 are sometimes said to be ‘fine-tuned,’ and we can understand the connection between these two notions of fine-tuning by observing that in order to have a causal model where there is a causal arrow from one variable to another but those variables appear to be statistically independent, it is necessary to ‘match up’ a variety of supposedly independent parameters of the structural equations defining the causal model, in order to ensure that the causal influences exactly cancel out in every possible configuration. So if we write down a generative program in which the parameters of the causal model are considered as ‘original vertices,’ then fine-tuning of the causal modelling type entails that the generative program will be fine-tuned. Of course, in most real situations (particularly when we are using causal models to describe causal phenomena in the macroscopic world) the parameters of the causal model will not really be original vertices and so there may be opportunities for them to arise through some dynamical or non-dynamical process which explains the connection between them - for example, this occurs in ‘equilibration’ explanations of fine-tuning, as discussed in ref [65]. However, when we are dealing with ‘fine-tuning’ not in ordinary causal models but in graphs encoding ontological priority of other kinds, it may be more plausible that the parameters of the causal model should genuinely be considered as original vertices, so failures of faithfulness may indeed indicate that the generative program is fine-tuned.
We can also use the relation between these two types of fine-tuning to observe that when we have a fine-tuned generative program, we cannot always overcome that problem by simply switching to a model in which one of the original vertices is no longer original, because that is liable to give rise to a DAG which does not obey faithfulness and thus we will still have a fine-tuning problem. For example, suppose we try to solve the fine-tuning problem for the cosmological constant [61] by postulating a generative program as represented by the DAG shown in figure 6, where the bare cosmological constant is an original vertex but the cut-off scale is no longer an original vertex - instead it is a function of , calculated in a way which ensures that will always remain small. Subsequently and are used to calculate in the usual way. But this DAG fails to obey faithfulness: although there are arrows of ontological priority from both and to , and are statistically independent, and likewise and are statistically independent, since regardless of the value of and , will always be very close to zero. So although we no longer have the problem that the values of and precisely match up despite the absence of any connection between them, we instead have the problem that the specific parameters used to calculate the relationships between and and have to match up in a highly specific way to make the dependence relations cancel out, so we have not in fact managed to eliminate fine-tuning by this simple stratagem.
5 Free Will
The generative program also offers a clear way to formalize and visualize various ongoing debates around free will, because concerns around ‘determination’ have long been central to the accounts that philosophers give of free will or its absence. Perhaps the most common concern raised about free will is that physics tells us the past determines the future, and thus human actions are fully determined by conditions in the distant past, which suggests that you are not really responsible for your actions at all - you are no more than a vessel through which the distant past exercises its influence on the present. Ref [66] puts the matter thus: ‘If determinism is true, then our acts are the consequences of the laws of nature and events in the remote past. But it is not up to us what went on before we were born [i.e., we do not have the ability to change the past], and neither is it up to us what the laws of nature are [i.e., we do not have the ability to break the laws of nature]. Therefore, the consequences of these things (including our present acts) are not up to us.’
5.1 ‘Undetermined’ free will
If ‘determination’ is interpreted in causal terms, then one natural way of defining ‘free will’ is to say that an action is free if it is not caused by anything other than the will of the agent, which in turn is ‘free’ in the sense that it is not caused by anything at all. However, within the generative program framework we may instead wish to interpret ‘determination’ in terms of ontological priority, which leads to the idea that an action is ‘free’ if it is, in some appropriate way, associated with an original vertex in a DAG representing the actual ontological priority structure of our world, as depicted in figure 7. In this section, we will discuss various different ways in which this ‘undetermined action’ conception of free will could be implemented in the generative program framework, focusing on the question of whether there is a consistent way to come up with a DAG modelling a realistic physical theory in which human actions are associated with original vertices.
First, if we are using a graph of ontological priority which coincides with a standard causal-dynamical picture of reality where the initial state is ontologically prior to all later events, as those who espouse a direct-access approach to the flow of time will generally be obliged to do, then there are quite strong limitations on the extent to which actions can be ‘undetermined.’ Indeed, in a deterministic causal-dynamical picture it follows immediately that no actions can be free in this sense, since all of the original vertices are found at the very beginning of time when presumably there are no agents around to take actions, as depicted in figure 8.
On the other hand, if we now adopt a perspectival approach to the flow of time, in which the experience of conscious agents is understood to supervene only on the nodes and not on the arrows of the DAG, then our graph of ontological priority is not required to coincide with a standard causal-dynamical picture of reality, and therefore we have many more options. For example, consider again a deterministic and time-reversal invariant theory as depicted in figure 8. Because the theory is time-reversal invariant, it is equally the case that the final state together with the laws of nature can determine the course of history, so if we take it that agents don’t have direct access into the process of ‘coming-into-being’ then there is another possible choice of DAG representing ontological priority which gives rise to exactly the same course of history but which has its original vertices at the final condition and all arrows running backwards in time, as depicted schematically in figure 9. So although it is true that according to our usual notion of strong causation, ‘our acts are the consequences of the laws of nature and events in the remote past’ it need not be the case that this convention reflects the real structure of ontological priority.
Now, one might object that although there are two valid DAGs representing ontological priority with arrows going in opposite directions, there is no such DAG in which your action is an original vertex, so either way your action is not an original vertex. But note that in our deterministic and time-reversal invariant theory, the initial and final conditions are not in any way special: in fact the state on any time-slice together with the laws of nature fully defines the course of history, as we can simply evolve both forwards and backwards in time from the given time-slice. So if we are dealing with a theory of this kind we can always choose a time-slice containing your action and then formulate a DAG representing ontological priority such that all the original vertices are on the time-slice , with arrows running both forwards and backwards in time from the vertices on , as depicted in figure 10. So, again assuming that agents do not have direct accesss into the process of ‘coming-into-being,’ there is a possible description of the theory in which your current action is indeed an original vertex, so it is free in the ‘undetermined’ sense.
Now, this rewriting only gives you ‘free will’ at a single moment in time. However, it is possible to imagine an even more radical rewriting of the DAG where the original vertices are scattered throughout time: this can be done if we can find a set of events across all of history which together contain enough information that the full set of them, along with the time-evolution laws, is enough to determine (but not overdetermine!) the entire course of history. And possibly it could be the case that some such set of events might contain the set of all human actions, or at least a reasonably large subset of human actions. If so, then it may be possible to come up with a DAG in which all of our actions are associated with original vertices, as depicted in figure 11, and thus despite the deterministic and time-reversible evolution, our choices can truly be regarded as free in the ‘undetermined’ sense.
One potential danger of scattering ‘original vertices’ across spacetime is that this could potentially give rise to either logical contradictions or fine-tuning. Of course, if one starts from a pre-existing block universe and simply identifies a set of original vertices that can be used to summarise it, the result will never contain any contradictions. But if one imagines a generative program as starting from a set of independent vertices and then producing the block from them, then in principle contradictions could arise - for example, if two distinct original vertices both fully determine a third vertex and they assign it different values. And if the values of the original vertices are specifically chosen in such a way as to avoid contradictions, this is liable to look like a kind of fine-tuning, since the vertices will not really be independent any more. So if we are going to scatter original vertices across spacetime, ideally we would like to do this in such a way that the structure of the program guarantees there will not be contradictions.
One possible strategy is suggested by Loewer’s account of free will, which allows that an action taken in the present can ‘influence’ both the past and future, where event E ‘influences’ O if and only if O counterfactually depends on E. Loewer places strict limitations on the kinds of influences an action can exert on the past - he argues for an approach known as the ‘Mentaculus,’ where we evaluate counterfactuals by holding constant the macrostate at the beginning of time and the probability distribution over microstates compatible with that macrostate, but not the specific initial microstate [67]. This prescription entails that we can in general ‘influence’ microstates in the past, but the macroscopic past history is held constant. Loewer’s account is a compatibilist one, but inspired by his work we can imagine an ‘undetermined’ version of his view in which human actions are associated with original vertices giving rise to arrows pointing both into the future and to the past. Because of the way in which Loewer limits influences on the past, it follows that the influence of a human action on an event in the past of that action can never be observed by anyone, since human actions can influence only microstates, not macrostates, and if a macroscopic observer were to use a measurement to learn about these influences then the influences would also be changing a past macrostate, i.e. the macrostate of the observer’s brain. So in this kind of model it is impossible to learn about a past microstaste in advance of the action that influences it, which means we can’t perform an observation of and then use the observation to select . Therefore despite the arrows going in both directions of time, the DAG for the ‘undetermined’ version of Loewer’s approach would remain acyclic. Loewer’s approach is presumably not the only possible way of achieving this, but it is a useful proof of principle that it is possible to come up with rules which define DAGs with unusual choices of original vertices without running into contradictions.
Of course, the mere existence of such possible DAGs does not lead straightforwardly to the conclusion that human actions are free, because if we adopt the perspectivalist position which suggest that conscious agents have no direct access into the process of coming-into-being, it follows that we cannot tell empirically which of the various possible ways of assigning ontological priority to events in spacetime is the correct one, and therefore we have no particular reason to think that the correct DAG is one in which our actions are original vertices. But we also have no particular reason to think that the correct DAG is the one where all the original vertices occur at the beginning of time, so for all we know, our actions could be the original vertices, in which case they could indeed be free choices. Thus this approach at least offers the possibility that our actions could be free in the ‘undetermined’ sense, even if it does not guarantee that this is the case. Indeed, since the perspectivalist will argue that changing the structure of the DAG while leaving the empirical data the same cannot change our experience in any way, they might be inclined to suggest that we should regard ourselves as always having free will, because our experience would be exactly the same if our actions were the original vertices.
This discussion demonstrates that there is still a possibility of ‘undetermined’ free will even in the context of deterministic laws. But of course our current understanding of quantum mechanics suggests that the laws of our actual world may not be deterministic, in which case we have even more possibilities for scattering original vertices across spacetime - for example, every time there is a random quantum event, we can have an original vertex which determines the precise outcome of a specific quantum event in order to get a DAG like the one in figure 7. So if it can be shown that random quantum events play an important role in the human brain, it would seem reasonable to assume that human actions always involve at least one original vertex and are thus ‘free’ in the undetermined sense. And in fact there are several ongoing research programs seeking to identify significant quantum effects in the brain - for example Hameroff and Penrose [68] have proposed models for long-term coherence in microtubule structures in the brain, while Kane [69] has argued that butterfly effects may magnify microscopic phenomena to higher scales so they can contribute to decision-making.
However there are some difficulties with simply associating free actions with original vertices - it is essentially an expression of event-causal libertarianism and thus is susceptible to the disappearing agent argument [70, 71, 72]. In the terminology of the generative program framework, this argument says that even if there is some original vertex which wholly or partly determines my action, it’s unclear that this original vertex has anything much to do with me, so I seem to ‘disappear’ at the moment of actually making the decision. After all, my character is in large part the result of the past events in the physical world that have shaped my life, together with genetic and biological facts which are likewise a part of the physical world: but any original vertices contributing to my actions are by definition completely independent of all of these things, so those original vertices are surely not in any meaningful sense a manifestation of me or my will. From this point of view, a ‘choice’ associated with an original vertex is no better than a completely random choice, since it has nothing to do with the actual person involved - they may determine the set of choices along with their relative probabilities, but not the choice itself. James made this point in 1907, arguing that ‘If a ‘free’ act be a sheer novelty, that comes not from me, the previous me, but ex nihilo, and simply tacks itself on to me, how can I, the previous I, be responsible? How can I have any permanent character that will stand still long enough for praise or blame to be awarded?’ [73] Similarly, various philosophers have argued that an action’s being caused by appropriate reasons is in fact the essence of free will - for example, Sartorio argues that when we act freely, our conduct is caused by reasons to act as we do and also the absences of reasons to do otherwise [74] - and from this point of view, an original vertex is not actually a meaningful expression of free will since it is by definition not caused by any reasons.
For those who are swayed by the disappearing agent argument, it would perhaps seem more natural to adopt an agent-causal libertarian approach. Here we will attempt to model Clarke’s agent-causal approach, which is based on the idea that physical events underlying beliefs and desires are not sufficient in themselves to bring about a free choice but require the agent’s exercise of agent-causal power to make those events causally efficient. This is represented in figure 12, where we add to the structure of ontological priority a ‘black box’ process, in which some information from previous physical events is synthesized together with some free volition in order to arrive at the decision made by the agent. This would still allow the decision to be at least somewhat ‘free’ in the undetermined sense.
Now, it is very natural to ask what exactly is supposed to be going on inside this black box. We can immediately conclude that the black box is not deterministic, because if it were the free volition would not actually be playing any role in deciding the output, and thus this would not represent a meaningful notion of freedom after all. So if we try to expand the contents of the box within the ontological priority framework, it seems we must represent it as something structurally similar to a probabilistic vertex: the inputs to the box are effectively determining a probability distribution, and then the ‘free volition’ is represented by an original vertex which plays the role of actually selecting the output in accordance with the specified probability distribution. But if that is the case then the DAG of figure 12 has the same information flow as the DAG of figure 11, which would entail that agent-causal libertarianism is not really meaningfully different from event-causal libertarianism once we clarify the nature of the information-flow in these models. Thus, if one wishes to maintain that agent-causal libertarianism is meaningfully different from event-causal libertarianism, it seems one must resist the move to expand the black box in this way. So the agent-causal libertarian must hold that whatever is going on inside the black box, it is neither deterministic nor random, so it apparently cannot be represented at all within the ontological priority framework described here.
From some points of view this may be regarded as a problem for the agent-causal libertarian: their account depends on postulating a mysterious ‘black-box’ form of information processing, but they cannot tell us anything about how this information processing actually works, and indeed since it is neither deterministic nor random it apparently doesn’t correspond to any known type of information-processing. People who are disinclined to believe in mysterious physical processes which can’t be clearly specified may take this as an argument against agent-causal libertarianism, at least until a better account of the operation of the black box can be provided. On the other hand, agent-causal libertarians might perhaps respond that consciousness is complex and mysterious and perhaps the exact nature of the information processing going on in the black box is just beyond the comprehension of agents like us with our limited cognitive capabilities. We will not attempt to adjudicate between these approaches here, but we think the two possible representations of agent-causal libertarianism given in figures 12 and 11 are helpful to clarify exactly what is at stake in this discussion and why one might wish for further clarification from the agent-causal libertarian.
5.2 Participatory Free Will
A completely different approach to free will can be defined within ‘all-at-once’ style models, as advocated in refs [75, 76]. Models of this kind do not postulate temporal evolution or any other directed process by which the course of history is generated: instead the laws specify a set of nomically possible histories and some history from this set is selected and instantiated in an atemporal way. The selection from the set can be thought of as akin to the selection of an initial state in an ordinary time-evolution model: we simply select and actualise the whole history rather than just the initial state.
In this kind of picture, no event in spacetime is ontologically prior to any other, as depicted in figure 13, and no set of events is really ‘the cause’ of the rest of history: events at the beginning of time have no better claim to this title than any other set of events. Rather, events happening across all of history stand in mutual, reciprocal relations to one another: there is an underlying structure of ontological priority, which includes some original vertices, but none of the original vertices are located in spacetime.
So what does this mean for free will? Well, if one is committed to the ‘undetermined’ account of freedom then one will have to conclude that in this picture nobody really has freedom, since no actions are associated with original vertices. But on the other hand, if one’s concerns about free will stem largely from the idea that all our actions are determined by factors outside of our control in the far distant past, then the fact that in these models no direction of time corresponds to the objectively correct direction of ontological priority should assuage those concerns, for this entails that it is no more true to say that our actions are determined by the distant past than it is to say that the past is determined by our actions. This leads to a conception of freedom which emphasizes the participatory nature of agency: in this picture the course of history is selected in such a way that all moments of time play an equal role, so your current actions are no more or less important than the events at the beginning and end of time. Your choices play a real part in determining the course of history, because all other events must be adjusted to fit with your choices, just as your choices are mutually adjusted to fit other events. This way of thinking allows for your actions to be ‘free’ in a genuine, ontological sense while also allowing that meaningful choices are not random but rather are appropriately linked with the events in your past that have formed your identity.
Obviously, this approach in itself is not a definition of free will: the argument applies equally well to all events in history and every action that a person takes, regardless of whether the action corresponds to what we would ordinarily call a ‘free choice’ or whether it is the result of someone holding a gun to the agent’s head and making demands. So there is still a need to formulate criteria which can be used to make individual judgements about which actions count as free for moral, legal and religious purposes. However, this approach does serve to defuse the blanket challenge to free will posed by the idea that everything is ultimately determined by the distant past; it allows for the conceptual possibility that at least some human actions may count as ‘free,’ so we can go on to decide which particular actions we wish to regard as free444Our own view is that one should not seek a univocal criterion which divides all actions neatly into ‘free’ and ‘not free’ - rather one should acknowledge that human actions are subject to a range of different internal and external constraints, and it is better to focus on understanding the complex interplay of constraints in a specific case than to assign a simple binary label classifying the action as ‘free’ or ‘not free’..
For example, it is often said that we have free will if it is possible for us to act otherwise, and determinism is supposed to undermine this kind of free will because, if we condition on the microstate of the world up until the time of the action, then only one action is possible. But if no events in spacetime are really ontologically prior to any others, then there is no reason why the whole of the past must be held constant when one assesses what is possible at a given time. Rather we are free to make contextual choices about what factors to hold constant, so for example if we are assessing whether an action is free we will most likely hold constant factors like the presence or absence of external compulsion, intoxicants, phobias etc, but we need not hold constant the exact past microstate. Now, this is a fairly standard compatibilist argument - for example, List suggests that the relevant notion of possibility in ‘it was possible to act otherwise’ involves conditioning not on the past microstate but on what he calls ‘the agential state’ which describes various psychological factors, and in many cases if we condition only on the agential state then we can conclude it was possible for us to act otherwise, so we can distinguish between cases in which we have free will and cases in which we do not [77]. The account we have given here adds legitimacy to approaches of this kind, because it entails that these approaches are not simply sweeping under the carpet the fact that the action is really determined by the whole past microstate - in fact all events determine all other events and therefore any meaningful statement of physical possibility must involve a choice about what we are going to hold constant555This does not apply to statements about physical possibility which refer to the whole universe at once, e.g. saying which entire courses of history are physically possible or impossible. However, evidently such statements are far too general to be applied to questions about free will in any particular case.. From this point of view, holding only the ‘agential state,’ or some such thing constant is no more or less correct at a physical level than holding the past microstate fixed, and thus it is perfectly reasonable to say that in some circumstances it is possible for us to act otherwise.
Note that this notion of freedom is similar to the one featuring in the definition of superdeterminism which we presented in section 4. This definition uses the past empirical experiences of an agent to define the space of possible choices which must be considered as counterfactuals, which is similar to using List’s ‘agential state’ to determine which actions were ‘possible,’ or in the terms of [57] ‘empirically plausible’: in models which are superdeterministic according to this definition, the agent’s conception of which actions they are free to perform, holding constant only the ‘agential state,’ does not match the reality of what the generative program allows, and hence the program is ‘conspiratorial’ in the sense that the agent’s experiences deceive them into believing that some impossible choices are possible. That said, even with a superdeterministic generative program, there is still some intermediate sense in which the agent can be free; provided the response function for a given system allows at least two empirically plausible inputs, and specifies empirically plausible outputs for all of them, then the agent is still free to choose among those in this same sense, even if there are other empirically plausible choices that are in fact impossible in the given program.
6 Conclusion
Within both physics and philosophy there is a long history of linking scientific realism to causation. But this approach is limiting - not every meaningful physical theory can be associated with a set of asymmetric relations linking different physical events, and in particular there is good reason to think that fundamental physics itself contains no asymmetric structures. Thus we would argue that if scientific realism is to continue being relevant to modern physics, it must cease to lean so heavily on causal notions.
Now, there seems to be a concern amongst some realists that giving up on causation means giving up on realism itself - and more specifically, there is a worry that we cannot talk meaningfully about heuristics like locality, superdeterminism, contextuality and so on without using causal notions. But in this article we have argued that this is not so. Realism does not require a commitment to causal structure in particular; it simply requires a commitment to the idea that the world has some objective structure, which can be encompassed in the claim that there is some fact of the matter about how empirical data is generated. To be precise, we suggest that realism involves a commitment to the idea that there exists some objectively correct generative program generating the empirical data composing our reality, which may or may not include substructures which can sensibly be described as causal. Furthermore, we argue that any theory which postulates an objectively correct generative program is implicitly a realist theory, despite some claims that theories like this are actually operational.
Moreover, the generative program framework can be used to offer precise definitions of terms like locality, superdeterminism, contextuality and so on, without appealing to causal notions - these claims are meaningful only if associated with some specific generative program. These definitions can also be applied to theories which do propose some causal and/or retrocausal structure, allowing a precise, unambiguous judgements as to whether they are local, superdeterministic and so on, and also making it possible to compare causal theories to non-causal theories within a single unifying framework. Thus we hope that the generative program framework will offer a way forward for scientific realism which can retain the commitment to objective, explanatory structures even if the world is not fundamentally causal.
Examples of generative programs for a number of well-known physical theories and interpretations of quantum mechanics are given in algorithmic form in the Supplemental Information (SI), along with some discussion of their properties. These are our best attempts to characterize how proponents of these interpretations describe them, but they are of course invited to provide their own improved versions of these programs. That is the purpose of this framework. To illustrate the utility of programs and their DAGs for particularly counterintuitive theories, Figs. 14 and 15 show two variant generative programs for the interpretation (see the SI) proposed by Aharonov of the time-symmetric quantum mechanics of Aharonov, Bergmann, and Lebowitz [78], one version with agent/event-causal libertarianism, and other without.
Another useful feature of the generative program approach is that putting a theory in the form of a generative program requires us to be precise about the role of agents and their relation to the rest of the physical world. This can help to highlight important structural features of theories that can otherwise be obscured by noncommittal language around agency. For example, we have noted that operationalism, particularly in the context of a theory like quantum mechanics, tends to lead to a picture of disconnected, isolated agents performing experiments with no relation to other agents or past versions of themselves. This is a significant disadvantage of operationalism which is not very evident from the language used by many operationalists in practice. We can also use the generative program framework to offer new clarity on some traditional definitions about free will - for example, it provides a sharp visual depiction of the disappearing agent argument. It can also be used to express an alternative approach to free will that arises in ‘all-at-once’ approaches to physics: freedom arises from the fact that agency is participatory, since an agent’s actions at some time are no more or less important than events occurring at the beginning or end of time.
References
- [1] C.. Wood and R.. Spekkens “The lesson of causal discovery algorithms for quantum correlations: causal explanations of Bell-inequality violations require fine-tuning” In New Journal of Physics 17.3, 2015, pp. 033002 DOI: 10.1088/1367-2630/17/3/033002
- [2] David Schmid, John H. Selby and Robert W. Spekkens “Unscrambling the omelette of causation and inference: The framework of causal-inferential theories”, 2020 arXiv:2009.03297 [quant-ph]
- [3] Judea Pearl “Causality” Cambridge university press, 2009
- [4] P. Spirtes, R. Scheines and C. Glymour “Causation, Prediction, and Search”, Adaptive computation and machine learning MIT Press, 2000 URL: https://books.google.ca/books?id=vV-U09kCdRwC
- [5] James Woodward “Making things happen: A theory of causal explanation” Oxford university press, 2005
- [6] R. Chaves, C. Majenz and D. Gross “Information-theoretic implications of quantum causal structures” In Nature Communications 6, 2015, pp. 5766 DOI: 10.1038/ncomms6766
- [7] Nikolai Miklin et al. “The entropic approach to causal correlations” In New Journal of Physics 19.11 IOP Publishing, 2017, pp. 113041 DOI: 10.1088/1367-2630/aa8f9f
- [8] Jacques Pienaar “Which causal structures might support a quantum–classical gap?” In New Journal of Physics 19.4 IOP Publishing, 2017, pp. 043021 DOI: 10.1088/1367-2630/aa673e
- [9] Jonathan Barrett, Robin Lorenz and Ognyan Oreshkov “Quantum Causal Models” arXiv, 2019 DOI: 10.48550/ARXIV.1906.10726
- [10] David Lewis “Counterfactual Dependence and Time?s Arrow” In Noûs 13.4 Wiley-Blackwell, 1979, pp. 455–476 DOI: 10.2307/2215339
- [11] Phil Dowe “Process Causality and Asymmetry” In Erkenntnis (1975-) 37.2 Springer, 1992, pp. 179–196 URL: http://www.jstor.org/stable/20012433
- [12] Huw Price “Causal Perspectivalism” In Causation, Physics, and the Constitution of Reality: Russell’s Republic Revisited Oxford University Press, 2005
- [13] Daniel M. Hausman and James Woodward “Independence, Invariance and the Causal Markov Condition” In The British Journal for the Philosophy of Science 50.4, 1999, pp. 521–583 DOI: 10.1093/bjps/50.4.521
- [14] David Deutsch “Quantum mechanics near closed timelike lines” In Phys. Rev. D 44 American Physical Society, 1991, pp. 3197–3217 DOI: 10.1103/PhysRevD.44.3197
- [15] H. Politzer “Simple quantum systems in spacetimes with closed timelike curves” In Phys. Rev. D 46 American Physical Society, 1992, pp. 4470–4476 DOI: 10.1103/PhysRevD.46.4470
- [16] O. Oreshkov, F. Costa and Č. Brukner “Quantum correlations with no causal order” In Nature Communications 3, 2012, pp. 1092 DOI: 10.1038/ncomms2076
- [17] O. Oreshkov and C. Giarmatzi “Causal and causally separable processes” In New J. Phys, 2015 eprint: quant-ph/1506.05449
- [18] Mateus Araú jo, Adrien Feix, Miguel Navascués and Časlav Brukner “A purification postulate for quantum mechanics with indefinite causal order” In Quantum 1 Verein zur Forderung des Open Access Publizierens in den Quantenwissenschaften, 2017, pp. 10 DOI: 10.22331/q-2017-04-26-10
- [19] Giulia Rubino et al. “Experimental Verification of an Indefinite Causal Order” In Science Advances 3, 2016 DOI: 10.1126/sciadv.1602589
- [20] K. Goswami et al. “Indefinite Causal Order in a Quantum Switch” In Physical Review Letters 121.9 American Physical Society (APS), 2018 DOI: 10.1103/physrevlett.121.090503
- [21] Gabriele Contessa “Constructive empiricism, observability and three kinds of ontological commitment” In Studies in History and Philosophy of Science Part A 37.3, 2006, pp. 454–468 DOI: https://doi.org/10.1016/j.shpsa.2006.06.007
- [22] Paul M. Churchland “The Ontological Status of Observables: In Praise of the Superempirical Virtues”, 1985
- [23] Tim Maudlin “104On the Passing of Time” In The Metaphysics Within Physics Oxford University Press, 2007 DOI: 10.1093/acprof:oso/9780199218219.003.0005
- [24] Kurt Gödel “A remark about the relationship between relativity theory and idealistic philosophy” In Revista Brasileira de Ensino de Física 28, 2005, pp. 521–524 DOI: 10.1590/S1806-11172006000400014
- [25] B.C. Fraassen “The Scientific Image”, Clarendon Library of Logic and Philosophy Clarendon Press, 1980 URL: https://books.google.co.uk/books?id=VLz2F1zMr9QC
- [26] Tuomas E. Tahko and E. Lowe “Ontological Dependence” In The Stanford Encyclopedia of Philosophy Metaphysics Research Lab, Stanford University, https://plato.stanford.edu/archives/fall2020/entries/dependence-ontological/, 2020
- [27] Tyler Hildebrand “Platonic Laws of Nature” In Canadian Journal of Philosophy 50.3, 2020, pp. 365–381 DOI: 10.1017/can.2019.41
- [28] Jonathan Schaffer “On What Grounds What” In Metametaphysics: New Essays on the Foundations of Ontology Oxford University Press, 2009, pp. 347–383
- [29] Jonathan Schaffer “Grounding in the Image of Causation” In Philosophical Studies 173.1 Springer Verlag, 2016, pp. 49–100 DOI: 10.1007/s11098-014-0438-1
- [30] Lina Jansson “When are Structural Equation Models Apt? Causation versus Grounding” In Oxford Scholarship Online, 2018
- [31] Craig J. Hogan “Why the universe is just so” In Reviews of Modern Physics 72.4 American Physical Society (APS), 2000, pp. 1149–1161 DOI: 10.1103/revmodphys.72.1149
- [32] Max Tegmark and Martin J. Rees “Why Is the Cosmic Microwave Background Fluctuation Level ?” In The Astrophysical Journal 499.2 American Astronomical Society, 1998, pp. 526–532 DOI: 10.1086/305673
- [33] Michael Esfeld “From the Measurement Problem to the Primitive Ontology Programme” In Fundamental Theories of Physics Springer International Publishing, 2020, pp. 95–108 DOI: 10.1007/978-3-030-46777-7˙8
- [34] Valia Allori “Primitive Ontology in a Nutshell” In International Journal of Quantum Foundations 1.3, 2015, pp. 107–122 URL: http://philsci-archive.pitt.edu/11651/
- [35] Michael Esfeld “The modal nature of structures in ontic structural realism” In International studies in the philosophy of science 23.2 Taylor & Francis, 2009, pp. 179–194
- [36] Peter Mark Ainsworth “What is Ontic Structural Realism?” In Studies in History and Philosophy of Science Part B: Studies in History and Philosophy of Modern Physics 41.1, 2010, pp. 50–57 DOI: 10.1016/j.shpsb.2009.11.001
- [37] Nora Berenstain and James Ladyman “Ontic Structural Realism and Modality” In Structural Realism: Structure, Object, and Causality Springer, 2012
- [38] Laura Felline “On Explaining Quantum Correlations: Causal vs. Non-Causal” In Entropy 23.5, 2021 DOI: 10.3390/e23050589
- [39] C.. Fuchs, N.. Mermin and R. Schack “An introduction to QBism with an application to the locality of quantum mechanics” In American Journal of Physics 82, 2014, pp. 749–754 DOI: 10.1119/1.4874855
- [40] C.. Fuchs “QBism, the Perimeter of Quantum Bayesianism” In ArXiv e-prints, 2010 arXiv:1003.5209 [quant-ph]
- [41] David Glick “QBism and the Limits of Scientific Realism”, 2019 URL: http://philsci-archive.pitt.edu/16382/
- [42] Christopher A. Fuchs “On Participatory Realism”, 2016 arXiv:1601.04360 [quant-ph]
- [43] Giulio Chiribella, Giacomo Mauro D’Ariano and Paolo Perinotti “Probabilistic theories with purification” In Phys. Rev. A 81 American Physical Society, 2010, pp. 062348 DOI: 10.1103/PhysRevA.81.062348
- [44] Samson Abramsky and Chris Heunen “Operational theories and Categorical quantum mechanics”, 2013 arXiv:1206.0921 [quant-ph]
- [45] Lucien Hardy “Operational General Relativity: Possibilistic, Probabilistic, and Quantum”, 2016 arXiv: https://arxiv.org/abs/1608.06940
- [46] Eric G. Cavalcanti “The View from a Wigner Bubble” In Foundations of Physics 51.2 Springer ScienceBusiness Media LLC, 2021 DOI: 10.1007/s10701-021-00417-0
- [47] Howard M. Wiseman and Eric G. Cavalcanti “Causarum Investigatio and the Two Bell’s Theorems of John Bell”, 2015 arXiv:1503.06413 [quant-ph]
- [48] Richard A. Healey “Scientific Objectivity and its Limits” To appear in the British Journal for the Philosophy of Science., 2021 URL: http://philsci-archive.pitt.edu/19401/
- [49] Daniela Frauchiger and Renato Renner “Quantum theory cannot consistently describe the use of itself” In Nature Communications 9.1 Springer ScienceBusiness Media LLC, 2018 DOI: 10.1038/s41467-018-05739-8
- [50] Kok-Wei Bong et al. “A strong no-go theorem on the Wigner’s friend paradox” In Nature Physics 16.12 Springer ScienceBusiness Media LLC, 2020, pp. 1199–1205 DOI: 10.1038/s41567-020-0990-x
- [51] Aníbal Utreras-Alarcón, Eric G. Cavalcanti and Howard M. Wiseman “Allowing Wigner’s friend to sequentially measure incompatible observables”, 2023 arXiv:2305.09102 [quant-ph]
- [52] Emily Adlam “Does Science need Intersubjectivity? The Problem of Confirmation in Orthodox Interpretations of Quantum Mechanics” In Synthese, 2022
- [53] Emily Adlam and Carlo Rovelli “Information is Physical: Cross-Perspective Links in Relational Quantum Mechanics” arXiv, 2022 DOI: 10.48550/ARXIV.2203.13342
- [54] Emily Adlam “Is There Causation in Fundamental Physics? New Insights from Process Matrices and Quantum Causal Modelling” arXiv, 2022 DOI: 10.48550/ARXIV.2208.02721
- [55] Adam Koberinski and Markus P. Müller “Quantum Theory as a Principle Theory: Insights from an Information-Theoretic Reconstruction” In Physical Perspectives on Computation, Computational Perspectives on Physics Cambridge University Press, 2017, pp. 257–280 DOI: 10.1017/9781316759745.013
- [56] R.. Spekkens “Contextuality for preparations, transformations, and unsharp measurements” In Physical Review A 71.5, 2005, pp. 052108 DOI: 10.1103/PhysRevA.71.052108
- [57] Mordecai Waegell and Kelvin McQueen “Refining the Assumptions of Bell’s Theorem”, 2023
- [58] G.W. Gibbons, S.W. Hawking and J.M. Stewart “A natural measure on the set of all universes” In Nuclear Physics B 281.3, 1987, pp. 736–751 DOI: https://doi.org/10.1016/0550-3213(87)90425-1
- [59] Alan H. Guth “Inflationary universe: A possible solution to the horizon and flatness problems” In Phys. Rev. D 23 American Physical Society, 1981, pp. 347–356 DOI: 10.1103/PhysRevD.23.347
- [60] Alexei Grinbaum “Which Fine-Tuning Arguments Are Fine?” In Foundations of Physics 42.5 Springer ScienceBusiness Media LLC, 2012, pp. 615–631 DOI: 10.1007/s10701-012-9629-9
- [61] David Wallace “Quantum Gravity at Low Energies” arXiv, 2021 DOI: 10.48550/ARXIV.2112.12235
- [62] C.. McCoy “Can Typicality Arguments Dissolve Cosmology?s Flatness Problem?” In Philosophy of Science 84.5, 2017, pp. 1239–1252 DOI: 10.1086/694109
- [63] Sabine Hossenfelder “Screams for explanation: finetuning and naturalness in the foundations of physics” In Synthese 198.S16 Springer ScienceBusiness Media LLC, 2019, pp. 3727–3745 DOI: 10.1007/s11229-019-02377-5
- [64] Dennis Lehmkuhl “The Equivalence Principle(S)” In The Routledge Companion to Philosophy of Physics, 2021
- [65] Denver Dash “Restructuring dynamic causal systems in equilibrium” In AISTATS 2005 - Proceedings of the 10th International Workshop on Artificial Intelligence and Statistics, 2005
- [66] Peter Van Inwagen “An Essay on Free Will” Oxford University Press, 1983
- [67] Barry Loewer “The Mentaculus Vision” In Statistical Mechanics and Scientific Explanation, pp. 3–29 DOI: 10.1142/9789811211720˙0001
- [68] Stuart Hameroff and Roger Penrose “Orchestrated reduction of quantum coherence in brain microtubules: A model for consciousness” In Mathematics and Computers in Simulation 40.3, 1996, pp. 453–480 DOI: https://doi.org/10.1016/0378-4754(96)80476-9
- [69] Robert Kane “Quantum Physics, Action and Free Will: How Might Free Will be Possible in a Quantum Universe?” In New Essays on the Mind-Body Relation in Quantum-Theoretical Perspective Berlin, Boston: De Gruyter, 2014, pp. 163–182 DOI: doi:10.1515/9783110351064.163
- [70] Derk Pereboom “Is Our Conception of Agent-Causation Coherent?” In Philosophical Topics 32.1/2 University of Arkansas Press, 2004, pp. 275–286 DOI: 10.5840/philtopics2004321/23
- [71] Derk Pereboom “The Disappearing Agent Objection to Event-Causal Libertarianism” In Philosophical Studies Springer, 2012, pp. 1–11 DOI: 10.1007/s11098-012-9899-2
- [72] Meghan Griffith “Why Agent-Caused Actions Are Not Lucky” In American Philosophical Quarterly 47.1 University of Illinois Press, 2010, pp. 43–56
- [73] W. James “Lecture III” In Pragmatism Longmans, GreenCo., 1907
- [74] Carolina Sartorio “Causation and Free Will” Oxford University Press UK, 2016
- [75] K. Wharton “The Universe is not a Computer” In Questioning the Foundations of Physics Springer, 2012, pp. 177–190 arXiv:1211.7081 [quant-ph]
- [76] Emily Adlam “Spooky Action at a Temporal Distance” In Entropy 20.1 Multidisciplinary Digital Publishing Institute, 2018, pp. 41
- [77] Christian List “Free Will, Determinism, and the Possibility of Doing Otherwise” In Nous 48.1, 2014, pp. 156–178 DOI: https://doi.org/10.1111/nous.12019
- [78] Yakir Aharonov, Peter G Bergmann and Joel L Lebowitz “Time symmetry in the quantum process of measurement” In Phys. Rev. 134.6B APS, 1964, pp. B1410