iBet uBet web content aggregator. Adding the entire web to your favor.
iBet uBet web content aggregator. Adding the entire web to your favor.



Link to original content: https://www.academia.edu/72611587/The_Fine_Tuning_for_Discoverability
(PDF) The Fine-Tuning for Discoverability | Robin Collins - Academia.edu
Academia.eduAcademia.edu

The Fine-Tuning for Discoverability

God and Cosmology

The Fine-Tuning for Discoverability Copyright, Robin Collins, 2014. DRAFT I. Introduction One of the most persuasive evidences for the existence of God from the cosmos is the argument from the fine-tuning of the cosmos for the existence of life, the so-called anthropic fine-tuning. This refers to the fact that laws, initial conditions, and the fundamental parameters of physics must be precisely set for life to exist. The relevant kind of life depends on the hypotheses that the evidence is supposed to support, which in the case of theism is embodied conscious agents who can interact with each other based on what they perceive as moral criteria. For convenience, I will simply refer to such agents as “observers.” The most commonly cited case of anthropic fine-tuning is that of the cosmological constant.i If it were not within one part in 10120 of its theoretical possible range of values, either the universe would expand, or collapse, too quickly for galaxies and stars to form. There have been a variety of challenges to the fine-tuning evidence itself, and whether it supports the existence of God or a multiverse. I have developed a detailed argument elsewhere (Collins, 2009) that the fine-tuning evidence does provide strong confirmatory evidence for theism over naturalism. Here I primarily want to explore another kind of fine-tuning and its implications for this debate: the fine-tuning of the universe for being discovered. By this fine-tuning, I mean that the laws, fundamental parameters, and initial conditions of the universe must be just right for the universe to be as discoverable as ours. After presenting examples to illustrate this kind of this fine-tuning, I will argue that if this kind of fine-tuning exists, in general it cannot be explained by a multiverse hypothesis – by far the leading non-theistic explanation for anthropic fine-tuning. Further, I will show how the idea that the universe is fine-tuned for discovery answers some other commonly raised objections against the fine-tuning argument, and finally I will look at its potential predictive and explanatory power. Finally, to be absolutely clear, my project in this paper is not so much to argue for the existence of God, but to explicate where one might look for new evidence one way or another. This is in keeping with the spirit of scientific inquiry. Background Many scientists and others have commented on the seemingly “miraculous” intelligibility and discoverability of physical reality, most famously Albert Einstein and Eugene Wigner. Recently this idea has been developed more carefully by Mark Steiner in his 1998 book Mathematics as a Philosophical Problem (Harvard University Press). Steiner presents an array of examples where, in their attempts to discover the underlying laws of nature, physicists successively used lines of reasoning that only make sense if they were implicitly assuming that the world was structured for discoverability. He concludes that the world “looks ‘user friendly.’ This is challenge to naturalism.” (p. 176.)ii In a project I am near completing, I have attempted to quantitatively test this idea that the universe is in some sense “finetuned” for discoverability by calculating the effects on our ability to discover the major and/or important domains of reality – such as cosmology, microbiology, and the past history of the earth – by varying the some of the fundamental parameters of physics. The cases I will cite involve original calculations; they all have been verified by at least two physicists. The calculations can be found on my website: just google my name, Robin Collins. Nonetheless, because they are not yet part of the peerreviewed literature, I do not expect everyone to accept their legitimacy. For those who do not, they should take this paper conditionally, as spelling out the implications this sort of fine-tuning would have if it is legitimate. I have found around a dozen cases of this kind of fine-tuning. Below are three illustrative examples involving the fine-structure constant. Except for the first example, a basic explanation of the physical lines of reasoning behind them is left for the appendix. Later, in the course of articulating the significance and predictive and explanatory power of this sort of fine-tuning, I will look at two other examples in the cosmological context. Examples of fine-tuning for discoverability. The first two examples involve the fine-structure constant, commonly designated by the Greek letter α. This is a physical constant that governs the strength of the electromagnetic force. If it were larger, the electromagnetic force would be stronger; if smaller, it would be weaker. A small increase in α would have resulted in all open wood fires going out; yet harnessing fire was essential to the development of civilization, technology, and science – e.g., the forging of metals. Why would an increase in α have this result? The reason is that in atomic units, everyday chemistry and the size of everyday atoms are not affected by up to a nine-fold increase or any decrease in α. Hence, the combustion rate of wood would remain the same with such a change. In these units, however, the rate of radiant output of a fire is proportional to α2 -- for example, a two-fold increase in α would cause the radiant output of an open fire to be four times as great. A small increase in α – around 10% to 40% -- causes the radiant energy loss of an open wood fire to become so great that the energy released by combustion cannot keep up, and hence the temperature of the fire would decrease to below the combustion point. The above argument applies to all forms of biomass, not just wood: since in atomic units, chemistry does not change with the changes in α considered above, their combustion rate would also remain the same. Although some biomass is much more combustible than wood – such as oil – these types of biomass are either not be readily available to primate carbon-based observers or would be less suitable for the size of fires needed for smelting metals; and hence it would be far less likely that primitive carbon-based observers would have regularly used them and thus discovered the smelting of metals. Going in the other direction, if α were decreased, light microscopes would have proportionality less resolving power without the size of living cells or other microscopic objects changing (when measured in atomic units). As is the maximum resolving power of light microscopes is about 0.2 microns, which happens to be the size of the smallest living cell. The only alternative to light microscopes for seeing the microscopic world is electron microscopes. Besides being very expensive and requiring careful preparation of the specimen, electron microscopes cannot be used to see living things. Thus, it is quite amazing that the resolving power of light microscopes goes down to that of the smallest cell (0.2 microns), but no further. If it had less resolving power, some cells could not be observed alive. The fine-structure constant, therefore, is just small enough to allow for open wood fires and just large enough for the light microscope to be able to see all living cells. Other constants also must be just right for other major domains to be discovered. For example, the ability to use radioactive dating – which plays a crucial role in geology, archeology, and paleontology -- depends on the density of radioactive elements in the crust of the planet on which observers evolve. As the strength of gravity is decreased (e.g., as measured by the force between two protons a unit distance apart), the density of radioactive elements must decrease to keep the number of volcanoes per unit area from increasing, which would decrease livability. Theses of Discoverability If the cases of discoverability are indeed coincidental under naturalism, the thesis they directly support is what I will call the discoverability thesis: Discoverability Thesis: This thesis is that the universe is non-accidentally structured in such a way as to be highly discoverable.iii If we define naturalism to include the thesis that any apparent teleology in the universe is accidental, then the discoverability thesis is in conflict with naturalism. In arguing for the discoverability thesis, I ultimately argue that the level of discoverability in our universe is much more coincidental than one would expect under naturalism. Specifically, I argue that among the alternative universes generated by varying the parameters of physics, a very small proportion are as discoverable as ours when the parameter itself is used as a natural measure of proportion. I call this the discoverability coincidence thesis. Finally, the data suggest a particularly strong version of the discoverability thesis, what I call the discoverability/livability optimality thesis (DLO): DLO: Within the range of values of a given parameter p that yield near-optimal livability, p will fall into that subrange of values that maximize discoverability (given constraints of elegance are not violated). In every case that I was able to make calculations regarding whether the fundamental parameters of physics are optimized in this way, they appear to pass the test.iv This alone is significant since this hypothesis is falsifiable in the sense that one could find data that potentially disconfirms it – namely, cases in which as best as we can determining, such as a case in which changing the value of a fundamental parameter – such as the fine-structure constant – increases discoverability while not negatively affecting livability.v Below, I will look at a case from cosmology where this thesis could have been disconfirmed but was not. New ideas are very subject to misinterpretation. So, before moving on, it is important to clear up a misunderstanding of what I am claiming, namely that I am arguing that this is the most discoverable possible universe. We can certainly imagine what initially might seem to be more discoverable universes and much less discoverable universes. We cannot draw any conclusions from this, however, unless we know the underlying laws of those universes – for example, an imagined more discoverable universe might require laws that are far more complex or inelegant, thereby taking away from their seeming discoverability. Rather, if we are going to test the thesis that the level of discoverability is coincidental, we must restrict ourselves to alternative universes chosen by a method that (1) is not a prior biased for or against the discoverability coincidence thesis and (2) such that we can make reasonable determinations of the level of discoverability of each universe. That is why the possible universes being considered are those with different values of the fundamental parameters, and why I restricted the DLO to those universes with different values for a given parameter – such as the finestructure constant -- but the same underlying laws. Further, as noted below, it is unsurprising under theism that the world has an elegant mathematical structure; in fact, physicists have often used elegance as a guide to new theories. This is why elegance was included in the discoverability/livability optimality thesis. So far, I have talked about the discoverability thesis. What is theism’s relation to this? God and Discoverability God is often defined as an omnipotent, omniscient, perfectly good being that created all contingent reality. Because God’s goodness is the only attribute that tells us anything about what God would do, theism leads us to expect that God would create a reality structured to realize a positive, and if possible an optimal, balance of good over evil. vi . Thus, theism renders unsurprising any feature of the world for which we can glimpse how it could be of moral or aesthetic value, such as the existence of embodied conscious agents or the elegance and harmony of the underlying laws of nature. If we can glimpse how discoverability might be of value, then theism would also render it unsurprising that the universe is highly discoverable. Why might discoverability be of value? First, it allows us to develop technology, which in turn allows us to greatly expand our ability to improve our conditions. Second, being able to understand the universe is widely perceived (at least on an implicit level) as being intrinsically valuable. If it were not, it is difficult to explain why many have sacrificed financial and other sorts of rewards to pursue fields such as cosmology, advanced physics, and the like. If one merely enjoyed solving puzzles, it would make much more sense to pursue a career that involved puzzle solving but in which the likelihood of employment was far higher. Further, the fact that governments spend billions of dollars on research into the fundamental structure of the cosmos, and that the public generally supports this, shows that collectively we find such knowledge of value. So, although theism does not require that the universe being highly discoverable, it renders it unsurprising and hence fits with it much better than naturalism. II. Significance Now, we are ready to see the significance of the above discoverability theses for the debate over God and cosmology. I begin with the multiverse hypothesis. Multiverse Hypothesis A significant number of philosophers and scientists respond to the anthropic fine-tuning evidence by claiming that it is brute fact that does not need any explanation. I find such a response incredible when one looks at the degree of fine-tuning: one part in 10120 – that is, one followed by 120 zeroes -- in the case of the effective dark energy density and a ridiculous 1 part in 10 to the 10123 ( ) for the probability of the universe having an entropy as low as ours. Speaking for myself, almost anything is more believable than that.vii Thus, in order to believe that, I would have to be given an exceedingly strong argument that there was no other alternative, not the kind of arguments typically offered against the theistic alternative. At the very least, such evidence puts a burden on naturalists to provide powerful reasons for rejecting the theistic explanation. A more common response among cosmologists is to invoke a multiverse to explain the finetuning. According to this hypothesis, there is a very large, if not infinite, number of regions of spacetime with different values of the fundamental parameters of physics, different initial conditions, and perhaps even different laws of nature. It then claims that in a sufficiently varied multiverse, it is no surprise that some universe is structured so that observers will arise in it. Finally, it invokes the so-called observer-selection principle, which is the tautological claim that embodied observers can only exist in a region of space- time that allows for them to exist. This renders it unsurprising that as observers we find ourselves in an observer-structured region of space-time since it is impossible for us to exist in any other type of region. The observer-selection principle is essential to the multiverse explanation because it prevents it from undercutting the need to explain other seemingly surprising events and features of the universe. For example, normally one would think that it is too coincidental for a six-sided die to land 50 times in a row on four just by chance. Yet, in a large enough multiverse, someone will observe this to happen. Nonetheless, it is still improbable that a generic observer in a generic multiverse will see such an occurrence. Hence, purportedly, via the observer selection principle the multiverse hypothesis combined with the observer-selection principle can render it unsurprising both that we exist and that we find ourselves in an observer-structured universe while at the same time not undercutting ordinary claims of improbability. Because of its reliance on the observer selection principle, the multiverse can only directly render unsurprising that we find ourselves in an observer-structured universe. Because of this limitation, it cannot, without additional hypotheses, explain the fine-tuning of the constants or the finetuning for discoverability. With regard to the former, all those universes that do not have life-permitting values for the fundamental parameters of physics – such as the cosmological constant -- and are sufficiently large (e.g., infinite) will be dominated by isolated observers arising from thermal fluctuations. Thus, the parameters of physics are not fine-tuned for observers. Rather, they are finetuned so that embodied conscious agents can arise that can significantly interact with one another. But, because of its reliance on the observer-selection effect, without additional postulates the multiverse only implies that we will find ourselves in an observer-structured universe. Thus, it cannot of itself directly explain the actual anthropic fine-tuning – that is, why as generic observers, we find ourselves in a universe whose fundamental parameters allow for embodied conscious agents. The existence of these Boltzmann brain universes, therefore, poses a problem not just for a multiverse explanation of the low entropy of the universe, but more generally for explaining the fine-tuning of the constants. According to the thesis I am proposing, the universe is not just fine-tuned so that ECAs can exist, but that they can develop technology and discover its nature. The multiverse hypothesis also runs into a major problem explaining this. The reason is that there seems to be no necessary connection between a universe being ECA-permitting and its being discoverable beyond that required for getting around in the everyday world. Thus if, because of the fine-tuning for discoverability, the proportion of observerpermitting universes that are as discoverable as ours is really small, it would be very improbable under a generic multiverse hypothesis that as generic observers we would find ourselves in such a universe. So, the fine-tuning for discoverability, if legitimate, presents a further problem for the multiverse as a complete explanation of the fine-tuning; since this discoverability is not surprising under theism, it provides further data to test these two hypotheses and thus move the debate forward. Irrelevant to Life Objection I will now consider how the discoverability theses help to answer a common objection, usually raised by physicists, against anthropic fine-tuning: namely, many features of the universe seem irrelevant for the existence of life. I will call this the irrelevant to life objection. This objection is nicely stated by Sean Carroll. After listing some reasons to be skeptical of fine-tuning claims, Carroll states that “But in fact there is a better reason to be skeptical of the fine-tuning claim: the indisputable fact that there are many features of the laws of nature which don't seem delicately adjusted at all, but seem completely irrelevant to the existence of life.” viii One commonly used example – for example by Carroll, Steven Weinberg, and Mario Livio -- is the existence of extra generations of particles, such as the muon, a particle that is in all ways like the electron except being much more massive; such particles do not seem in any way needed for life.ix Along similar lines, Carroll takes issue with a theistic explanation of the low entropy of the universe, correctly noting that only a universe with a local region of low entropy is needed for life, Carroll rightly notes that in creating a universe with low entropy throughout, God would have had to fine-tune the universe far more than would have been necessary for life. This, he states, poses “a bigger problem for the God hypothesis than for the multiverse.” x Contrary to what Carroll assumes, it is not clear why under theism every feature of the universe would have to serve a purpose. But even if theism did, discoverability (and perhaps considerations of elegance) could offer a way of explaining them. The extra generations of particles could very well help with humans discovering the fundamental structure of matter. For instance, these extra generations fall into a highly symmetric pattern, three for each of the two types of quark and three for the two types of leptons. This symmetry suggests that they are clues to an even deeper, more elegant theory. Further, they could help in discoverability in other ways. For example, according to the June 2012 issue of Symmetry Magazine (a joint publication of Fermilab and Stanford Linear Accelerator Center), the muon – the particle perhaps most commonly cited as an added extra whose existence seems purposeless -- is playing an increasingly important role in particle physics. According to Chris Polly, one of the Fermilab physicists involved in muon research, one reason the muons are special is that “they are light enough to be produced copiously, yet heavy enough that we can use them experimentally to uniquely probe the accuracy of the Standard Model.” Further, the article notes that “Today scientists can manipulate the muon and use it as a tool not only for particle physics research but also for cosmology, archeology and public safety.xi With regard to the low entropy of the universe, having a low entropy throughout the entire universe makes it more discoverable for at least two reasons. First, a universe that has a low entropy over a vast region is necessary for us to observe other stars and galaxies, and thus to understand the big bang origin of our own universe. (The existence of stars and galaxies requires low entropy.) As Carroll notes, if the region of law entropy were not large enough, the universe would be devoid of other galaxies.xii Second, to solve the equations of general relativity in closed form for the entire cosmos – which is central to doing cosmology – one must assume that the distribution of matter is nearly uniform at large scales, a cosmological scenario known as Friedmann cosmology. This would not be true if the universe was not in a low entropy state throughout. Of course, one could respond that God could have created a universe with one galaxy, which would have been even more discoverable. The problem is that we have no clue what the underlying laws of such a universe would have to be, and thus whether they would be as discoverable as ours. As I noted above, to make any kind of judgments about overall discoverability, we must restrict ourselves to alternative universes in which the fundamental parameters of physics are changed – in this case, the spatial distribution of entropy; because of this restriction, the above example universe is irrelevant to my argument. In light of the possibilities that discoverability offer for understanding why the universe is this way, let me suggest that objections like the above run the risk of being an ungodly appeal to gaps: just because we do not understand the reason God would have created the world with a particular fundamental feature does not mean there is not one. Besides its ability to make sense of the items mentioned above, below I consider an examples of the discoverability/livability optimality thesis’s (DLO) potential predictive power. III. Predictive and Explanatory Power of Discoverability Cosmic Microwave Background Radiation The most dramatic confirmation of the DLO is the dependence of the Cosmic Microwave Background Radiation (CMB) on the baryon to photon ratio. The CMB is leftover radiation that permeates space from the big bang that has been redshifted into the microwave region of the electromagnetic spectrum. Because its source is in the big bang, the CMB tells us critical information about the large scale structure of the universe: for example, physicists John Barrow and Frank Tipler point out that “The background radiation has turned out to be the ‘Rosetta stone’ on which is inscribed the record of the Universe’s past history in space and time.”xiii Much of the information in the CMB is in very slight variations in its intensities of less than one part in 100,000 in different parts of the sky. Since the CMB is already fairly weak, this implies that within limits, the more intense it is, the smaller the fluctuations it can detect, and hence the better a tool it is for discovering the universe. Now, the intensity of CMB depends on the photon to baryon ratio, ηγb , which is the ratio of the average number of photons per unit volume of space to the average number of baryons (protons plus neutrons) per unit volume. At present this ratio is approximately a billion to one (109), but it could be anywhere from one to infinity; it traces back to the degree of asymmetry in matter and anti-matter right after the beginning of the universe – for approximately every billion particles of antimatter, there was a billion and one particles of matter. So far it is a mystery why this ratio is what it is. Even if physicists can give a further explanation for why this ratio has the value it does, the question would still arise as to why that deeper physics instead of some other. The only livability effect this ratio has is on whether or not galaxies can form that have nearoptimally livability zones. As long as this condition is met, the value of this ratio has no further effects on livability. Hence, the DLO predicts that within this range, the value of this ratio will be such as to maximize the intensity of the CMB as observed by typical observers. According to my calculations – which have been verified by three other physicists -- to within the margin of error of the experimentally determined parameters (~20%), the value of the photon to baryon ratio is such that it maximizes the CMB. This is shown in Figure 1 below. CMB/CMB0 1.5 1 0.5 0 0.1 1 10 100 1000 ηγb/ηγb0 Figure 1: Variation of the intensity of the cosmic microwave background radiation (CMB) observed by a typical observers for various values of the photon to baryon ratio (ηγb). CMB0 and ηγb0 are the values of the CMB and ηγb in our universe. Notice that the CMB is maximal at ηγb/ηγb0 = 1 -- that is, for the value of the photon/baryon ratio in our universe. It is easy to see that this prediction could have been disconfirmed. In fact, when I first made the calculations in the fall of 2011, I made a mistake and thought I had refuted this thesis since those calculations showed the intensity of the CMB maximizes at a value different than the photon-baryon ratio in our universe. So, not only does the DLO lead us to expect this ratio, but it provides an ultimate explanation for why it has this value, whatever other explanation we find based on some deeper physics. This is a case of a teleological thesis serving both a predictive and an ultimate explanatory role. IV. Objections Now I turn to a couple of objections. Discoverability-Selection Objection One objection to the above argument is that there is a discoverability selection effect. One form this objection takes is the claim that if a domain were not discoverable, we would not know what we are missing. This objection is similar to the “weak anthropic principle” response to the anthropic finetuning: if the fundamental parameters of physics were not observer-permitting, there would be no one here to observe the fact. The first, and simplest, answer to this objection is that this is not generally the case, as the following examples will demonstrate. As our first example, suppose that α were 50 times smaller. In that case, light microscopes could only resolve objects down to 10 microns. Yet, if observers could exist in such a universe, they could observe some cells and thus develop cell theory, and gain indirect knowledge that there were cells less than 10μm in size, and yet wish that they had an instrument that could see them. Or, consider radioactivity dating. Even without high enough levels of radioactive elements in the earth’s crust to be of much use in geology, there could be enough radioactive elements for us to learn about radioactivity; we would then be able to determine that if only the earth had a higher density of radioactive elements, we could have a useable dating method. In fact, it should be noted here, the level of background radiation is as about as high as it could be without posing a significant threat to life, thus meeting the expectations of the DLO (since the higher the level, the more useful radioactive dating would be.) Finally, we could have lived in a universe in which the photon to baryon ratio was substantially different, but where a very weak CMB was still detectable. Then one could show that the photon to baryon ratio in our universe would maximize the intensity of the CMB and hence would make it more useful than any other value of this ratio. xiv Cherry Picking Objection Another possible response to cases of discoverability is that they involve “cherry picking”: one looks for and finds the information that confirms one’s hypothesis, ignoring all the disconfirming evidence. There are several types of “cherry picking.” First, one might pick features of the world that are helpful for discovery and ignore those features of the world that hurt discovery; since there are so many domains that could be discovered, it seems likely that one could always find some features of the world that are helpful for discovery with almost any kind of world that could give rise to observers. This objection can be avoided by restricting ourselves to only considering widely recognized major domains – cosmology, geology, cell biology, the fundamental microscopic structure of the world, and the like – and the major possible tools for those domains, such as the light microscope, radioactive dating, and the CMB. Further, one might worry that since the tools used for discovery are often not obvious until they are developed, if the values of the parameters were different, there might be other possible tools that observers could develop that we are not presently aware of. Thus, although the usefulness of our present tools of discovery depends on the parameters falling into a restricted range of values, this does not mean that an observer’s ability to discover the domains in question depends on a small range. For the major domains of discovery, this is typically not the case. For example, alternative values for α will not give rise to alternative forms of radiation that would be as good as light for observing cells; hence, for at least a very wide range of values of α there will not be any adequate replacement for the light microscope. Or consider the CMB. Different values for the photon to baryon ratio will not give rise to an alternative form of radiation that is a substitute for the CMB. Similarly, a lower value of the strength of gravity is not going to give rise to a “clock” that is as good as radioactive dating. V. Conclusion In conclusion, looking for cases of fine-tuning for discoverability has the potential of providing a new set of empirically-based evidence with regard to the debate whether the universe is teleologically structured or indifferent to our existence. It thus has the promise of substantially moving the debate forward. Appendix A: Dark Energy and Cosmic Coincidence Problem Another explanatory potential of the DLO involves the so-called “cosmic coincidence problem.” This problem has to do with the effective dark energy density being about the same as the energy density of matter at the present age of the universe. This is thought a coincidence because the matter density falls off very fast with the universe’s expansion (specifically, with the inverse cube of the amount of expansion) whereas the effective dark energy density is nearly constant. As Sean Carroll notes, if these two densities are comparable today, “in the past the vacuum energy would be undetectably small, while in the future the matter density will be negligible. This ‘coincidence problem’ has thus far proven a complete mystery.” (2004, P. 359). Initially, the existence of this effective dark energy might seem to disconfirm the DLO since it has caused the universe to expand about 20% more than it otherwise would, thereby weakening the CMB by a factor of about two. Such would be the case if there were not at least two positive discoverability effects from a small dark energy density. First, the positive dark energy density has made the Hubble parameter about 50% larger than it would otherwise be. The Hubble parameter determines the rate at which galaxies appear to be moving away relative to distance; the larger the Hubble parameter, the faster they are receding for the same distance away. If one looks at Hubble’s original data, it is clear the Hubble parameter was just large enough to keep Hubble’s original data from being swamped by the local (peculiar) motion of galaxies within galactic clusters. So, the dark energy helped with both our discovering the expansion of the universe, and then later with establishing a reliable relation between the distance of a galaxy and its redshift, something very important for galactic astrophysics. Second, arguably if the effective dark energy density were much smaller, it would not yet have been – and perhaps never would be – detected. Detection of the dark energy has mostly relied on the universe’s having undergone accelerated expansion starting approximately six billion years ago; as is, it was only until the late 1990s that we had precise enough instruments to detect this accelerated expansion. It is easy to show that if the dark energy density were less than about ¼ its value in our universe, the universe would yet to begin this accelerated expansion phase. Yet knowledge of this dark energy density tells us many important things about the universe (over and above merely the knowledge that the universe contains dark energy.) For example, it lends credence to one of the most speculative, yet crucial, postulate of inflationary cosmology – the possibility of a large effective dark energy density at the beginning of the universe. Further, it tells us that there is no fundamental symmetry that requires that the effective dark energy density be zero, a symmetry that would have solved the anthropic finetuning problem of the cosmological constant mentioned at the beginning of this paper. Thus, the detection of the dark energy has kept cosmologists from mistakenly pursuing such a solution and has rightly kept the problem on the table. Finally, the attempt to understand the origin of this dark energy has given a significant boost to cosmology. Given these positive discoverability contributions are significant, the DLO renders it unsurprising that the effective dark energy density would be large enough to guarantee its positive discoverability contribution, but no larger because of the negative discoverability contribution to the CMB. 1 This is precisely what we find. Thus, it has not only passed another potential disconfirmation, but has shown once again its potential explanatory power. Finally, it should be emphasized that the explanations given above are not in competition with a deeper scientific explanations. Even if we found a deeper set of laws or parameters that explained the values for photon to baryon ratio and the dark energy density, the DLO would provide an ultimate explanation of why the universe has these set of laws and parameters instead of some other that yields different values for the photon to baryon ratio and the effective dark energy density. Appendix B: Brief Physical Arguments for the Cases Above 1. Light Microscope The basic reason for the dependence of the resolving power of a light microscope on α goes as follows. First, the maximum resolving power of a light microscope is approximately ½ the wavelength of the light being used. Second, in atomic units, decreasing α does not significantly affect the size and 1 What about a negative effective dark energy density? In that case, the universe would switch from expanding to contracting when the dark energy density was equal to the matter density. This would leave two possibilities: either observers occur before the switchover in an expanding phase or after the switchover in a contracting phase. If they occur in an expanding phase, then observers would observe a greater deceleration of the universe than in a pure matter universe. Even if the effective dark energy density was negative enough to be detectable, such a greater deceleration would likely not be as convincing evidence for dark energy as an accelerating universe, partly because it is not qualitatively different from what the universe is already doing (decelerating), but also because positive curvature can cause a greater deceleration. Even if such a greater deceleration could be as direct or convincing evidence for dark energy as the acceleration that has occurred in the last six billion years, the Hubble parameter would have to be considerably smaller than in a pure matter universe, with the amount by which it is smaller being greater for nearby galaxies. This would create worse conditions for measuring the Hubble parameter, which would have at least impeded obtaining evidence for the big bang along with many other aspects of cosmology that rely on the Hubble parameter. If observers occur in a contracting phase, then the Hubble parameter could be as large as in our universe, but light from nearby galaxies would be blue-shifted whereas light from galaxies farther away would be redshifted. This would likely make it much more difficult to establish the big bang theory, and much more difficult to establish the relationship between distance and degree of redshift or blue-shift. chemical properties of atoms involved in biochemistry, since written in these units the non-relativistic Schrödinger equation is not dependent on α; thus, the size of living cells would not change. Finally, in atomic units the minimum wavelength of light eyes can see is inversely proportional to α. Thus, for instance, if α were 1/100 its current value, the maximum resolving power would be 20 microns, not 0.2 microns. To see why the minimum wavelength that observers can see is inversely proportional to α first note that I atomic units, the speed of light = c = 1/α. However, the energy of a photon = E = hf, where h is Planck’s constant and f is the frequency of light. (h = 1 in atomic units). But, a photon of visible light cannot have more energy than the bonding energy of typical biochemical molecules, otherwise it would destroy the molecules in an organism’s eye. In our world, this requires that for light microscopes, f < 800 trillion cycles per second. Since this bonding energy remains the same with a change in α, f < 800 trillion cycles per second in the alternative world. Now, the wavelength of light = λ = c/f. In our world, the above restriction on frequency means that λ > 0.35 microns. Since c = 1/α, as α decreases, c increases, which causes the minimum wavelength of light for a light microscope to increase; this in turn means that the resolving power of light microscopes will decrease. 2. Radioactive Dating Radioactive dating is dependent on the strength of gravity being sufficiently large. The strength of gravity is commonly given by the dimensionless gravitational fine-structure constant αG = G(mp)2 /ℏc, where mp is the mass of the proton. If αG is decreased, to retain an atmosphere the radius, Rh, of a habitable planet must increase: specifically, Rh ∝ (1/αG)1/2. Now, the ratio of volume to surface area of a planet is proportional to its radius. Since the amount of heat produced in a planet via radioactive decay is proportional to the volume of the planet, unless the density of radioactive elements decreases with a decrease in αG, the amount of heat energy per unit area exiting the planet’s surface – and hence the amount of volcanic activity per unit area – will increase as αG decreases, at some point drastically decreasing the ability of the planet to support observers that can engage in advanced science. Thus, to retain a world livable enough for science, the density of radioactive elements in the crust must at some point decrease, making the world worse for radioactive dating. (It is assumed here that ratio between the density of radioactive elements in the crust and in the interior of the planet does not substantially increase as the planet gets larger.) i More accurately, it is the fine-tuning of the effective dark energy density of the universe, which is the sum of Einstein’s original cosmological constant with various energy fields that of themselves would cause an accelerated expansion or contraction of the universe. However, the cosmological literature has typically talked about it in terms of the fine-tuning of the cosmological constant. ii In the Privileged planet (2003), Gonzalez and Richards cite a significant number of examples where what they call habitability (the condition under which a planet can exist and be habitable) appear to coincide with the conditions under which the universe is optimally discoverable. In their treatment, however, they provide no methodology for determining whether the universe is more discoverable than would be expected by chance. iii In this definition, “non-accidental” is to be taken in the minimalist sense that is used in philosophical discussions of the metaphysical status of the laws of nature. In these discussions, an accidental regularity is one such that observed instances of the regularity do not give us any reason to believe that the regularity will continue in unobserved cases. For example, if I ask five people in my class whether they have a dime in their pockets, and they all answer yes – but having such a dime was not prearranged but just occurred by chance – this would give me no reason to think the next person I ask would have a dime in their pockets. Or, if I flip a coin ten times in a row, and each time it comes up heads, and I check the coin and it is a two-sided, perfectly symmetrical coin (except one side has heads engraved on it and one side has tails engraved on it), I assume this just occurred by accident. Thus, the fact that it came up heads the first ten times gives me no reason to believe it will continue. In contrast, suppose that I freeze a certain substance – call it substance X – ten times and every time it freezes, it contracts. That would th give me good reason to believe that the 11 time it freezes it will contract. Why? Because I take the regularity to be non-accidental – the result of underlying laws. In philosophical discussions of the metaphysical nature of the laws of nature, everyone agrees that a distinguishing mark of accidental versus non-accidental regularities is the ability of observed instances of a nonaccidental regularity to give us good reason to believe that unobserved cases (in relevantly the same circumstances) will also fall under the regularity. In the language of Nelson Goodman, non-accidental regularities could be said to be projectable. The theories divide on what they think accounts for this projectability, specifically when it comes to the most fundamental regularities in the world. Two prominent views are those of the so-called necessitarians and those of the regularists. Necessitarians think non-accidental regularities – the regularities expressed by the laws of nature -- are ultimately the result of underlying necessities in nature. For example, assuming that regularity that masses always attract each other cannot be explained by any deeper regularity, a necessitarian would say that they must attract each other due to some underlying property of the world that is beyond science to discover. On the other hand, advocates of the regularity theory claim that the non-accidental regularities expressed by are must fundamental theories are just brute facts, without any deeper explanation. Since advocates of the regularity theory hold that all regularities are merely brute facts, they then must give an account of what distinguishes accidental regularities from non-accidental (lawlike) regularities. Typically, such accounts appeal to the role such regularities play in our models of the world. The point of the above summary is not to resolve the dispute between the various accounts of the laws of nature but to simply clarify what I mean by non-accidental. Thus, the discoverability hypothesis entails that the fact that the world has been found to be highly discoverable can give us strong reason to believe the pattern will continue. iv I only considered parameters in models one level above the standard model of particle physics. However, the fact that they passed the test in this case strongly suggests that they would at a deeper level if we could do the calculations. v The way this hypothesis would be disconfirmed is similar to the way that the hypothesis that animals were optimally constructed was disconfirmed by anatomical studies in the nineteenth and twentieth centuries, namely by showing how an organ (such as the back) could be constructed better to fulfill its function. vi The idea that non-abstract reality is non-accidentally structured so that moral (and aesthetic) value is positively, or optimally, realized is what has become known as the axiarchic thesis, a thesis that dates back to Plato and has been taken up recently by other philosophers, such as John Leslie (1979). Theism entails axiarchism, but axiarchism does not entail theism; so one can be an axiarchist without being a theist, but not vice versa; further, it is only via its entailing the axiarchic thesis that theism leads to any expectations about the structure of the universe, and hence it is only via its entailment of this thesis that it can explain the fine-tuning. For our purposes, the significance of axiarchism is that with some auxiliary assumptions about what kinds of things are likely to be of moral and aesthetic value, it entails certain expectations about the structure of the universe without making any metaphysical commitments about realities beyond the universe. Specifically, if can glimpse how discoverability could be of value, then the axiarchic thesis would render it unsurprising that the universe is highly discoverable. vii It is not its mere seeming improbability that makes it cry out for explanation, since there are many occurrences in the world that are enormously improbable. Rather, it is a combination of being seemingly enormously improbable with its being special – e.g., required for the existence of embodied conscious agents. viii “Why Most Cosmologists Are Atheists,” page __. In file). ix Dreams of a Final Theory, pp. 253-254. x Sean Carroll, “Does the Universe Need God,” in The Blackwell Companion to Natural Theology, J. B. Stump and Alan Padgett, editors, Wiley-Blackwell, Malden, MA, 2012, p. 192 (185-197). xi Sarah Charley, “Through a muon’s eyes,” Symmetry: A joint Fermilab/Slac Publication, June 12, 2012, available at http://www.symmetrymagazine.org/cms/?pid=1000961. Accessed July 20, 2012. xii op. cit., pp. 192 – 193. xiii John Barrow and Frank Tipler, The Cosmological Anthropic Principle, 1986, p. 380. xiv Our second response notes that even if we would not have known the difference in some cases, that does not affect the claim that it is coincidental that we have the tools to discover the domain in question. To see, consider the following example. Suppose a government a nation of a billion people has what they call a grace-lottery, which only the highest level government officials and the past winners know about. Without even buying a ticket, a million dollars is given to one person a year, supposedly chosen at random among the entire population; part of the requirements for receiving the money is that one is to tell no one where the money came from, but offer an official “lie” about its source. Now, suppose some poor farmer – call him Omaz -- receives a million dollars from this program. He not only finds out about the program, but when he uses several thousands of dollars of the money to check into his family history, he discovers that he has a half-brother who is now a high official in the government and the primary person in control of the lottery. Omaz now becomes suspicious that the lottery was rigged, reasoning as follows: Corruption is rampant in this government, and we are taught from birth that one’s happiness in the afterlife depends on taking care of one’s closest blood relatives. Consequently, it is not unlikely that my half-brother was responsible for me receiving the million dollars through this program. It is very, very unlikely – around one in a twenty million over a fifty year life span – that I would have received this money by chance. Hence, my receiving the money strongly confirms that my brother had a role in this over the chance hypothesis. Certainly Omaz’s reasoning seems correct. If Omaz had not been selected, however, he would not have known about the lottery, and thus that he was not selected. Yet, this does not take away from the confirmation that receiving the money gives to his rigging hypothesis.