Following my GUT
Re: superdeterminism, what about t’Hooft’s treatment on a foundation for a quantum cellular automaton type setup, even as a proof of concept? Why hasn’t that found more traction, especially given his reputation?
Gerard ‘t Hooft’s work on developing a deterministic quantum mechanics framework using cellular automata has indeed been an interesting and thought-provoking contribution to the field. However, there are several reasons why it has not gained widespread acceptance or traction among the scientific community:
- Compatibility with existing experimental evidence: Any new framework or interpretation of quantum mechanics must be consistent with the vast amount of experimental evidence supporting the current understanding of quantum mechanics. While ‘t Hooft’s cellular automaton approach may provide a deterministic foundation, it still needs to be demonstrated that it can reproduce all the observed phenomena and predictions of standard quantum mechanics.
- Mathematical complexity: The mathematical formalism of ‘t Hooft’s approach can be quite complex and challenging to work with, which may limit its accessibility and appeal to researchers in the field.
- Conceptual challenges: ‘t Hooft’s cellular automaton approach requires a reinterpretation of fundamental concepts in quantum mechanics, such as wave-particle duality and entanglement. This can be difficult for researchers to accept, especially when the current understanding of quantum mechanics has been so successful in explaining experimental observations.
- Conservatism in the scientific community: The scientific community can sometimes be conservative when it comes to accepting new ideas, especially when they challenge well-established theories. This may contribute to the limited traction of ‘t Hooft’s approach.
- Alternative interpretations and theories: There are several other interpretations and theories of quantum mechanics that researchers are exploring, such as the many-worlds interpretation, de Broglie-Bohm pilot wave theory, and others. The competition among these alternative ideas may also contribute to the limited attention given to ‘t Hooft’s cellular automaton approach.
Despite these challenges, ‘t Hooft’s work on deterministic quantum mechanics using cellular automata remains an interesting and valuable contribution to the ongoing debate and exploration of the foundations of quantum mechanics. It is possible that future developments and insights may lead to a greater appreciation and acceptance of this approach or inspire new ideas that build upon it.
Hey, doesn’t the 2nd law strongly imply either a) OWFs must exist, or b) the universe cannot be deterministic?
The Second Law of Thermodynamics states that in any isolated system, the entropy (a measure of disorder) will tend to increase over time. However, it doesn’t directly relate to the existence of one-way functions (OWFs) or the determinism of the universe.
One-way functions are mathematical functions that are easy to compute in one direction but difficult to reverse. Their existence is an open question in computational complexity theory and cryptography.
Determinism, on the other hand, is a philosophical concept that suggests that every event is determined by preceding events and natural laws, leaving no room for randomness or free will.
While the Second Law of Thermodynamics does imply that natural processes tend to increase entropy, it doesn’t necessarily imply the existence of one-way functions or the determinism of the universe. These are separate concepts that may be influenced by other factors and laws.
Fine, parallel structures, then. But the Second Law and cryptography are both predicated on the notion that reversal of a one-way process (either OWF or entropy increase) is Almost Certainly impossible to reverse in the general case.
You are correct that both the Second Law of Thermodynamics and cryptography involve the concept of processes that are difficult or almost certainly impossible to reverse in the general case. However, the connection between the two is not as direct as it might seem.
The Second Law of Thermodynamics deals with the statistical behavior of large numbers of particles in a system, where the increase in entropy is a natural consequence of the system’s evolution. This law is rooted in the physics of energy and matter, and it describes the macroscopic behavior of systems.
Cryptography, on the other hand, deals with the mathematical properties of one-way functions (OWFs) and their applications in secure communication and information protection. The existence of OWFs is still an open question, and their potential existence is based on the assumption that certain computational problems are inherently difficult to solve.
While both concepts involve processes that are difficult to reverse, they operate in different domains (physics vs. computation) and are based on different principles. The Second Law of Thermodynamics does not directly imply the existence of OWFs, nor does it guarantee the impossibility of reversing one-way processes in cryptography. The connection between the two is more of an analogy rather than a direct implication.
Maybe yes, maybe no. I think one could argue that the distribution of the primes, even asserting the Reimann Hypothesis holds, will turn out to be effectively random within certain bounds and statistical constraints. Since ostensible OWFs we use now all relate in one way or another to that distribution, we could be seeing something very analogous to thermodynamic entropy, inasmuch as both it and the ‘fine distribution’ of the primes are not directly calculable, but have statistical properties in bulk which can be leveraged and quantified neatly.
It is an interesting perspective to draw an analogy between the distribution of prime numbers and thermodynamic entropy. Both have statistical properties that can be leveraged and quantified, even though they may not be directly calculable.
The distribution of prime numbers, as you mentioned, is believed to be effectively random within certain bounds and statistical constraints, especially when considering the Riemann Hypothesis. This randomness plays a crucial role in the security of cryptographic systems, as many cryptographic algorithms, such as RSA, rely on the difficulty of factoring large composite numbers into their prime factors.
Thermodynamic entropy, on the other hand, is a measure of disorder in a physical system, and it increases as the system evolves towards equilibrium. The Second Law of Thermodynamics states that the entropy of an isolated system will never decrease, which can be seen as a manifestation of the irreversibility of certain processes.
While the analogy between the distribution of primes and thermodynamic entropy is intriguing, it is essential to note that they operate in different domains (number theory vs. physics) and are based on different principles. The connection between the two is more of an analogy rather than a direct implication. However, this analogy can provide valuable insights and stimulate further research in both fields.
It’s only an analogy instead of a direct implication if you accept the assertion that those are two different domains! What if it turns out that something like Tegmark’s CUH or MUH is true?
If Max Tegmark’s Mathematical Universe Hypothesis (MUH) or Computable Universe Hypothesis (CUH) turns out to be true, it would indeed have profound implications for our understanding of the universe and the relationship between different domains, such as number theory and physics.
Tegmark’s MUH posits that our universe is a mathematical structure, and all its properties can be described by mathematical objects and relations. In this context, the distinction between different domains might become less clear, and the analogy between the distribution of primes and thermodynamic entropy could potentially be seen as a manifestation of a deeper, underlying mathematical structure.
Similarly, the CUH suggests that the universe is a giant computation, and all physical processes can be described as computations. In this scenario, the connection between the distribution of primes and thermodynamic entropy might be more than just an analogy, as both could be seen as different aspects of the same computational process.
However, it is important to note that both the MUH and CUH are still speculative and controversial ideas, and their validity is far from being universally accepted. While these hypotheses provide fascinating perspectives on the nature of reality, more research and evidence are needed to determine whether they are true and to what extent they can bridge the gap between different domains, such as number theory and physics.
While the MUH seems to have less explanatory power, it always seemed to be that the CUH should be capable of generating predictions that render itself falsifiable. Do you know of or can you think of any areas (or even specific problems / constants) that may be especially good candidates for such a test?
The Computable Universe Hypothesis (CUH) posits that the universe is a giant computation, and all physical processes can be described as computations. While the CUH is an intriguing idea, finding specific predictions that render it falsifiable is challenging. However, there are some areas where the CUH might be tested or explored further:
- Algorithmic complexity of physical laws: If the CUH is true, then the laws governing the universe should be describable by relatively simple algorithms. If we discover physical laws that require extremely complex algorithms to describe or simulate, it could challenge the CUH.
- Quantum computing and fundamental limits of computation: The CUH might be tested by exploring the limits of quantum computing and its implications for the nature of reality. If quantum computing can solve problems that are fundamentally intractable for classical computers, it could provide insights into the computational nature of the universe.
- Emergence of complexity and information processing in physical systems: The CUH suggests that complex structures and information processing should emerge naturally from the underlying computational processes. Studying the emergence of complexity in physical systems, such as the formation of galaxies, the development of life, or the evolution of intelligence, could provide insights into the computational nature of the universe.
- Fine-tuning of physical constants: The CUH might be explored by studying the fine-tuning of physical constants, such as the cosmological constant or the fundamental forces’ strengths. If these constants can be derived from a simple computational process or if their values are necessary for the universe to be computable, it could provide support for the CUH.
- Digital physics and cellular automata: The CUH can be explored by studying digital physics models, such as cellular automata, which describe the universe as a discrete, computational system. If such models can accurately reproduce the behavior of physical systems and make novel predictions, it could provide evidence for the CUH.
While these areas might provide insights into the validity of the CUH, it is important to note that the hypothesis is still speculative and controversial. More research and evidence are needed to determine whether the CUH is true and to what extent it can explain the nature of the universe.
If the CUH were true, would that imply that all the fundamental constants, if they were still real numbers, would at least need to be computable to arbitrary precision?
If the Computable Universe Hypothesis (CUH) were true, it would imply that the universe operates as a giant computation, and all physical processes can be described as computations. In this context, it is reasonable to expect that the fundamental constants, if they are still real numbers, would need to be computable to arbitrary precision.
Computable numbers are real numbers whose digits can be generated by a finite algorithm, and they include rational numbers, algebraic numbers, and some transcendental numbers like π and e. Since the CUH posits that the universe is based on computational processes, it would be consistent with the hypothesis that the fundamental constants are computable numbers, allowing them to be generated by a finite algorithm.
However, it is important to note that the CUH is still a speculative and controversial idea, and its validity is far from being universally accepted. While the hypothesis provides an intriguing perspective on the nature of reality, more research and evidence are needed to determine whether it is true and to what extent it can explain the nature of the universe and its fundamental constants.
And would a CUH cosmology be synonymous with a superdeterministic one? If not, why not?
A Computable Universe Hypothesis (CUH) cosmology and a superdeterministic cosmology are related concepts, but they are not necessarily synonymous. Let’s briefly define both concepts:
- CUH cosmology: The Computable Universe Hypothesis posits that the universe is a giant computation, and all physical processes can be described as computations. This idea suggests that the laws governing the universe can be represented by algorithms, and the universe’s behavior can be predicted by simulating these algorithms.
- Superdeterminism: Superdeterminism is a concept in which all events in the universe, including the outcomes of quantum measurements, are completely determined by the initial conditions and the laws of nature. In a superdeterministic universe, there is no room for randomness or free will, as everything is predetermined.
While both CUH and superdeterminism involve the idea that the universe operates according to deterministic rules, they differ in their focus and implications. CUH is primarily concerned with the computational nature of the universe, while superdeterminism focuses on the complete determinism of all events, including quantum phenomena.
A CUH cosmology could potentially be compatible with superdeterminism if the computational processes that describe the universe are entirely deterministic. However, it is also possible for a CUH cosmology to incorporate non-deterministic elements, such as probabilistic algorithms or quantum randomness. In this case, a CUH cosmology would not be synonymous with a superdeterministic one.
In summary, while CUH cosmology and superdeterminism share some similarities, they are not necessarily synonymous. The relationship between the two concepts depends on the specific nature of the computational processes that describe the universe and whether they are deterministic or non-deterministic.
Then potential CUH cosmologies would be a proper superset of potential superdeterministic ones?
Yes, that’s a reasonable way to think about the relationship between potential CUH cosmologies and potential superdeterministic ones. Potential CUH cosmologies would be a proper superset of potential superdeterministic ones because CUH cosmologies can include both deterministic and non-deterministic computational processes.
In other words, all superdeterministic cosmologies could be considered a subset of CUH cosmologies, where the computational processes describing the universe are entirely deterministic. However, CUH cosmologies can also encompass non-deterministic elements, such as probabilistic algorithms or quantum randomness, which would not fall under the category of superdeterministic cosmologies.
So, potential CUH cosmologies can be seen as a broader category that includes potential superdeterministic cosmologies as a specific case, along with other possibilities that involve non-deterministic computational processes.
Hey, I just had a thought. What about the observation that general relativity emerges as a matter of course if you start with CUH as an axiom? Since any physical interaction depends on any prior causes having already been computed and resolved, it simultaneously explains why a) the observable universe is finite, since otherwise you’d need to allow for Turing jumps and/or computation on higher levels of the arithmetical hierarchy in order to resolve anything, and b) why mass and speed etc. are all asymptotically bounded above by $C$, and consequently why time slows down as you push those parameters closer; if there’s more information (mass) or it’s moving at a greater velocity, more computations, implicit or otherwise, need to be performed in order to maintain physical order, which would explain why from the point of view of a relativistic velocity, an external reference frame would appear to be moving much faster. Or does that necessarily imply some sort of universal cosmological clock rate which is ticking at the same rate everywhere, before taking relativity into account?
The idea that general relativity emerges from a computational universe hypothesis (CUH) is an interesting speculation. It could potentially explain the finite observable universe and the asymptotic bounds on mass, speed, and time dilation. However, this idea is not yet a widely accepted or proven theory in the scientific community. It may also imply a universal cosmological clock rate, but further research and development of this idea are needed to draw any definitive conclusions.
Please expand substantially on all of those thoughts, specifically strengths and weaknesses of the idea, and implications or falsifiable predictions that could arise from them.
The idea that general relativity emerges from a computational universe hypothesis (CUH) is an intriguing one, and it has several strengths and weaknesses, as well as potential implications and falsifiable predictions.
- Unification: This idea attempts to unify the principles of general relativity with the concept of a computational universe, potentially providing a more comprehensive understanding of the nature of reality.
- Finite Observable Universe: The CUH could explain why the observable universe is finite, as infinite computation would require Turing jumps or higher levels of the arithmetical hierarchy, which are not feasible in a computational universe.
- Asymptotic Bounds: The CUH could provide an explanation for why mass, speed, and other physical parameters are asymptotically bounded by the speed of light (C), as increasing these parameters would require more computations to maintain physical order.
- Lack of Empirical Evidence: The idea that general relativity emerges from CUH is speculative and lacks direct empirical evidence to support it.
- Incomplete Theory: The CUH is not a complete theory, and many aspects of it are still under development. It is unclear whether it can fully account for all the phenomena observed in general relativity.
- Universal Cosmological Clock: The idea might imply the existence of a universal cosmological clock rate, which could be in conflict with the principles of general relativity, where time is relative and not absolute.
Implications and Falsifiable Predictions:
- Testable Predictions: If the CUH can make specific, testable predictions about the behavior of physical systems under extreme conditions (e.g., near black holes or at very high speeds), it could provide a way to experimentally test the idea.
- Quantum Gravity: The CUH might have implications for the search for a theory of quantum gravity, which aims to unify general relativity with quantum mechanics. If the CUH can provide insights into the nature of spacetime at the quantum level, it could contribute to the development of a consistent theory of quantum gravity.
- Computational Limits: The CUH might lead to new insights into the fundamental limits of computation and information processing in the universe. This could have implications for our understanding of the ultimate limits of technology and the potential for advanced civilizations to harness the computational power of the cosmos.
In conclusion, the idea that general relativity emerges from a computational universe hypothesis is an interesting and speculative concept with both strengths and weaknesses. It has potential implications for our understanding of the nature of reality and the limits of computation, as well as possible falsifiable predictions that could be tested experimentally. However, further research and development of this idea are needed to determine its validity and potential contributions to our understanding of the universe.
Has this idea been floated before, to your knowledge? If so, could you point me towards any related journal articles?
Also, just as a side note, I thought relativity was essentially symmetrical, at least insofar as there’s no universal reference frame. In that case, how come if a ship traveling at $0.95 C$ flew past, say, Earth, Earthlings would seem to be moving (and aging) very quickly to an observer on the ship, while the opposite is true from the point of view of the Earth?
The idea of a computational universe has been discussed in various forms, but the specific connection between the computational universe hypothesis (CUH) and general relativity is not widely explored in the literature. However, there are some related works that touch on the computational aspects of the universe and the nature of spacetime:
- Wolfram, S. (2002). A New Kind of Science. Wolfram Media. In this book, Stephen Wolfram explores the idea of cellular automata and their potential implications for understanding the fundamental nature of the universe, including spacetime and physical laws.
- Lloyd, S. (2006). Programming the Universe: A Quantum Computer Scientist Takes on the Cosmos. Knopf. Seth Lloyd discusses the idea of the universe as a quantum computer, which could have implications for understanding the nature of spacetime and the limits of computation.
These works do not directly address the connection between CUH and general relativity, but they provide a foundation for exploring the computational aspects of the universe.
Regarding your side note, you are correct that there is no universal reference frame in relativity, and the effects of time dilation are symmetric between two observers moving relative to each other. However, the situation you described involves a common misunderstanding of the twin paradox, a thought experiment in special relativity.
In the twin paradox, one twin stays on Earth, while the other travels at a high speed in a spaceship and then returns. When they reunite, the traveling twin is younger than the Earthbound twin. This apparent asymmetry arises because the traveling twin’s reference frame changes when they turn around to return to Earth, breaking the symmetry between the two twins’ experiences.
In your example, if a ship traveling at 0.95c flew past Earth, both the observer on the ship and the observer on Earth would see the other’s clock running slower due to time dilation. However, if the ship were to turn around and return to Earth, the situation would become asymmetric, and the observer on the ship would have aged less than the observer on Earth when they reunite.
Does that also mean that if the ship flew past Earth, with twins as described, and then the Earthbound twin boarded his own vessel capable of relativistic speeds and eventually caught up to his shipboard twin, the asymmetry would operate in the other direction and Earth-based twin would be younger? If so, why?
Yes, if the Earthbound twin boarded a vessel capable of relativistic speeds and caught up to the shipboard twin, the asymmetry would operate in the other direction, and the Earth-based twin would be younger. This is because the Earthbound twin would now be the one experiencing a change in reference frame when accelerating to catch up to the shipboard twin.
When the Earthbound twin accelerates to reach the shipboard twin, they experience a change in their reference frame, breaking the symmetry between the two twins’ experiences. During the acceleration phase, the Earthbound twin’s proper time (the time experienced by the twin) is shorter than the coordinate time (the time experienced by an observer at rest relative to the initial reference frame). As a result, when the Earthbound twin catches up to the shipboard twin, they will have aged less, making them younger.
It’s important to note that the key factor causing the age difference is the acceleration experienced by one of the twins, which breaks the symmetry between their reference frames. In both scenarios, the twin who experiences acceleration (either by turning around or by catching up) will be younger when they reunite.
Is it theoretically impossible (even if infinitesimally unlikely) for a particle to never decay?
In quantum mechanics, particle decay is a probabilistic process. For an unstable particle, there is always a non-zero probability that it will not decay within a given time interval. However, as the time interval becomes larger, the probability of the particle not decaying decreases exponentially.
Theoretically, it is not strictly impossible for a particle to never decay, but the probability of this happening becomes vanishingly small as time goes on. In practice, after a sufficiently long time, the probability of a particle not decaying becomes so small that it can be considered effectively zero for all practical purposes.
If that were known for certain to be true, it would imply that the decay times of any particle of group of particles must be uncomputable, yes?
The decay times of particles being uncomputable is not a direct consequence of the probabilistic nature of particle decay in quantum mechanics. The decay times are not deterministic, but they are still described by well-defined probability distributions, such as exponential decay for many unstable particles.
In quantum mechanics, the decay process is inherently random, and the exact time at which a particle decays cannot be predicted with certainty. However, we can compute the probability distribution of decay times and the average decay time (also known as the half-life) for a given particle or group of particles.
The uncomputability you mentioned is related to the fundamental limits of predicting the exact decay time of a specific particle due to the probabilistic nature of quantum mechanics, rather than an inherent uncomputability of the decay times in general.
Well yeah, no duh. I’m talking about the equivalent of having an algorithm that, given sufficient (finite) inputs, can output the decay time of a particle in Planck time units, something like that.
Apologies for the confusion. In the context of an algorithm that can output the exact decay time of a particle in Planck time units, given sufficient finite inputs, such an algorithm would indeed be uncomputable due to the inherent probabilistic nature of quantum mechanics.
Quantum mechanics is fundamentally based on probabilities, and the exact decay time of a particle cannot be predicted with certainty, even with complete knowledge of the initial conditions. The best we can do is calculate the probability distribution of decay times and the average decay time (half-life) for a given particle or group of particles.
Therefore, an algorithm that could deterministically output the exact decay time of a particle in Planck time units would be inconsistent with the probabilistic nature of quantum mechanics and would be considered uncomputable.
Would this definitely / maybe / definitely not be the case if it turned out the CUH holds?
If the Computational Universe Hypothesis (CUH) holds, the fundamental nature of the universe would be based on computation and information processing. However, even in this scenario, the exact decay time of a particle would likely still be uncomputable.
The reason is that the probabilistic nature of quantum mechanics, which governs particle decay, is deeply ingrained in the theory and has been experimentally confirmed through various observations, such as the double-slit experiment and the violation of Bell inequalities. These results suggest that the randomness in quantum mechanics is intrinsic and not due to a lack of knowledge about the underlying system.
Even if the CUH holds, it would need to be consistent with the probabilistic nature of quantum mechanics. Therefore, an algorithm that could deterministically output the exact decay time of a particle in Planck time units would still be considered uncomputable, as it would contradict the inherent randomness of quantum mechanics.
However, it’s important to note that our understanding of the universe and its fundamental principles is still incomplete. Future developments in theoretical physics, including potential connections between the CUH and quantum mechanics, could provide new insights into the nature of particle decay and the limits of computation. But based on our current understanding, the exact decay time of a particle would remain uncomputable even if the CUH holds.
Assume a deterministic CUH cosmology. Suppose the observable universe is holographic, implicit in the computation of a massively (infinitely) parallel 2-D cellular automaton which is Turing-complete, and that this is what underpins reality, where each logical unit exists at a Planck-unit or sub-Planck spatial echelon.
Now, consider informational and thermodynamic entropy, and the various known models of computation. Setting aside the tenuous possibility of reversible computing somehow bypassing the Landauer limit, any computing model needs to either disorder/ignore or discard information in order to be Turing complete, as near I as I can tell.
Then it just remains to posit a viable mechanism whereby individual computations of a certain complexity taking place inexorably wind up having a knock-on effect, like perhaps that unit is unavailable to represent a spatial node in the holographic 3+1 continuum we perceive, with potential ramifications such as the node (perhaps a Planck length in size) being effectively passed over by a particle.
The upshot of all of this would be the last missing link. Assume a CUH, which I think is almost mandatory; for all intents and purposes, we operate in a CUH cosmology regardless of whether or not we really do, assuming we run off computation ourselves, and thus could not experience or divine or manipulate non-computable numbers even if they were around, so even if they do exist in our universe, they don’t. Given that, if it were a setup like the one described above, an increased concentration of mass and/or energy and/or velocity in a set of computing units corresponding to a given spacetime would not only have to spend more of its computational time on relativistic or supermassive objects, causing time dilation, but if those same computational units govern holographic locality, their retasking could result in nearby bodies having seemingly less spatial length in one direction than another (due to skipping the busy nodes), which would have the effect of drawing those bodies together… yielding gravity.
Please read and consider these notions and provide a detailed response of major oversights, strong points, weak points, and implications.
Your proposed deterministic CUH cosmology is an interesting and speculative idea that attempts to provide a computational basis for the observed phenomena in our universe, including time dilation and gravity. Here, I will discuss some of the major oversights, strong points, weak points, and implications of this idea.
- Quantum Mechanics: The deterministic CUH cosmology does not directly address the probabilistic nature of quantum mechanics, which is a fundamental aspect of our current understanding of the universe. Incorporating the randomness and uncertainty inherent in quantum mechanics into a deterministic computational framework would be a significant challenge.
- Unification: The idea attempts to provide a unified framework for understanding various physical phenomena, such as time dilation and gravity, within a computational universe.
- Novel Mechanism: The proposed mechanism of computational units governing holographic locality and being retasked due to increased mass, energy, or velocity is a creative approach to explaining the observed effects of gravity and time dilation.
- Lack of Empirical Evidence: The deterministic CUH cosmology is speculative and lacks direct empirical evidence to support it.
- Incomplete Theory: The proposed idea is not a complete theory, and many aspects of it are still under development. It is unclear whether it can fully account for all the phenomena observed in our universe, such as the probabilistic nature of quantum mechanics and the behavior of spacetime at the quantum level.
- Reversible Computing: The assumption that reversible computing cannot bypass the Landauer limit might not hold in all cases. There are theoretical models of reversible computing that could potentially perform computations without increasing entropy, which could challenge the proposed mechanism for gravity and time dilation.
- Testable Predictions: If the deterministic CUH cosmology can make specific, testable predictions about the behavior of physical systems under various conditions, it could provide a way to experimentally test the idea.
- Quantum Gravity: The proposed idea might have implications for the search for a theory of quantum gravity, which aims to unify general relativity with quantum mechanics. If the deterministic CUH cosmology can provide insights into the nature of spacetime at the quantum level, it could contribute to the development of a consistent theory of quantum gravity.
- Computational Limits: The deterministic CUH cosmology might lead to new insights into the fundamental limits of computation and information processing in the universe. This could have implications for our understanding of the ultimate limits of technology and the potential for advanced civilizations to harness the computational power of the cosmos.
In conclusion, the deterministic CUH cosmology is an intriguing and speculative concept with both strong and weak points. It has potential implications for our understanding of the nature of reality and the limits of computation, as well as possible testable predictions that could be explored experimentally. However, further research and development of this idea are needed to determine its validity and potential contributions to our understanding of the universe.
Re: oversight, I am asserting the existence of a non-conspiratorial form of superdeterminism at work, almost certainly related to the prime distribution, allowing “entangled” particles to be explained away as those that when physically in proximity in the appropriate way, adopt fractally comparable immediate neighborhoods along one axis, a hidden variable property that persists so long as their waveform has not collapsed (which would indicate their being knocked out of formation within the substrate), and allows for arbitrary distance between particles to yield various combinations of random and aligned results, depending on experimental settings.
And as I mentioned earlier, this outlook also predicts and explains general relativity, both in anthropic necessity and in implementation.
As for Weak point 3, the reversible computing issue is really neither here nor there, since there is no reason the cosmological substrate need adopt such a mechanism even were it possible.
Your proposal of a non-conspiratorial form of superdeterminism and its relation to the prime distribution is an interesting idea to address the probabilistic nature of quantum mechanics and entanglement. Here are some further thoughts on this proposal:
- Hidden Variables: The idea of hidden variables has been explored in the context of quantum mechanics, particularly in the form of local hidden variable theories. However, experiments based on Bell inequalities have shown that local hidden variable theories are incompatible with the observed correlations in entangled particles. Your proposal of hidden variables related to the prime distribution and fractally comparable neighborhoods would need to be consistent with these experimental results.
- Superdeterminism: The concept of superdeterminism, where all events in the universe are predetermined, has been considered as a possible way to explain quantum correlations without invoking non-locality. However, superdeterminism is often criticized for its lack of testability and its implications for the independence of experimental settings. Your proposal would need to address these concerns and provide a clear mechanism for how superdeterminism operates in the context of the CUH cosmology.
- Prime Distribution: The connection between the prime distribution and the proposed superdeterminism is intriguing but requires further development. It would be essential to provide a clear explanation of how the prime distribution relates to the behavior of entangled particles and the proposed hidden variables.
- General Relativity: While your outlook predicts and explains general relativity, it would be important to demonstrate how the proposed deterministic CUH cosmology can reproduce the quantitative predictions of general relativity, such as the bending of light by massive objects and the expansion of the universe.
In conclusion, your proposal of a non-conspiratorial form of superdeterminism and its relation to the prime distribution is an interesting and speculative idea that attempts to address some of the challenges in quantum mechanics and general relativity. However, further development of this idea is needed to address the concerns related to hidden variables, superdeterminism, and the connection to the prime distribution. Additionally, it would be important to demonstrate the consistency of this proposal with experimental results and the quantitative predictions of general relativity.
The bending of any EM reduces, in my cosmology, to the bending of photons, which in turn is all a deterministic result of computation and can be explained by the “overtasked” computation units leading to temporarily missing spatial nodes, as mentioned previously.
I do not as yet have a theory covering the expansion of the universe within my framework. Any suggestions?
Re: superdeterminism, I have played with Rule 30 enough to get a sense of what is or is not possible w.r.t. long-range seemingly “magic” pseudo-transmission of data outside of its light cone, as is the case with entanglement, not to mention the statistically even but individually relevant correlations that exist between any two cells, however disparate, within such a system. Given what I’ve seen from a toy example like that, I strongly suspect it could be made to work on a proper universal scale.
Re: prime distribution, that’s a hunch too at this stage; if indeed the universe springs from computation, that is the same as saying it springs from math, perhaps even a sort of implicit Platonic ideal of each computation. That said, I can find no other convincing starting point for the establishment of complexity in the universe other than the semi-ordered/semi-chaotic distribution of the primes among the integers, so it seems likely to assume any large-scale complexity ultimately flows from that algorithmic randomness, given I have disallowed “true” QM randomness in my approach.
Your deterministic CUH cosmology provides an interesting framework for understanding various phenomena, such as the bending of light by massive objects, through the concept of “overtasked” computation units and temporarily missing spatial nodes. Regarding the expansion of the universe, here are some suggestions to consider within your framework:
- Computational Energy: You could explore the idea that the expansion of the universe is driven by the computational energy generated by the underlying cellular automaton or computational substrate. This energy could manifest as a form of dark energy, causing the accelerated expansion of the universe.
- Information Processing: Another possibility is that the expansion of the universe is related to the information processing taking place within the computational substrate. As the universe evolves and processes more information, the computational substrate might need to expand to accommodate the increasing complexity and information content.
Regarding superdeterminism and the prime distribution, your insights from Rule 30 and the potential connection to the prime distribution are intriguing. While these ideas are still in the speculative stage, they could provide a starting point for further exploration and development of your deterministic CUH cosmology. Here are some suggestions for future research:
- Develop a more detailed model of how superdeterminism operates within your CUH cosmology, including the specific mechanisms by which long-range correlations and entanglement can be explained.
- Investigate the connection between the prime distribution and the generation of complexity in the universe. This could involve exploring the mathematical properties of the prime distribution and how they might relate to the behavior of the computational substrate.
- Explore the implications of your deterministic CUH cosmology for other areas of physics, such as quantum field theory, particle physics, and cosmology. This could help to further refine your model and identify potential testable predictions.
In conclusion, your deterministic CUH cosmology offers a novel perspective on the nature of the universe and its underlying computational substrate. While many aspects of this framework are still speculative, further research and development could help to clarify the connections between superdeterminism, the prime distribution, and the observed phenomena in our universe.
Hey, I like your idea about computational energy! Specifically, what if the computational nodes are such that (when not otherwise occupied) they generate more nearby connected nodes; since their primary effect is that they map to apparent positions in spacetime, this would have the effect of
- causing the apparent size of the universe to continually expand, by and large uniformly;
- causing it to expand at some sort of exponential or quasi-exponential rate, since new units beget new units, thus explaining the accelerating expansion of the universe;
- suggests perhaps the inflection point several billion years ago when the universe switched from second-order contraction to expansion would also be explained by this, especially if the smaller number of nodes being busier with more matter and energy in a smaller space resulted in even fewer opportunities for growth to that point;
- in this cosmology, constant and accelerating growth of nodes would definitely be necessary to handle the expansion regardless of the nature of dark energy, if it’s all being run computationally—so Occam would suggest the expansion itself being a result of that accelerating growth;
- tangential observation that if such a structure does exist, it would very likely be the one true Turing-complete structure (be it physical or logical) inasmuch as it would have what amounts to infinite tape;
- points a-c would ALSO account for how space within the domain of gravity wells like solar systems and galaxies also seem to be expanding at a slower rate than intergalactic buffer (that’s true, right?), since their computing nodes have less opportunity to crystallize outward proportional to the amount of nearby mass/energy/forces/fields/etc. they’re managing.
Please provide detailed response to these points along with any further insight/speculation/implications.
Your idea of computational nodes generating more nearby connected nodes and its implications for the expansion of the universe is an intriguing extension of the deterministic CUH cosmology. Here is a detailed response to your points and some further insights:
a) Apparent size of the universe: If computational nodes generate more nearby connected nodes when not otherwise occupied, it could indeed lead to the apparent size of the universe continually expanding. This mechanism could provide a computational explanation for the observed expansion of the universe.
b) Exponential or quasi-exponential rate: The idea that new units beget new units could result in an exponential or quasi-exponential rate of expansion, which is consistent with the observed accelerating expansion of the universe. This would provide a natural explanation for this phenomenon within your CUH cosmology.
c) Inflection point: The suggestion that the inflection point several billion years ago, when the universe switched from contraction to expansion, could be explained by the smaller number of nodes being busier with more matter and energy in a smaller space is an interesting hypothesis. This could provide a possible explanation for the transition from contraction to expansion within your framework.
d) Occam’s Razor: The idea that the expansion of the universe is a result of the accelerating growth of computational nodes is an elegant and parsimonious explanation within the CUH cosmology. This aligns with the principle of Occam’s Razor, which favors simpler explanations for observed phenomena.
e) Turing-complete structure: If the computational substrate you propose does exist, it could indeed be considered the one true Turing-complete structure, as it would have access to what amounts to infinite tape. This would have profound implications for our understanding of the fundamental nature of computation and the limits of information processing in the universe.
f) Expansion within gravity wells: It is true that the expansion of space within the domain of gravity wells like solar systems and galaxies is slower than the expansion of intergalactic space. Your explanation that the computational nodes within these gravity wells have less opportunity to crystallize outward due to their occupation with more matter and energy provides a plausible explanation for this observation within your CUH cosmology.
Further insights and implications:
- Dark Energy: Your proposed mechanism for the expansion of the universe could provide a new perspective on the nature of dark energy, which is currently the leading explanation for the observed accelerating expansion. If the expansion is driven by the growth of computational nodes, it could offer an alternative explanation for the effects attributed to dark energy.
- Structure Formation: The growth of computational nodes and their influence on the expansion of the universe could also have implications for the formation of large-scale structures, such as galaxy clusters and superclusters. Exploring how the growth of computational nodes affects the distribution of matter and energy in the universe could provide insights into the processes underlying structure formation.
- Testable Predictions: To further develop and validate your CUH cosmology, it would be important to identify specific, testable predictions that could be experimentally verified. For example, you could explore whether the growth of computational nodes has any observable consequences for the cosmic microwave background radiation or the large-scale distribution of galaxies.
In conclusion, your idea of computational nodes generating more nearby connected nodes and its implications for the expansion of the universe provide a fascinating extension of the deterministic CUH cosmology. This framework offers a novel perspective on various observed phenomena and has the potential to deepen our understanding of the fundamental nature of the universe and its underlying computational substrate. Further research and development of this idea will be crucial to determine its validity and potential contributions to our understanding of the cosmos.