Category Archives: Cosmology

Adventures in Theoretical Physics I – Entropy

Updated 3Jun21

Entropy is one of those slippery concepts floating around in theoretical physics that defies easy explication. The following quote is from a small subsection of an extensive Wikipedia entry on entropy:

The following is a list of additional definitions of entropy from a collection of textbooks:

-a measure of energy dispersal at a specific temperature.

-a measure of disorder in the universe or of the availability of the energy in a system to do work.

-a measure of a system’s thermal energy per unit temperature that is unavailable for doing useful work.

In Boltzmann’s definition, entropy is a measure of the number of possible microscopic states (or microstates) of a system in thermodynamic equilibrium. Consistent with the Boltzmann definition, the second law of thermodynamics needs to be re-worded as such that entropy increases over time, though the underlying principle remains the same.

Note that this is a “list of additional definitions”, there are seemingly innumerable others peppered throughout the article. It would be a hard to caricature the overall impression of incoherence, oozing like quicksand and sucking under any rational attempt to say concisely what, exactly, entropy is, beyond the usual handwaving about a vaguely specified measure of disorder.

The first sentence of the Wiki article states unequivocally that “Entropy is… a measurable physical property…”. That statement is false. In section 7.8 it is made clear that entropy cannot be measured directly as a property. Rather, discrete energy transfers are dropped into an equation that ostensibly “describes how entropy changes dS when a small amount of energy dQ is introduced into the system at a certain temperature T.” The only measurements made are of energy and temperature. Entropy is a made-up mathematical concoction; It is not a real property of physical reality.

The reader will also be pleased to note that the same article informs us that “In 1865, Clausius named the concept of “the differential of a quantity which depends on the configuration of the system,” entropy… “.

The real problem with the concept of entropy is not that it may be useful in doing some thermodynamic calculations. Clearly, it is of some mathematical value? What it does not have is any clear explanatory power or physical meaning, and well, math just isn’t physics. Nonetheless, entropy has been elevated to a universal, physical law – the Second Law of Thermodynamics, where the situation doesn’t get any less opaque. In fact the opacity doesn’t get any more opaque than this gem:

Nevertheless, this principle of Planck is not actually Planck’s preferred statement of the second law, which is quoted above, in a previous sub-section of the present section of this present article, and relies on the concept of entropy.

Section 2.9 of the 2nd Law wiki entry above

Here is Planck’s quote from “…a previous sub-section of the present section of this present article…”:

Every process occurring in nature proceeds in the sense in which the sum of the entropies of all bodies taking part in the process is increased. In the limit, i.e. for reversible processes, the sum of the entropies remains unchanged.

Of course, the only question that remains then is, what exactly is entropy?

The final word on the nature of entropy should be this comment by John von Neumann which appears in a sidebar of this section, wherein he advises Claude Shannon to call his information uncertainty concept entropy because “… In the second place, and more important, nobody knows what entropy really is, so in a debate you will always have the advantage.” Can’t argue with that logic. It appears that the definition of entropy is entropic. It is getting ever more disordered with the passage of time.

So what’s the problem? This is after all, a relatively trivial matter compared to the Grand Delusions of Theoretical Physics, like Dark Matter, Dark Energy, Quarks & etc. Or is it? Well, actually the Entropy Paradigm represents a major blind spot in fundamental physics. Theoretical physics has this misbegotten, laboratory based idea that all ordered systems are in some sense, “running down” or losing the ability to do work, or becoming more disordered or, or… something, called Entropy.

Entropy is all downhill, all the time. In the context of a designed laboratory experiment, the original, closed-system condition is some prearranged, ordered-state, that is allowed to “run down” or “become disordered” or whatever, over time and that, more or less is Entropy! Fine so far, but when you make Entropy a universal law – The Second Law of Thermodynamics, all is lost, or more exactly half of reality is lost.

Why? Because there are no closed systems outside the laboratory. What you get when you ignore that fact is this kind of bloated nonsense:

The law that entropy always increases holds, I think, the supreme position among the laws of Nature. If someone points out to you that your pet theory of the universe is in disagreement with Maxwell’s equations – then so much the worse for Maxwell’s equations. If it is found to be contradicted by observation – well, these experimentalists do bungle things sometimes. But if your theory is found to be against the second law of thermodynamics I can give you no hope; there is nothing for it but to collapse in deepest humiliation. 

Sir Arthur Stanley Eddington, The Nature of the Physical World (1927)

Entropy is correct, if all it boils down to is, ordered, closed, systems will become disordered. But entropy doesn’t tell you how the systems came to be ordered in the first place. If you extrapolate from a closed-system experiment, a Universal Law that says, or implies, that all systems rundown, but you have no theory of how those systems came to be ordered in the first place, you have exactly half a picture of reality – because over here in the real world, ordered systems do arise from disorder.

It’s hard to do good physics if you have this half-blind view of reality. You start babbling about the Arrow of Time, the heat death of the Universe, not to mention whatever this is. You may have to invoke an unobservable Creation Miracle like the Big Bang to get the ball rolling.

Babble Update (3Jun21)

“If you want your clock to be more accurate, you’ve got to pay for it,” study co-author Natalia Ares, a physicist at the University of Oxford, told Live Science. “Every time we measure time, we are increasing the universe’s entropy.”

Livescience.com

Meanwhile, physical reality doesn’t work like that. Physical reality is locally cyclic – like life itself. There is, for instance, a considerable body of observational evidence suggesting that galaxies reproduce themselves. That body of evidence has been dismissed with spurious statistical arguments, and the well-established astronomer who made them, Halton Arp, was denied telescope time to continue making those observations. No one has since been permitted to pursue that line of observational research.

Meanwhile, there is also no comparable body of observational evidence for the Big Bang supported model of galaxy formation, via accretion, from a primordial, post-BB cloud. Like the rest of the Big Bang’s defining features, galaxy formation by accretion is not an observed phenomenon in the Cosmos.

So, an over-generalization (to the point of incoherence), of some simplistic closed-system laboratory experiments done in the 18th and 19th centuries, has come to aid and abet, the absurd Big Bang model of the Cosmos. Modern theoretical physics is an incoherent, unscientific mess.

Starts With a Bang, Ends With Nonsense

Ethan Siegel’s columns for Forbes magazine are tour de forces of the illogical circular reasoning that passes for scientific discourse in the modern theoretical physics community. They are best avoided out of respect for the good name of science. It is, however, sometimes instructive, not to mention eye-opening, to slog through one, in order to appreciate just how self-deluded the pseudo-science known as modern cosmology has become.

Case in point: this article from Dec 22, 2020. Therein, Siegel is at great pains to demonstrate that dark matter and dark energy are the only scientific way of accounting for observed cosmological phenomena in the context of the Standard Model of Cosmology, commonly known as the Big Bang model. This is true of course, the BB model can only account for cosmological observations if it is allowed to invoke the existence dark matter and dark energy – so therefore dark matter and dark energy must exist despite the absence of any empirical evidence for their existence. Reasoning can’t get any more circular than that.

In some of Siegel’s arguments there is almost no logical argument at all, just a recitation of certain facts followed by an assertion unsupported by the facts presented, that those factual observations were predictions rather than post-dictions of the model (emphasis added):

What’s remarkable is that, because the laws of physics that govern particles (and nuclear fusion) are so well-understood, we can compute exactly — assuming the Universe was once hotter, denser, and expanded and cooled from that state — what the different ratios of these different light elements ought to be. We’ve even studied the reactions in the lab directly, and things behave precisely as our theory predicts. The only factor we vary is the photon-to-baryon ratio, which tells us how many cosmic photons (particles of light) there are for every proton or neutron (the baryons) in our Universe.

We’ve now measured it all. Satellites like COBE, WMAP, and Planck have measured how many photons there are in the Universe: 411 per cubic centimeter of space. Intervening clouds of gas that appear between us and a distant light source, like a luminous galaxy or quasar, will absorb a fraction of the light as it travels through the Universe, teaching us the abundances of these elements and isotopes directly. When we add it all up, only ~5% of the total energy in the Universe can be normal matter: no more and no less.

What a mess! The central assumption (emphasized above) is simply the BB model, which is itself based on two foundational assumptions:

  1. The cosmos is a unitary, coherent, simultaneous entity – a “Universe”.
  2. The cosmological redshift is caused by some form of recessional velocity.

Neither of those assumptions has any empirical basis and both date to the early 20th century when the scale of the cosmos was barely grasped. They are however, sacrosanct among cosmologists. Those foundational assumptions are, by some common but unstated agreement, true by definition in the cosmological community. They are more than just assumptions, they are beliefs, and they are almost certainly wrong. But, what of that… Assuming our assumptions are correct we can use math to massage our model into agreement with observations – just like good old Ptolemy did way back around the dawn of the previous Dark Ages.

More recently Siegel has posted a similarly obtuse argument concerning the supposed successes of the Big Bang model and its unchallengeable status. It features the same kind of circular reasoning as above: Our model is correct as long as our assumptions and postulates are correct, therefore the things we assume and postulate are correct, because the model is correct. The result of this transparently illogical syllogism, which underlies all of the author’s triumphalist claims for the Big Bang model, is the nonsensical narrative that winds from inexplicable Big Bang to the now 95% invisible “Universe”.

The reason there are no significant challenges to the Big Bang orthodoxy from within the scientific community is that the community members have been trained to accept the 100 year old premises of the model. So, again, the author is correct to this extent: if you accept the model’s premises you are going to wind up with some ludicrous depiction of the Cosmos like LCDM. However, the simplistic assumptions of 100 years ago have proven a disastrous foundation for modern cosmology. The model depicts an inane, unscientific, “Universe” that does not at all resemble, in its defining features, the Cosmos we actually observe.

In the Cosmos we observe there is no Big Bang event, no inflation, no expanding spacetime, no dark matter, no dark energy. All of those things are essential elements of LCDM . None of those things are present in the Cosmos we observe. Physical reality does not look like the LCDM “Universe”. People like the author, who, nonetheless, believe the absurd LCDM story, do so because they are functionally, mathematicists.

Mathematicism is an ancient, half-baked philosophy whose proponents believe that mathematics underlies and determines the nature of physical reality. So, dark matter (& etc.) cannot be empirically demonstrated to exist? That is of no concern to a mathematicist; it all has to be there because a peculiar model (LCDM) requires it in order to reconcile itself with observations, and therefore dark matter (& etc.) have to exist. Pay no attention to the testimony of your lyin’ eyes (and telescopes)!

Mathematicism is junk philosophy. It is not science. How we got to this state of affairs is going to provide plenty of work for future historians of science. For now though, there is much to do just to pry the cold, dead hands of mathematicism from the throat of theoretical physics (which is now barely alive in any scientific sense).

Science rests on observation (detection) and measurement, not the febrile imaginings of reality-challenged, nerd-mathematicists with a cult-model to defend. What passes for science among the theoretical physics community of the academy these days is not science in any meaningful sense. The Big Bang model is Exhibit A for that proposition.

(Parts of this post were taken from a comment I made on Siegel’s Medium article.)

Junk Science in The New Yorker

So, an interesting and even-handed account of the recent acknowledgement by the US government of ongoing investigations into UFO phenomena contains this whopper:

Interstellar travel by living beings still seems like a wildly remote possibility, but physicists have known since the early nineteen-nineties that faster-than-light travel is possible in theory, and new research has brought this marginally closer to being achievable in practice.

This is not an accurate scientific assessment of the possibility of faster-than-light travel. What it appears to be referring to is the mathematicist fantasy of an Alcubierre warp-drive. You can find a succinct presentation of this “theory” here. It is a short and not technically difficult presentation which seems to reach a straightforward conclusion:

At present, such a thing just doesn’t seem to be entirely within the realm of possibility. And attempts to prove otherwise remain unsuccessful or inconclusive.

Things would have been fine if they left it at that but the nerd community, having had their minds warped by too many repeated viewings of Star Wars-Trek fantasy-fiction- movies, had to add a more hopeful note:

But as history has taught us, what is considered to be impossible changes over time. Someday, who knows what we might be able to accomplish?

Unfortunately history has also taught us that the human imagination is not a reliable guide to the nature of physical reality. The Alcubierre drive rests on the scientifically baseless presumption that spacetime is a causally interacting element of physical reality. There is no empirical evidence for such a presumption.

There is, however, a significant contingent of Academy Certified Scientists who believe in the existence of a physical spacetime. This belief is entirely based on the assumption, by the mid-20th century neo-Relativists (Misner, Thorne, Wheeler, et al), that such a spacetime exists. However, real science cannot be based on what MTW (or you or I) believe. Real science rests on empirical evidence, not on, in this case, the spurious reification of relational concepts.

Like Star Wars itself, the Alcubierre drive is a pseudo-scientific fantasy that amuses and confuses a certain reality-challenged subset of humanity, many of whom, unfortunately, are puking up this indigestible, hairball of irrationality in the halls of the Scientific Academy. The New Yorker is not to blame for this mess, but thoughtlessly repeating such unscientific nonsense, as if it had a scientific basis, is not helpful.

The Unprincipled Principle & The Persistence of Mistakes

Over at Triton Station, Stacy McGaugh has a series of posts on the recent, and ongoing, successful predictions of MOND at the galactic scale, beginning with the December 31, 2020 post, 25 Years A Heretic, which provides an excellent historical overview of the situation. It is thankfully, solidly couched in empirical observations.

The last two posts in the Comments section of the linked post are by Apass and Phillip Helbig. I had intended to reply to them in a timely fashion but by the time I got around to it the Comment section was closed. Following are the replies I intended to post:

@Apass,

… through what mechanism would this new disk model give rise to the EFE felt by a satellite galaxy isolated by tens of thousands of light years where you cannot invoke anymore viscosity-like effects?

Given that the mechanism through which gravity causes its effects is generally unknown, that’s an odd question. If a theoretical model of disk dynamics limning the mass distribution replicated MOND’s success in describing the rotational curves, I would be surprised if it didn’t also account for the EFE. After all, satellite galaxies are not, by definition, gravitationally isolated from their host galaxy.

The fact that DM didn’t predict the EFE is probably a function of the fact that DM doesn’t successfully predict anything except itself. It is always deployed, as a patch to a failure of an implementation, on significantly larger scales, of the standard gravitational model of the solar system. If the EFE holds up, which I expect it will, there will likely be a concerted effort to post-dict it with DM. This would present an existential problem to the SMoC proponents however, as the EFE overturns the unprincipled Strong Equivalence Principle.

The SEP is not an expression of empirical fact, the way the weak principle of equivalence is (the equality of inertial and gravitational mass); it is little more than an idle conjecture of mathematical convenience, adopted mid-twentieth century, right at the onset of the post-Einsteinian revision of General Relativity.

The SEP is not foundational to the SMoC, but it is well integrated into the mindset of its practitioners and has consequences for the theoretical “study” of theoretical entities like black holes and multibody systems like galaxy clusters. Dispensing with the SEP would, for one thing, mean that cosmologists could now stop believing, contrary to the empirical evidence and GR prediction, that the speed of light in a vacuum is a universal constant.

@Phillip Helbig

Note that in the standard cosmological model, only 1920s cosmology is used. Despite the huge amount of high-quality data now available, classical cosmology needs nothing beyond what was known about relativistic cosmology in the 1920s.

Well yes, and that’s why we need a 21st century cosmology with, shall we say, more realistic underpinnings. I really don’t understand why anyone would think clinging to the naive and dubious assumptions of 100 years ago is a scientifically sound approach, especially given the manifest failures of the current model. I understand that the SMoC is graded on the curve, with negative empirical results deeply discounted in favor of theoretical requirements – but still, the model’s defining structural elements all lie totally in the realm of the empirically undetected.

The fact that there appears to be absolutely no interest in revisiting the standard model’s foundational assumptions, even as an exercise in intellectual curiosity, says nothing good about the socio-economic incentives of the modern scientific academy. The Big Bang model is the best possible you say? I would agree, but only with the caveat that it is only so in the context of the model’s essentially pre-modern assumptions. As soon as those are jettisoned, the big bang creation myth will almost certainly become nothing more than a historical curiosity.

Nobel Prize in Physics Awarded to 3 Scientists for Work on Fairy Tale

It’s sadly, depressingly true. Some mathematicist nonsense that attends modern cosmological theory has been awarded science’s grandest prize:


This award has been given to mathematical work on a mathematical theory that has nothing to do with actual physics. Black holes, the ostensible subject matter of the work are fictitious entities that only arise in a mathematical models (such as Schwarzschild solution to the field equations of General Relativity) that ignore one of the fundamental consequences of General Relativity – the speed of light varies with position in a gravitational well.

Specifically, the steeper the gravitational well, the slower the speed of light. It follows directly from this, an observed fact, that gravitational collapse is a self-limiting event. In the Schwarzschild and similar solutions, the speed of light is held constant and results in a theoretical collapsed state (a black hole) that is physically nonsensical, containing as it does a singularity (a mass of zero-volume and therefore infinite density).

Modern theoretical cosmology is a mess, in which Einstein’s General Relativity has been dumbed-down in service of mathematical convenience. The result has been dumb-downed, mathematical models (the Big Bang, black holes), that present physically nonsensical accounts of observed reality. These models do not remotely resemble the physical reality we can actually observe and measure. They are mathematical fairy tales.

The Nobel Prize in Physics was just awarded for work on a dumbed-down mathematical model that has nothing to do with real physics.

A Brief History Of The Universe According To ΛCDM

Once upon a time, 13.8 billion years ago, the entirety of the ΛCDM model Universe was compressed into a volume smaller than a gnat’s ass (a statement of universal simultaneity forbidden by Relativity Theory).

The original condition of this model Universe is inexplicable in terms of known physics. Regardless, the model Universe erupted from its inexplicable condition, for an unknown reason, in a somewhat explicable cosmic flatulence event.

With the aid of an inflaton field the model Universe then expanded into today’s model Universe, which is 95% composed of things that are not observed in the Cosmos (Dark Matter and Dark Energy). The remaining 5% of the model Universe consists of the things that are observed in the Cosmos. 

This model should not be classified as science. It constitutes nothing more than the feverish imaginings of the mathematicist mindset that has utterly debased the field of cosmological studies.

A General Critique of The Standard Model of Cosmology (ΛCDM)

Foundational Assumptions

  1. The Cosmos is a singular, unified, simultaneous entity (a Universe) that can be modeled as such using the field equations of General Relativity. This assumption invokes the existence of a universal frame – the FLRW metric, which is antithetical to the foundational premise of Relativity Theory, that a universal frame does not exist. Solving the equations of GR for a universal frame is an oxymoronic exercise. The existence of a universal frame is impossible to empirically verify.
  2. The cause of the observed cosmological redshift is some form of recessional velocity. A recessional velocity produces a redshift, but not all redshifts are caused by recessional velocities. There is no empirical evidence supporting this assumption.

Consequent Supporting Hypotheses

  1. The model Universe had a singular origin – the Big Bang event. The Big Bang is not an observed component of the Cosmos. It is only a model dependent inference.
  2. Subsequent to the Big Bang event, the model Universe underwent an Inflation event driven by an inflaton field. Neither the Inflation event nor the inflaton field are observed components of the Cosmos. They constitute a model dependent hypothesis necessary to reconcile the model with observations.
  3. The Universe is expanding because spacetime is expanding and driving the galaxies apart. There is no evidence for the existence of a substantival spacetime, that is, for the existence of a spacetime that can interact causally with matter and energy. A substantival spacetime, one that can drive the galaxies apart, is not an observed component of the Cosmos.
  4. Dark Matter comprises @ 85% of the total matter content of the model Universe. The only salient characteristic of Dark Matter is that it reconciles the standard model with observations. Dark Matter is not an observed component of the Cosmos.
  5. Dark Energy comprises @ 69% of the matter-energy content of the model Universe. The only salient characteristic of Dark Energy is that it reconciles the standard model with observations. Dark Energy is not an observed component of the Cosmos

The Big Bang Hustle

(Correction appended 2020-06-21)

Comes before us then, the cosmologist duo of Luke A. Barnes and Geraint F. Lewis, a couple of Aussies, with a Universe they want to sell you. The pitch, as presented in this promo for their new book, is that the Big Bang explains certain observations that cannot be explained by any competing model and any competing model must explain them. The argument rests on three points that aren’t so much about observations, but in two cases are merely superficial claims to explaining things that are essentially self-explanatory in terms of known physics, without invoking the big bang. The third point rests on the unjustified contention that the big bang interpretation is the only acceptable, existing explanation for the cosmological redshift.

The three points discussed in the video and their fundamental weaknesses are:

Olber’s Paradox – Olber’s paradox is inherently resolved by the cosmological redshift and this resolution does not depend on the Big Bang assumption that the redshift is a consequence of some form of recessional velocity. The cosmological redshift exists and if the cosmos is of sufficient size there will be a distance beyond which light from a galaxy is redshifted to a null energy state or, more likely, will be reduced to a minimum state where it will pool with all other such distant sources to form a microwave “background” from the point of view of a local observer.

The Cosmological Redshift – That the redshift is caused by some form of recessional velocity, from which cosmologists reason backwards to the Big Bang event, is simply an assumption of the standard model. If the assumption is rather that what expands are the spherical wavefronts of light being emitted by galaxies, you get a picture that is consistent with observations and which can be treated with a standard General Relativity formalism to yield a cosmological redshift-distance relationship – no big bang, with its ludicrously inexplicable original condition, required.

Nucleosynthesis – The cosmic abundances of the elements are what they are. BB theorists have simply imported nuclear theory and fit it to their model. Model fitting is a technique as old as Ptolemy. As the linked article makes clear there are many possible pathways to nucleosynthesis. The BB version also doesn’t even work that well.

Don’t know much about dark matter and dark energy… Barnes and Lewis kind of joke about this, which is appropriate I guess, since according to the standard model, 95% of the model’s Universe is composed of this unobserved and, by their account, poorly understood stuff. Which brings us to one of the central conceits of the video and, presumably, the book it is promoting, which is that the standard model “explains” satisfactorily, all of our cosmological observations.

The problem with the standard model’s “explanations” for observed phenomena is precisely that they are not scientifically satisfactory. They are not satisfactory because “explanations” such as dark matter, and dark energy are not empirically resolvable. Dark matter and dark energy are necessary to the standard model, in order to fit the model to observations. Physical reality, however, has no use for either, in as much as, no evidence for their physical existence can be found beyond this urgent need to somehow fit the standard model to observations.

By invoking, as necessary components of physical reality, entities for which there is no observational evidence, modern cosmology has descended into a pre-Enlightenment medieval mind-set. Dark matter and dark energy are just the modern, secular version of angels and devils dancing on the head of a pin. Their scientific, “explanatory” power in nil.

The video also features such an egregious scientific error as to be almost breathtaking. At the 5:50 minute mark, in an attempt to justify their claim that only the expanding universe interpretation of redshift accounts for Olber’s Paradox, Barnes makes the completely fallacious argument that the more distant a luminous object is, its apparent size diminishes, but not its apparent brightness. That is simply wrong.

Even the most perfunctory treatment of light theory would include a discussion of the luminosity-distance relationship. That someone with a Phd (I presume in science), would choose to ignore basic scientific knowledge in order to advance a spurious claim about the explanatory power of the standard model, says all you need to know about the state of modern cosmology. It’s a mess.

Correction (2020-06-21): In a conversation with Louis Marmet over at A Cosmology Group, he pointed out that technically Barnes was correct; he was speaking of surface brightness, which does not change with distance for an extended object. Stars, however, are not resolvable as extended objects by the human eye, only by the largest telescopes. The apparent magnitude of point-like sources, such as stars varies according to the luminosity-distance relationship. Barnes’ argument remains false – on its own terms.

Forbes’ Physics Follies & The Broken Clock Syndrome

Sure enough, it had to happen. Ethan Siegel finally wrote a column for Forbes that isn’t just a thoughtless regurgitation of, and apologia for, the inanities of modern theoretical physics. In fact, it’s a column I can wholeheartedly endorse. It should be required reading for all physics students. A copy should be handed out with every science diploma. Here is the key takeaway:

Mathematics wasn’t at the root of the physical laws governing nature; it was a tool that described how the physical laws of nature manifested themselves. The key advance that happened is that science needed to be based in observables and measurables, and that any theory needed to confront itself with those notions. Without it, progress would be impossible.

Those are fine scientific sentiments, indeed! Unfortunately, there is no evidence of these noble precepts being applied in any of Siegel’s now numerous scientific writings for Forbes, or at least not in any I have read. There we find only the unthinking mathematicism that has turned modern theoretical physics into a caricature of bad science.

There is no more reliable purveyor of the modern scientific orthodoxy than Ethan Siegel. His Forbes column, Starts With A Bang, with the exception noted above, relentlessly flogs theoretical assertions that flagrantly violate the principles quoted above. Even in cases where he gets the orthodoxy completely wrong, he and his editors simply plow ahead, barely acknowledging an error.

So, you can still find this piece online despite the fact that the grandiose claims of the title (Scientists Discover Space’s Largest Intergalactic Bridge, Solving A Huge Dark Matter Puzzle), and of the concluding paragraphs are completely wrong. Here is the triumphal conclusion:

…If this same type of structure that exists between Abell 0399 and Abell 0401 also exists between other colliding clusters, it could solve this minor anomaly of the Bullet cluster, leaving dark matter as the sole unchallenged explanation for the displacement of gravitational effects from the presence of normal matter.

It’s always an enormous step forward when we can identify a new phenomenon. But by combining theory, simulations, and the observations of other colliding galaxy clusters, we can push the needle forward when it comes to understanding our Universe as a whole. It’s another spectacular victory for dark matter, and another mystery of the Universe that might finally be solved by modern astrophysics. What a time to be alive.

So what’s wrong? Well immediately following the last paragraph is this subsequently added correction:

Correction: after a Twitter exchange with one of the study’s scientists, the author regrets to inform the reader that the acceleration imparted by the magnetic fields to the electrons along this intergalactic bridge is likely unrelated to the velocity anomaly of the Bullet cluster. Although both may be explained by hydrodynamic effects, the effects that cause this radio emission and the acceleration of electrons are unrelated to the measured high velocity of the Bullet cluster’s collisional elements and X-ray gas. Ethan Siegel regrets the error.

Oh. So this correction completely negates the claim that the observations as described, of the colliding galaxy clusters Abell 399 and Abell 401, somehow clear up a known problem that the Dark Matter model has with another pair of colliding galaxies known as the Bullet Cluster.

The proponents of Dark Matter like to cite the Bullet Cluster as strong evidence for DM, but their model cannot account for the high collisional velocity of the Bullet Cluster’s component galaxies. It was this problem that Siegel incorrectly interpreted the Abell cluster observations to have solved. So this was just another strained attempt to justify the failed Dark Matter hypothesis (by inference only, of course) and even by the low-bar of the modern scientific academy, it was an abject failure.

Which brings us to a more recent piece featuring another central conceit of Big Bang Cosmology that, like Dark Matter, has no empirical evidence to support it. This is the belief that there exists a thing called “space” that physically interacts with the matter and energy in the cosmos. There is absolutely no empirical evidence to support this belief. Despite this absence of evidence, the belief in a substantival “space” is widely held among among modern cosmologists.

It is a characteristic of modern theoretical physics, that a lack of empirical evidence for its existence, does not and cannot, diminish the claim to existence for an entity necessary to reconcile either one of the standard models with observed reality. In the modern scientific paradigm, a theoretical model is the determinant of physical reality.

The entire premise of this article then is, that something which cannot be demonstrated to exist must, nonetheless, have a specific physical characteristic in order for the modern conception of theoretical physics to be consistent with itself. Non-existent space must be continuous not discrete because otherwise modern theoretical physics doesn’t make sense.

But this is nothing more than an old whine encoded in a new frequency. It is not news that the modern conception of Relativity Theory is inconsistent with Quantum Theory. Modern theoretical physics, generally speaking, is inconsistent, with both empirical reality, and with itself. Both standard models make assertions about physical reality that cannot be empirically demonstrated to be physically meaningful.

The only evidence on offer for all the empirically baseless assertions of the two standard models is that they are necessary to make the models work. Once the models have been ad hoc corrected for any discrepancies with observed reality, the “success” of the models is touted as proof of their description of physical reality.

There is much assertive hand-waving going on in this article, as is typical of Siegel’s work. There is one particular assertion I’d like to single out for close examination:

In Einstein’s relativity, an observer who’s moving with respect to another observer will appear to have their lengths compressed and will appear to have their clocks run slow.

At first glance this seems a reasonably accurate statement, but it is, in fact a false assertion. “Einstein’s relativity” encompasses both the Special and General theories. The statement as presented, in which a moving observer “appears to have their their lengths compressed and… their clocks run slower”, applies only to the Special Relativity case involving two inertial frames. It absolutely does not apply to those conditions where General Relativity applies, i.e. non-inertial frames (see this previous post).

In the SR case, two inertial frames are moving with a constant velocity with respect to each other and each observer sees the other’s lengths compressed, and their clocks slowed. Neither observer actually has lengths compressed or clocks slowed; they only appear that way to the remote observer.

Under GR conditions, where a change of reference frame has taken place (via acceleration), there are physical consequences for the observer in the frame that has accelerated. In the accelerated frame, lengths are contracted in the direction of the acceleration, and clocks do run slower. There are physical consequences to changing frames.

Modern theorists like Siegel are, however, inclined to believe that the SR and GR cases can be conflated, whenever it is convenient to do so, based only on an erroneous and logically unjustified over-extension of the Principle of Equivalence. On the basis of empirical observations and sound theoretical considerations, however, it is possible to say, unequivocally, that they are wrong, both theoretically and empirically. And you can’t be any more wrong than that.

There is only one way out of this theoretical mess and that is to apply the principles set forth in the No, The Universe Is Not Purely Mathematical In Nature column to the two standard models. On the evidence, unfortunately, Professor Siegel does not practice what he preaches, nor do most of his cohorts in the theoretical physics community. Thus, the “crisis in physics”.

On the Redshift-Distance Relationship

The quote below is from a comment by @Apass on Stacy McGaugh’s blog, Triton Station. Stacy suggested we continue the conversation elsewhere. The quote comes from the comment section on this blogpost. The complete thread to this point can be found there. My response follows the quote.

@budrap …
Friedmann is irrelevant for the discussion we have. It is true that Friedmann was the first to derive the solution for the the expanding Universe, but It was Lemaitre who proposed what he later called the “primeval atom” – i.e. the first idea / model for BB.
And let’s not confuse the model’s assumptions – he assumed that the Universe is expanding (after all, all galaxies seem to recess), but this single assumption does not constrain how the galaxies at distance should move away from us. They might be recessing at a lower pace, they might be recessing at a constant speed or they might even be not recessing at all in the very far distance. At the time when the observations that showed the nebulae redshifts were made, there was no identified correlation between redshift and distance so all those possibilities were open. Only when you add to the model also GR, the correlation (as later observed by Hubble) can be derived – and he did so, before Hubble.
To me, that counts as a prediction confirmed later by empirical observations.
As for Hubble – again, it’s irrelevant if he accepted or not the interpretation of the redsifts. That was his opinion to which he was fully entitled.
As for the bias – yes, any scientific model should be based solidly on verifiable statements, but I’m not that easily going to throw the baby with the bathwater.
In case you must discount some verifiable observations that appear to you or to anyone else to not conform with the model, you should give a thorough justification of why those observations are not relevant. And if you cannot provide this solid justification you’ll need to start questioning, at first, your assumptions / understanding of the model (and here is a very big bias filter) – maybe those observations really fit in there but you don’t see how.
And if this still doesn’t allow room for the observations, then you’ll need to question the model. Don’t throw it right away, but look at what is salvageable, if something, in the light of what can be empirically observed or tested (that is, don’t introduce exotic, non-baryonic dark matter and call the day).
And when / where the model doesn’t provide sufficient explanatory power, use “as if” instead of definitive statements.

Apass,

So, you admit the assumption, which is all I’ve claimed – that the recessional velocity interpretation is an assumption. I guess your argument is that because that assumption can produce an “expanding universe” model which predicts a redshift-distance relationship in agreement with observations, the matter is settled.

It is not, because you can get the same result with the counter assumption – that the redshift-distance relation is not caused by a recessional velocity but is simply a consequence of the loss of energy by an expanding spherical wavefront of light as it traverses the cosmos. To calculate it, all you need do is apply the GR equation for gravitationl redshift to the expanding sphere over significant cosmological intervals, incorporating all the enclosed mass at each iteration. You can do it on a spreadsheet. You get a redshift-distance relation.

To this comment: “… yes, any scientific model should be based solidly on verifiable statements, but I’m not that easily going to throw the baby with the bathwater“, I can only ask, what exactly constitutes the “baby” in your opinion? To me, the BB model is all bathwater, there is nothing salvageable. The situation is exactly analogous to geocentric cosmology; the foundational assumptions are simply are wrong. The model presents a completely inaccurate, not to mention unrealistic account of observed reality.

The Speed Of Light Is Not A Universal Constant

One of the great unforced errors of late 20th century theoretical physics was to declare the speed of light in a vacuum, a universal constant. This was done despite the fact that, according to General Relativity, the speed of light varies with position in a gravitational field. This variation in light speed has been observed.

The idea that the speed of light varies with position in a gravitational field is not new. It comes on good authority:

…according to the general theory of relativity, the law of the constancy of the velocity of light in vacuo, which constitutes one of the two fundamental assumptions in the special theory of relativity and to which we have already frequently referred, cannot claim any unlimited validity. A curvature of rays of light can only take place when the velocity of propagation of light varies with position. Now we might think that as a consequence of this, the special theory of relativity and with it the whole theory of relativity would be laid in the dust. But in reality this is not the case. We can only conclude that the special theory of relativity cannot claim an unlimited domain of validity: its results hold only so long as we are able to disregard the influences of gravitational fields on the phenomena (e.g. of light).

Albert Einstein, Relativity The Special And The General Theory, 15th edition

The claim of those who wish to maintain the universal constancy of light speed (c) is that it is justified by the Equivalence Principle. The rather lengthy linked discussion goes into some detail but no mention is made of the constancy of c being consequent on the EP. None of the tests of the principle cited in the article involve measuring the value of c directly.

Einstein invoked the EP in the derivation of General Relativity to provide an interpretation of the observed fact that inertial mass and gravitational mass are equivalent. Given the above quote, he obviously did not find justification therein for asserting the universal constancy of c.

The only reasonable conclusion is that the EP does not justify the claim of a universally constant light speed. The claim appears to be only a lazy mathematicist assumption of mathematical convenience. Theoretical physics is a remarkably irrational, unscientific mess because of this sort of nonsense.

More on LIGO

The latest from the New York Times:

https://www.nytimes.com/2017/10/16/science/ligo-neutron-stars-collision.html?hp&action=click&pgtype=Homepage&clickSource=story-heading&module=second-column-region&region=top-news&WT.nav=top-news

I have to admit that I find this type of pseudo-scientific puff piece conducive to depression. Nothing discussed here involves anything remotely to do with an actual observation. What is described instead, is a mathematical fiction, with such a slender and tenuous connection, to a dubious, claimed detection, it boggles the scientific imagination.

What we have here is a display, not of science, but of mathematicism, a disease of modern culture more reductive than reductionism. A large piece of 19th century machinery, polished and constructed, to spit out reams of data, destined to be massaged by elegant algorithms of deep complexity, in silent computers running ceaselessly, and laboring mightily to bring forth a shred of a signal, so flimsily beneath any possible sensitivity of the employed machinery, as to be laughable.

From this shredded evidentiary bean of dubious provenance, is spun a mighty mathematical beanstalk, a tale of fantastical proportions, with beasts of impossible construction, swirling about in a dance of destruction and laboring mightily to bring forth a shred of a signal, so flimsily beneath any possible sensitivity of the detecting machinery, as to be laughable.

This piece is nothing but a rewrite of a vapid press release, as substance free as a DJT tweet touting his magnificent intelligence. Science weeps, bound in a dungeon of mathematical formalisms.

[Comment submitted to NYT on 10/16/17. It may or may not get published.]

The LIGO Fiasco Rolls On

It has become clear now, after three so-called gravitational wave detections, that the LIGO enterprise is a wheezing exercise in mathematical fantasizing. It’s claims of a GW detection are without any real physical basis. LIGO is all math all the time with some updated 19th century technology appended to provide source data. That source date is then mathematically sculpted such that a quantum level signal detection is claimed which just happens to more or less agree with one out of a set of 250,000 precalculated ‘signal templates’.

Leave aside the question of whether it is physically possible for a large, classical scale, mechanical device to detect a quantum scale (10^{-20} m) displacement on an interferometer with two 4 km long arms. For now it is enough to point out that the more-or-less agreement between ‘signal’ and template is hardly sufficient scientific basis for the subsequent elaborate and exquisitely precise description of a binary black hole merger that supposedly took place in a distant, unobservable past.

Even allowing for the possibility of such a dubious physical detection there is no scientific justification for then running an inferential chain of logic back up a baroque mathematical model, filigreed as it is with numerous unsubstantiated assumptions about the nature of physical reality, as if there  no other possible source for such a quantum level ‘signal’. The claim is, of course, that all other possibilities have been excluded. On top of everything else we are thereby treated to a ludicrous mathematicist claim of quantum omniscience.

It is interesting to note that the problematic claim in the first GW detection paper (discussed here), of having received an intact signal from the entire black hole merger event, has been downplayed somewhat in the third paper. It is alluded to in a fashion that suggests a desire for the claim to be plausibly deniable as a mere misunderstanding at some future time when the issue is pressed.