August 2018

Keating, Brian. Losing the Nobel Prize. New York: W. W. Norton, 2018. ISBN 978-1-324-00091-4.
Ever since the time of Galileo, the history of astronomy has been punctuated by a series of “great debates”—disputes between competing theories of the organisation of the universe which observation and experiment using available technology are not yet able to resolve one way or another. In Galileo's time, the great debate was between the Ptolemaic model, which placed the Earth at the centre of the solar system (and universe) and the competing Copernican model which had the planets all revolving around the Sun. Both models worked about as well in predicting astronomical phenomena such as eclipses and the motion of planets, and no observation made so far had been able to distinguish them.

Then, in 1610, Galileo turned his primitive telescope to the sky and observed the bright planets Venus and Jupiter. He found Venus to exhibit phases, just like the Moon, which changed over time. This would not happen in the Ptolemaic system, but is precisely what would be expected in the Copernican model—where Venus circled the Sun in an orbit inside that of Earth. Turning to Jupiter, he found it to be surrounded by four bright satellites (now called the Galilean moons) which orbited the giant planet. This further falsified Ptolemy's model, in which the Earth was the sole source of attraction around which all celestial bodies revolved. Since anybody could build their own telescope and confirm these observations, this effectively resolved the first great debate in favour of the Copernican heliocentric model, although some hold-outs in positions of authority resisted its dethroning of the Earth as the centre of the universe.

This dethroning came to be called the “Copernican principle”, that Earth occupies no special place in the universe: it is one of a number of planets orbiting an ordinary star in a universe filled with a multitude of other stars. Indeed, when Galileo observed the star cluster we call the Pleiades, he saw myriad stars too dim to be visible to the unaided eye. Further, the bright stars were surrounded by a diffuse bluish glow. Applying the Copernican principle again, he argued that the glow was due to innumerably more stars too remote and dim for his telescope to resolve, and then generalised that the glow of the Milky Way was also composed of uncountably many stars. Not only had the Earth been demoted from the centre of the solar system, so had the Sun been dethroned to being just one of a host of stars possibly stretching to infinity.

But Galileo's inference from observing the Pleiades was wrong. The glow that surrounds the bright stars is due to interstellar dust and gas which reflect light from the stars toward Earth. No matter how large or powerful the telescope you point toward such a reflection nebula, all you'll ever see is a smooth glow. Driven by the desire to confirm his Copernican convictions, Galileo had been fooled by dust. He would not be the last.

William Herschel was an eminent musician and composer, but his passion was astronomy. He pioneered the large reflecting telescope, building more than sixty telescopes. In 1789, funded by a grant from King George III, Herschel completed a reflector with a mirror 1.26 metres in diameter, which remained the largest aperture telescope in existence for the next fifty years. In Herschel's day, the great debate was about the Sun's position among the surrounding stars. At the time, there was no way to determine the distance or absolute brightness of stars, but Herschel decided that he could compile a map of the galaxy (then considered to be the entire universe) by surveying the number of stars in different directions. Only if the Sun was at the centre of the galaxy would the counts be equal in all directions.

Aided by his sister Caroline, a talented astronomer herself, he eventually compiled a map which indicated the galaxy was in the shape of a disc, with the Sun at the centre. This seemed to refute the Copernican view that there was nothing special about the Sun's position. Such was Herschel's reputation that this finding, however puzzling, remained unchallenged until 1847 when Wilhelm Struve discovered that Herschel's results had been rendered invalid by his failing to take into account the absorption and scattering of starlight by interstellar dust. Just as you can only see the same distance in all directions while within a patch of fog, regardless of the shape of the patch, Herschel's survey could only see so far before extinction of light by dust cut off his view of stars. Later it was discovered that the Sun is far from the centre of the galaxy. Herschel had been fooled by dust.

In the 1920s, another great debate consumed astronomy. Was the Milky Way the entire universe, or were the “spiral nebulæ” other “island universes”, galaxies in their own right, peers of the Milky Way? With no way to measure distance or telescopes able to resolve them into stars, many astronomers believed spiral neublæ were nearby objects, perhaps other solar systems in the process of formation. The discovery of a Cepheid variable star in the nearby Andromeda “nebula” by Edwin Hubble in 1923 allowed settling this debate. Andromeda was much farther away than the most distant stars found in the Milky Way. It must, then be a separate galaxy. Once again, demotion: the Milky Way was not the entire universe, but just one galaxy among a multitude.

But how far away were the galaxies? Hubble continued his search and measurements and found that the more distant the galaxy, the more rapidly it was receding from us. This meant the universe was expanding. Hubble was then able to calculate the age of the universe—the time when all of the galaxies must have been squeezed together into a single point. From his observations, he computed this age at two billion years. This was a major embarrassment: astrophysicists and geologists were confident in dating the Sun and Earth at around five billion years. It didn't make any sense for them to be more than twice as old as the universe of which they were a part. Some years later, it was discovered that Hubble's distance estimates were far understated because he failed to account for extinction of light from the stars he measured due to dust. The universe is now known to be seven times the age Hubble estimated. Hubble had been fooled by dust.

By the 1950s, the expanding universe was generally accepted and the great debate was whether it had come into being in some cataclysmic event in the past (the “Big Bang”) or was eternal, with new matter spontaneously appearing to form new galaxies and stars as the existing ones receded from one another (the “Steady State” theory). Once again, there were no observational data to falsify either theory. The Steady State theory was attractive to many astronomers because it was the more “Copernican”—the universe would appear overall the same at any time in an infinite past and future, so our position in time is not privileged in any way, while in the Big Bang the distant past and future are very different than the conditions we observe today. (The rate of matter creation required by the Steady State theory was so low that no plausible laboratory experiment could detect it.)

The discovery of the cosmic background radiation in 1965 definitively settled the debate in favour of the Big Bang. It was precisely what was expected if the early universe were much denser and hotter than conditions today, as predicted by the Big Bang. The Steady State theory made no such prediction and was, despite rear-guard actions by some of its defenders (invoking dust to explain the detected radiation!), was considered falsified by most researchers.

But the Big Bang was not without its own problems. In particular, in order to end up with anything like the universe we observe today, the initial conditions at the time of the Big Bang seemed to have been fantastically fine-tuned (for example, an infinitesimal change in the balance between the density and rate of expansion in the early universe would have caused the universe to quickly collapse into a black hole or disperse into the void without forming stars and galaxies). There was no physical reason to explain these fine-tuned values; you had to assume that's just the way things happened to be, or that a Creator had set the dial with a precision of dozens of decimal places.

In 1979, the theory of inflation was proposed. Inflation held that in an instant after the Big Bang the size of the universe blew up exponentially so that all the observable universe today was, before inflation, the size of an elementary particle today. Thus, it's no surprise that the universe we now observe appears so uniform. Inflation so neatly resolved the tensions between the Big Bang theory and observation that it (and refinements over the years) became widely accepted. But could inflation be observed? That is the ultimate test of a scientific theory.

There have been numerous cases in science where many years elapsed between a theory being proposed and definitive experimental evidence for it being found. After Galileo's observations, the Copernican theory that the Earth orbits the Sun became widely accepted, but there was no direct evidence for the Earth's motion with respect to the distant stars until the discovery of the aberration of light in 1727. Einstein's theory of general relativity predicted gravitational radiation in 1915, but the phenomenon was not directly detected by experiment until a century later. Would inflation have to wait as long or longer?

Things didn't look promising. Almost everything we know about the universe comes from observations of electromagnetic radiation: light, radio waves, X-rays, etc., with a little bit more from particles (cosmic rays and neutrinos). But the cosmic background radiation forms an impenetrable curtain behind which we cannot observe anything via the electromagnetic spectrum, and it dates from around 380,000 years after the Big Bang. The era of inflation was believed to have ended 10−32 seconds after the Bang; considerably earlier. The only “messenger” which could possibly have reached us from that era is gravitational radiation. We've just recently become able to detect gravitational radiation from the most violent events in the universe, but no conceivable experiment would be able to detect this signal from the baby universe.

So is it hopeless? Well, not necessarily…. The cosmic background radiation is a snapshot of the universe as it existed 380,000 years after the Big Bang, and only a few years after it was first detected, it was realised that gravitational waves from the very early universe might have left subtle imprints upon the radiation we observe today. In particular, gravitational radiation creates a form of polarisation called B-modes which most other sources cannot create.

If it were possible to detect B-mode polarisation in the cosmic background radiation, it would be a direct detection of inflation. While the experiment would be demanding and eventually result in literally going to the end of the Earth, it would be strong evidence for the process which shaped the universe we inhabit and, in all likelihood, a ticket to Stockholm for those who made the discovery.

This was the quest on which the author embarked in the year 2000, resulting in the deployment of an instrument called BICEP1 (Background Imaging of Cosmic Extragalactic Polarization) in the Dark Sector Laboratory at the South Pole. Here is my picture of that laboratory in January 2013. The BICEP telescope is located in the foreground inside a conical shield which protects it against thermal radiation from the surrounding ice. In the background is the South Pole Telescope, a millimetre wave antenna which was not involved in this research.

BICEP2 and South Pole Telescope, 2013-01-09

BICEP1 was a prototype, intended to test the technologies to be used in the experiment. These included cooling the entire telescope (which was a modest aperture [26 cm] refractor, not unlike Galileo's, but operating at millimetre wavelengths instead of visible light) to the temperature of interstellar space, with its detector cooled to just ¼ degree above absolute zero. In 2010 its successor, BICEP2, began observation at the South Pole, and continued its run into 2012. When I took the photo above, BICEP2 had recently concluded its observations.

On March 17th, 2014, the BICEP2 collaboration announced, at a press conference, the detection of B-mode polarisation in the region of the southern sky they had monitored. Note the swirling pattern of polarisation which is the signature of B-modes, as opposed to the starburst pattern of other kinds of polarisation.

B-mode polarisation in BICEP2 observations, 2014-03-17

But, not so fast, other researchers cautioned. The risk in doing “science by press release” is that the research is not subjected to peer review—criticism by other researchers in the field—before publication and further criticism in subsequent publications. The BICEP2 results went immediately to the front pages of major newspapers. Here was direct evidence of the birth cry of the universe and confirmation of a theory which some argued implied the existence of a multiverse—the latest Copernican demotion—the idea that our universe was just one of an ensemble, possibly infinite, of parallel universes in which every possibility was instantiated somewhere. Amid the frenzy, a few specialists in the field, including researchers on competing projects, raised the question, “What about the dust?” Dust again! As it happens, while gravitational radiation can induce B-mode polarisation, it isn't the only thing which can do so. Our galaxy is filled with dust and magnetic fields which can cause those dust particles to align with them. Aligned dust particles cause polarised reflections which can mimic the B-mode signature of the gravitational radiation sought by BICEP2.

The BICEP2 team was well aware of this potential contamination problem. Unfortunately, their telescope was sensitive only to one wavelength, chosen to be the most sensitive to B-modes due to primordial gravitational radiation. It could not, however, distinguish a signal from that cause from one due to foreground dust. At the same time, however, the European Space Agency Planck spacecraft was collecting precision data on the cosmic background radiation in a variety of wavelengths, including one sensitive primarily to dust. Those data would have allowed the BICEP2 investigators to quantify the degree their signal was due to dust. But there was a problem: BICEP2 and Planck were direct competitors.

Planck had the data, but had not released them to other researchers. However, the BICEP2 team discovered that a member of the Planck collaboration had shown a slide at a conference of unpublished Planck observations of dust. A member of the BICEP2 team digitised an image of the slide, created a model from it, and concluded that dust contamination of the BICEP2 data would not be significant. This was a highly dubious, if not explicitly unethical move. It confirmed measurements from earlier experiments and provided confidence in the results.

In September 2014, a preprint from the Planck collaboration (eventually published in 2016) showed that B-modes from foreground dust could account for all of the signal detected by BICEP2. In January 2015, the European Space Agency published an analysis of the Planck and BICEP2 observations which showed the entire BICEP2 detection was consistent with dust in the Milky Way. The epochal detection of inflation had been deflated. The BICEP2 researchers had been deceived by dust.

The author, a founder of the original BICEP project, was so close to a Nobel prize he was already trying to read the minds of the Nobel committee to divine who among the many members of the collaboration they would reward with the gold medal. Then it all went away, seemingly overnight, turned to dust. Some said that the entire episode had injured the public's perception of science, but to me it seems an excellent example of science working precisely as intended. A result is placed before the public; others, with access to the same raw data are given an opportunity to critique them, setting forth their own raw data; and eventually researchers in the field decide whether the original results are correct. Yes, it would probably be better if all of this happened in musty library stacks of journals almost nobody reads before bursting out of the chest of mass media, but in an age where scientific research is funded by agencies spending money taken from hairdressers and cab drivers by coercive governments under implicit threat of violence, it is inevitable they will force researchers into the public arena to trumpet their “achievements”.

In parallel with the saga of BICEP2, the author discusses the Nobel Prizes and what he considers to be their dysfunction in today's scientific research environment. I was surprised to learn that many of the curious restrictions on awards of the Nobel Prize were not, as I had heard and many believe, conditions of Alfred Nobel's will. In fact, the conditions that the prize be shared no more than three ways, not be awarded posthumously, and not awarded to a group (with the exception of the Peace prize) appear nowhere in Nobel's will, but were imposed later by the Nobel Foundation. Further, Nobel's will explicitly states that the prizes shall be awarded to “those who, during the preceding year, shall have conferred the greatest benefit to mankind”. This constraint (emphasis mine) has been ignored since the inception of the prizes.

He decries the lack of “diversity” in Nobel laureates (by which he means, almost entirely, how few women have won prizes). While there have certainly been women who deserved prizes and didn't win (Lise Meitner, Jocelyn Bell Burnell, and Vera Rubin are prime examples), there are many more men who didn't make the three laureates cut-off (Freeman Dyson an obvious example for the 1965 Physics Nobel for quantum electrodynamics). The whole Nobel prize concept is capricious, and rewards only those who happen to be in the right place at the right time in the right field that the committee has decided deserves an award this year and are lucky enough not to die before the prize is awarded. To imagine it to be “fair” or representative of scientific merit is, in the estimation of this scribbler, in flying unicorn territory.

In all, this is a candid view of how science is done at the top of the field today, with all of the budget squabbles, maneuvering for recognition, rivalry among competing groups of researchers, balancing the desire to get things right with the compulsion to get there first, and the eye on that prize, given only to a few in a generation, which can change one's life forever.

Personally, I can't imagine being so fixated on winning a prize one has so little chance of gaining. It's like being obsessed with winning the lottery—and about as likely.

In parallel with all of this is an autobiographical account of the career of a scientist with its ups and downs, which is both a cautionary tale and an inspiration to those who choose to pursue that difficult and intensely meritocratic career path.

I recommend this book on all three tracks: a story of scientific discovery, mis-interpretation, and self-correction, the dysfunction of the Nobel Prizes and how they might be remedied, and the candid story of a working scientist in today's deeply corrupt coercively-funded research environment.

 Permalink

Kroese, Robert. The Dream of the Iron Dragon. Seattle: CreateSpace, 2018. ISBN 978-1-9837-2921-8.
The cover tells you all you need to know about this book: Vikings!—spaceships! What could go wrong? From the standpoint of a rip-roaring science fiction adventure, absolutely nothing: this masterpiece is further confirmation that we're living in a new Golden Age of science fiction, made possible by the intensely meritocratic world of independent publishing sweeping aside the politically-correct and social justice warrior converged legacy publishers and re-opening the doors of the genre to authors who spin yarns with heroic characters, challenging ideas, and red-blooded adventure just as in the works of the grandmasters of previous golden ages.

From the standpoint of the characters in this novel, a great many things go wrong, and there the story begins. In the twenty-third century, humans find themselves in a desperate struggle with the only other intelligent species they'd encountered, the Cho-ta'an. First contact was in 2125, when a human interstellar ship was destroyed by the Cho-ta'an while exploring the Tau Ceti system. Shortly thereafter, co-ordinated attacks began on human ships and settlements which indicated the Cho-ta'an possessed faster-than-light travel, which humans did not. Humans formed the Interstellar Defense League (IDL) to protect their interests and eventually discovered and captured a Cho-ta'an jumpgate, which allowed instantaneous travel across interstellar distances. The IDL was able to reverse-engineer the gate sufficiently to build their own copies, but did not understand how it worked—it was apparently based upon some kind of wormhole physics beyond their comprehension.

Humans fiercely defended their settlements, but inexorably the Cho-ta'an advanced, seemingly driven by an inflexible philosophy that the universe was theirs alone and any competition must be exterminated. All attempts at diplomacy failed. The Earth had been rendered uninhabitable and evacuated, and most human settlements destroyed or taken over by the Cho-ta'an. Humanity was losing the war and time was running out.

In desperation, the IDL set up an Exploratory Division whose mission was to seek new homes for humans sufficiently distant from Cho-ta'an space to buy time: avoiding extinction in the hope the new settlements would be able to develop technologies to defend themselves before the enemy discovered them and attacked. Survey ship Andrea Luhman was en route to the Finlan Cluster on such a mission when it received an enigmatic message which seemed to indicate there was intelligent life out in this distant region where no human or Cho-ta'an had been known to go.

A complex and tense encounter leaves the crew of this unarmed exploration ship in possession of a weapon which just might turn the tide for humanity and end the war. Unfortunately, as they start their return voyage with this precious cargo, a Cho-ta'an warship takes up pursuit, threatening to vaporise this last best hope for survival. In a desperate move, the crew of the Andrea Luhman decide to try something that had never been attempted before: thread the needle of the rarely used jumpgate to abandoned Earth at nearly a third of the speed of light while evading missiles fired by the pursuing warship. What could go wrong? Actually a great deal. Flash—darkness.

When they got the systems back on-line, it was clear they'd made it to the Sol system, but they picked up nothing on any radio frequency. Even though Earth had been abandoned, satellites remained and, in any case, the jumpgate beacon should be transmitting. On further investigation, they discovered the stars were wrong. Precision measurements of star positions correlated with known proper motion from the ship's vast database allowed calculation of the current date. And the answer? “March sixteen, 883 a.d.

The jumpgate beacon wasn't transmitting because the jumpgate hadn't been built yet and wouldn't be for over a millennium. Worse, a component of the ship's main drive had been destroyed in the jump and, with only auxiliary thrusters it would take more than 1500 years to get to the nearest jumpgate. They couldn't survive that long in stasis and, even if they did, they'd arrive two centuries too late to save humanity from the Cho-ta'an.

Desperate situations call for desperate measures, and this was about as desperate as can be imagined. While there was no hope of repairing the drive component on-board, it just might be possible to find, refine, and process the resources into a replacement on the Earth. It was decided to send the ship's only lander to an uninhabited, resource-rich portion of the Earth and, using its twenty-third century technology, build the required part. What could go wrong? But even though nobody on the crew was named Murphy he was, as usual, on board. After a fraught landing attempt in which a great many things go wrong, the landing party of four finds themselves wrecked in a snowfield in what today is southern Norway. Then the Vikings show up.

The crew of twenty-third century spacefarers have crashed in the Norway of Harald Fairhair, who was struggling to unite individual bands of Vikings into a kingdom under his rule. The people from the fallen silver sky ship must quickly decide with whom to ally themselves, how to communicate across a formidable language barrier and millennia of culture, whether they can or dare meddle with history, and how to survive and somehow save humanity in what is now their distant future.

There is adventure, strategy, pitched battles, technological puzzles, and courage and resourcefulness everywhere in this delightful narrative. You grasp just how hard life was in those days, how differently people viewed the world, and how little all of our accumulated knowledge is worth without the massive infrastructure we have built over the centuries as we have acquired it.

You will reach the end of this novel wanting more and you're in luck. Volume two of the trilogy, The Dawn of the Iron Dragon (Kindle edition), is now available and the conclusion, The Voyage of the Iron Dragon, is scheduled for publication in December, 2018. It's all I can do not to immediately devour the second volume starting right now.

The Kindle edition is free for Kindle Unlimited subscribers.

 Permalink

Rand, Ayn. Ideal. New York: New American Library, 2015. ISBN 978-0-451-47317-2.
In 1934, the 29 year old Ayn Rand was trying to establish herself in Hollywood. She had worked as a junior screenwriter and wardrobe person, but had not yet landed a major writing assignment. She wrote Ideal on speculation, completing the 32,000 word novella and then deciding it would work better as a stage play. She set the novella aside and finished the play version in 1936. The novella was never published nor was the play produced during her lifetime. After her death in 1982, the play was posthumously published in the anthology The Early Ayn Rand, but the novella remained largely unknown until this edition, which includes both it and the play, was published in 2015.

Ideal is the story of movie idol Kay Gonda, a beautiful and mysterious actress said to have been modeled on Greta Garbo. The night before the story begins, Gonda had dinner alone with oil baron Granton Sayers, whose company, it was rumoured, was on the brink of ruin in the depths of the Depression. Afterwards, Sayers was found in his mansion dead of a gunshot wound, and Gonda was nowhere to be found. Rumours swirled through the press that Gonda was wanted for murder, but there was a blackout of information which drove the press and her studio near madness. Her private secretary said that she had not seen Gonda since she left for the dinner, but that six pieces of her fan mail were missing from her office at the studio, so she assumed that Gonda must have returned and taken them.

The story then describes six episodes in which the fugitive Kay Gonda shows up, unannounced, at the homes of six of her fans, all of whom expressed their utter devotion to her in their letters. Five of the six—a henpecked manager of a canning company, an ageing retiree about to lose the house in which he raised his children, an artist who paints only canvases of Ms Gonda who has just won first prize in an important exhibition, an evangelist whose temple faces serious competition from the upstart Church of the Cheery Corner, and a dissipated playboy at the end of his financial rope—end up betraying the idol to whom they took pen to paper to express their devotion when confronted with the human being in the flesh and the constraints of the real world. The sixth fan, Johnnie Dawes, who has struggled to keep a job and roof over his head all his adult life, sees in Kay Gonda an opportunity to touch a perfection he had never hoped to experience in his life and devises a desperate plan to save Gonda from her fate.

A surprise ending reveals that much the reader has assumed is not what really happened, and that while Kay Gonda never once explicitly lied, neither did she prevent those to whom she spoke from jumping to the wrong conclusions.

This is very minor Ayn Rand. You can see some of the story telling skills which would characterise her later work beginning to develop, but the story has no plot: it is a morality tale presented in unconnected episodes, and the reader is left to draw the moral on his or her own. Given that the author was a struggling screenwriter in an intensely competitive Hollywood, the shallowness and phoniness of the film business is much on display here, although not so explicitly skewered as the later Ayn Rand might have done. The message is one of “skin in the game”—people can only be judged by what they do when confronted by difficult situations, not by what they say when words are cheap.

It is interesting to compare the play to the novella. The stories are clearly related, but Rand swaps out one of the fans, the elderly man, for a young, idealistic, impecunious, and totally phoney Communist activist. The play was written in 1936, the same year as We the Living, and perhaps the opportunity to mock pathetic Hollywood Bolsheviks was too great to pass by.

This book will mostly be of interest to those who have read Ayn Rand's later work and are curious to read some of the first fiction she ever wrote. Frankly, it isn't very good, and an indication of this is that Ayn Rand, whose reputation later in life would have made it easy to arrange publication for this work, chose to leave it in the trunk all her life. But she did not destroy the manuscript, so there must have been some affection for it.

 Permalink