Centauri Dreams
Imagining and Planning Interstellar Exploration
New Explanations for the Enigmatic Wow! Signal
The Wow! signal, a one-off detection at the Ohio State ‘Big Ear’ observatory in 1977, continues to perplex those scientists who refuse to stop investigating it. If the signal were terrestrial in origin, we have to explain how it appeared at 1.42 GHz, while the band from 1.4 to 1.427 GHz is protected internationally – no emissions allowed. Aircraft can be ruled out because they would not remain static in the sky; moreover, the Ohio State observatory had excellent RFI rejection. Jim Benford today discusses an idea he put forward several years ago, that the Wow signal could have originated in power beaming, which would necessarily sweep past us as it moved across the sky and never reappear. And a new candidate has emerged, as Jim explains, involving an entirely natural process. Are we ever going to be able to figure this signal out? Read on for the possibilities. A familiar figure in these pages, Jim is a highly regarded plasma physicist and CEO of Microwave Sciences, as well as being the author of High Power Microwaves, widely considered the gold standard in its field.
by James Benford
The 1977 Wow! signal had the potential of being the first signal from extraterrestrial intelligence. But searches for recurrence of the signal heard nothing. Interest continues, as two lines of thought continue to ponder it.
An Astronomical Maser
A recent paper proposes that the Wow! signal could be the first recorded event of an astronomical maser flare in the hydrogen line [1]. (A maser is a laser-like coherent emission at microwave frequencies. The maser was the precursor to the laser.) The Wow frequency was at the hyperfine transition line of neutral hydrogen, about 1.4 GHz. The suggestion is that the Wow was a sudden brightening from stimulated emission of the hydrogen line in interstellar gas driven by a transient radiation source behind a hydrogen cloud. The group is now going through archival data searching for other examples of abrupt brightening of the hydrogen line.
Maser Wow concept: A transient radiative source behind a cold neutral hydrogen (HI) cloud produced population inversion in the cloud near the hydrogen line, emitting a narrowband burst toward Earth [1].
Image courtesy of Abel Mendez (Planetary Habitability Laboratory, University of Puerto Rico at Arecibo).
Could aliens use the hydrogen clouds as beacons, triggered by their advanced technology? Abel Mendez has pointed out that this was suggested by Bob Dixon in a student’s thesis in 1976 [2]! From that thesis [3]:
“If it is a beacon built by astroengineering, such as an extraterrestrial civilization that is controlling the emission of a natural hydrogen cloud and using it as a beacon, then the only way that it could be ascertained as such, is by some time variation. And we are not set up to study time variation.”
How could such a beacon be built? It would require producing a population inversion in a substantial volume of ionized hydrogen. That might perhaps be done by an array of thermonuclear explosives optimized to produce a narrowband emission into such a volume [4]. Exploded simultaneously, they could produce that inversion, creating the pulse seen on Earth as the Wow.
Why does the Wow! Signal have narrow bandwidth?
In 2021, I published a suggestion that the enigmatic Wow Signal, detected in 1977, might credibly have been leakage from an interstellar power beam, perhaps from launch of an interstellar probe [5]. I used this leakage to explain the observed features of the Wow Signal: the power density received, the Signal’s duration and frequency. The power beaming explanation for the Wow accounted for all four of the Wow parameters, including the fact that the Wow observation has not recurred.
At the 2023 annual Breakthrough Discuss meeting, Mike Garrett of Jodrell Bank inquired “I was thinking about the Wow signal and your suggestion that it might be power beam leakage. But it’s not obvious to me why any technical civilization would limit their power beam to a narrow band of <= 10 kHz. Is there some kind of technical advantage to doing that or some kind of technical limitation that would produce such a narrow-band response?”
After thinking about it, I have concluded that there is ‘some kind of technical advantage’ to narrow bandwidth. In fact, it is required for high-power beaming systems.
Image: The Wow Signal was detected by Jerry Ehman at the Ohio State University Radio Observatory (known as the Big Ear). The signal, strong enough to elicit Ehman’s inscribed comment on the printout, was never repeated.
A Beamer Made of Amplifiers?
High power systems involving multiple sources are usually built using amplifiers not oscillators, for several technical reasons. For example, the Breakthrough Starshot system concept has multiple laser amplifiers driven by a master oscillator, a so-called master oscillator-power amplifier (MOPA) configuration. Amplifiers are themselves characterized by the product of amplifier gain (power out divided by power in) and bandwidth, which is fixed for a given type of device, their ‘gain-bandwidth product’. This product is due to phase and frequency desynchronization between the beam and electromagnetic field outside the frequency bandwidth [6].
Therefore, for wide bandwidth, a lower power per amplifier follows. That means many more amplifiers. Likewise, to go to high power, each amplifier will have a small bandwidth. (Then the number of amplifiers is determined by the power required.) For power beaming applications, to get high power on target is essential: higher power is required, so smaller bandwidth follows.
So why do you get narrow bandwidth? You use very high gain amplifiers to essentially “eat up” the gain-bandwidth product. For example, in a klystron, you have multiple high-Q cavities that result in high gain. The high-gain SLAC-type klystrons had gains of about 100,000. Bandwidths for high power amplifiers on Earth are about one percent of one percent, 0.0001, 10-4. The Wow! bandwidth is 10 kHz/1.41 GHz, about 10-5.
So yes, the physics of amplifiers limits bandwidth in beacons and power beams because both would be built to provide very high power. So, with very high gain in the amplifiers, small bandwidth is the result.
This fact about amplifiers is another reason I think power beaming leakage is the explanation for the Wow. Earth could have accidentally received the beam leakage. Since stars constantly move relative to each other, later launches using the Wow! beam will not be seen from Earth.
Therefore I predicted that each failed additional search for the Wow! to repeat is more evidence for this explanation.
The Wow search goes on
These two very different explanations for the origin of the Wow! have differing future possibilities. I predicted that it wouldn’t be seen again. Each failed additional search for the Wow! to repeat (and there have been many) is more evidence for this explanation. Mendez and coworkers are looking to see if their process has occurred previously. They can prove their explanation by finding such occurrences in existing data. These are two very different possibilities. Only the Mendez concept can be realized soon.
References
1. Abel Mendez,1 Kevin Ortiz, Ceballos, and Jorge I. Zuluaga, “Arecibo Wow! I: An Astrophysical Explanation for the Wow! Signal,” arXiv:2408.08513v2 [astro-ph.HE], 2024.
2. Abel Mendez, private communication.
3. Cole, D. M. (1976). “A Search for Extraterrestrial Radio Beacons at the Hydrogen Line,” Doctoral dissertation, Ohio State University, 1976.
4. Taylor, T., 1986, “Third generation nuclear weapons,” Sci. Am., 256, 4, 1986.
5. James Benford “Was the Wow Signal Due to Power Beaming Leakage?”, JBIS 74 196-200, 2021.
6. James Benford, Edl Schamiloglu, John A. Swegle, Jacob Stephens and Peng Zhang, Ch. 12 in High Power Microwaves, Fourth Edition, Taylor and Francis, Boca Raton, FL, 2024.
Into Titan’s Ocean
When it comes to oceans beneath the surface of icy moons, Europa is the usual suspect. Indeed, Europa Clipper should have much to say about the moon’s inner ocean when it arrives in 2030. But Titan, often examined for the possibility of unusual astrobiology, has an internal ocean too, beneath tens of kilometers of ice crust. The ice protects the mix of water and ammonia thought to be below, but may prove to be an impenetrable barrier for organic materials from the surface that might enrich it.
I’ve recently written about abused terms in astrobiological jargon, and in regards to Titan, the term is ‘Earth-like,’ which trips me every time I run into it. True, this is a place where there is a substance (methane) that can appear in solid, liquid or gaseous form, and so we have rivers and lakes – even clouds – that are reminiscent of our planet. But on Earth the cycle is hydrological, while Titan’s methane mixing with ethane blows up the ‘Earth-like’ description. For methane to play this role, we need surface temperatures averaging −179 °C. Titan is a truly alien place.
In Titan we have both an internal ocean world and a cloudy moon with an atmosphere thick enough to keep day/night variations to less than 2 ℃ (although the 16 Earth day long ‘day’ also stabilizes things). It’s a fascinating mix. We’ve only just begun to probe the prospects of life inside this world, a prospect driven home by new work just published in the Planetary Science Journal. Its analysis of Titan’s ocean as a possible home to biology doesn’t come up completely short, but it offers little hope that anything more than scattered microbes might exist within it.
Image: This composite image shows an infrared view of Titan from NASA’s Cassini spacecraft, acquired during a high-altitude fly-by, 10,000 kilometers above the moon, on Nov. 13, 2015. The view features the parallel, dark, dune-filled regions named Fensal (to the north) and Aztlan (to the south). Credit: NASA/JPL/University of Arizona/University of Idaho.
The international team led by Antonin Affholder (University of Arizona) and Peter Higgins (Harvard University) looks at what makes Titan unique among icy moons, the high concentration of organic material. The idea here is to use bioenergetic modeling to look at Titan’s ocean. Perhaps as deep as 480 kilometers, the ocean can be modeled in terms of available chemicals and ambient conditions, factoring in energy sources (likely from chemical reactions) that could allow life to emerge. Ultimately, the models rely on what we know of the metabolism of Earth organisms, which is where we have to start.
It turns out that abundant organics – complex carbon-based molecules – may not be enough. Here’s Affholder on the matter:
“There has been this sense that because Titan has such abundant organics, there is no shortage of food sources that could sustain life. We point out that not all of these organic molecules may constitute food sources, the ocean is really big, and there’s limited exchange between the ocean and the surface, where all those organics are, so we argue for a more nuanced approach.”
Nuance is good, and one way to explore its particulars is through fermentation. The process, which demands organic molecules but does not require an oxidant, is at the core of the paper. Fermentation converts organic compounds into the energy needed to sustain life, and would seem to be suited for an anaerobic environment like Titan’s. With abiotic organic molecules abundant in Titan’s atmosphere and accumulating at the surface, a biosphere that can feed off this material in the ocean seems feasible. Glycine, the simplest of all known amino acids and an abundant constituent of matter in the early Solar System, serves as a useful proxy, as it is widely found in asteroids and comets and could sink through the icy shell, perhaps as a result of meteorite impacts.
The problem is that this supply source is likely meager. From the paper:
Sustained habitability… requires an ongoing delivery mechanism of organics to Titan’s ocean, through impacts transferring organic material from Titan’s surface or ongoing water–rock interactions from Titan’s core. The surface-to-ocean delivery rate of organics is likely too small to support a globally dense glycine-fermenting biosphere (<1 cell per kg water over the entire ocean). Thus, the prospects of detecting a biosphere at Titan would be limited if this hypothetical biosphere was based on glycine fermentation alone.
Image: This artist’s concept of a lake at the north pole of Saturn’s moon Titan illustrates raised rims and rampart-like features as seen by NASA’s Cassini spacecraft. Credit: NASA/JPL-Caltech.
More biomass may be available in local concentrations, perhaps at the interface of ocean and ice, and the authors acknowledge that other forms of metabolism may add to what glycine fermentation can produce. Thus what we have is a study using glycine as a ‘first-order approach’ to studying the habitability of Titan’s ocean, one which should be followed up by considering other fermentable molecules likely to be in the ocean and working out how they might be delivered and in what quantity. We still wind up with only approximate biomass estimates, but the early numbers are low.
The paper is Affholder et al., “The Viability of Glycine Fermentation in Titan’s Subsurface Ocean,” Planetary Science Journal Vol. 6, No. 4 (7 April 2025), 86 (full text).
On ‘Sun-like’ Stars
The thought of a planet orbiting a Sun-like star began to obsess me as a boy, when I realized how different all the planets in our Solar System were from each other. Clearly there were no civilizations on any planet but our own, at least around the Sun. But if Alpha Centauri had planets, then maybe one of them was more or less where Earth was in relation to its star. Meaning a benign climate, liquid water, and who knew, a flourishing culture of intelligent beings. So ran my thinking as a teenager, but then other questions began to arise.
Was Alpha Centauri Sun-like? Therein hangs a tale. As I began to read astronomy texts and realized how complicated the system was, the picture changed. Two stars and perhaps three, depending on how you viewed Proxima, were on offer here. ‘Sun-like’ seemed to imply a single star with stable orbits around it, but surely two stars as close as Centauri A and B would disrupt any worlds trying to form there. Later we would learn that stable orbits are indeed possible around both of these stars, and in the habitable zone, too (I still get emails from people saying no such orbits are possible, but they’re wrong).
So just how broad is the term ‘Sun-like’? One textbook told me that Centauri A was a Sun-like star, while Centauri B was sometimes described that way, and sometimes not. I thought the confusion would dissipate once I knew enough about astronomical terms, but if you look at scientific papers today, you’ll see that the term is still flexible. The problem is that by the time we move from a paper into the popular press, the term Sun-like begins to lose definition.
Image:A direct image of a planetary system around a young K-class star, located about 300 light-years away and known as TYC 8998-760-1. The news release describes this as a Sun-like star, raising questions as to how we define the term. Image credit: ESO.
G-class stars are the category into which the Sun fits, with its lifetime on the order of 10 billion years, and if we expand its parameters a bit, we can take in stars around the Sun’s mass – perhaps 0.8 to 1.1 solar masses – and in the same temperature range – 5300 to 6000 K. Strictly speaking, then, Centauri B, which is a K-class star, doesn’t fit the definition of Sun-like, although K stars seem to be good options for habitable worlds. They’re less massive (0.45 to 0.8 solar masses) and cooler (3900 to 5300 K), and they’re more orange than yellow, and they’re longer lived than the Sun, a propitious thought in terms of astrobiology.
So we don’t want to be too doctrinaire in discussing what kind of star a habitable planet might orbit, but we do need to mind our definitions, because I see the term ‘Sun-like’ in so many different contexts. When I see a statement like “one Earth-class planet around every Sun-like star,” I have to ask whether we’re talking about G-class stars or something else. Because some scientists expand ‘Sun-like’ to include not only G- and K- but F-class stars as well. Why do this? Such stars are, like the Sun, long-lived. F stars are hotter and more massive than the Sun, but like G- and K-class stars, they’re stable. Some studies, then, consider ‘Sun-like’ to mean all FGK-type stars.
Some examples. ‘COROT finds exoplanet orbiting Sun-like star’ is the title of a news release that describes a star a bit more massive than the Sun. So the comparison is to G-class stars. ‘Astronomy researchers discover new planet around a ‘Sun-like’ star’ describes a planet around an F-class star, so we are in the FGK realm. ‘First Ever Image of a Multi-Planet System around a Sun-like Star Captured by ESO Telescope’ describes a planet around TYC 8998-760-1, a K-class dwarf.
So there’s method here, but it’s not always clarified as information moves from the academy (and the observatory) into the media. Confounding the picture still further are those papers that use ‘Sun-like’ to mean all stars on the Main Sequence. This takes in the entire range of stars OBAFGKM. This is a rare usage, but there is a certain logic here as well. If you’re looking for habitable planets, it’s clear that stars in the most stable phase of their lives are the ones to examine, burning hydrogen to produce energy. No brown dwarfs here, but the category does take in M-class stars, and the jury is out on whether such stars can support life. And they take in the huge majority of stars in the galaxy.
So if we’re talking about hydrogen burning, the Main Sequence offers up everything from hot blue stars all the way down to cool red dwarfs. End the hydrogen burning and a different phase of stellar evolution begins, producing for example the kind of white dwarf that the Sun will one day become. A paper’s context usually makes it perfectly clear which of the three takes on ‘Sun-like’ it is using, but the need to clarify the term in news releases, particularly when dealing with a wide range through F-, G- and K-class stars is evident.
All of this matters to the popular perception of what exoplanet researchers do because it wildly affects the numbers. G-class stars are thought to comprise about 7 percent of the stars in the galaxy, while K-class stars take in about 12 percent, and M-class dwarfs as high as 80 percent of the stellar population. Saying there is an Earth-class planet around every Sun-like star thus could mean ‘around 7 percent of the stars in the Milky Way.’ Or it could mean ‘around 22 percent of the stars in the Milky Way, if we mean FGK host stars.
If we included red dwarf stars, it could mean ‘around about 95 percent of the stars in the galaxy,’ excluding evolved, non-Main Sequence objects like white dwarfs, neutron stars and red giants. Everything depends upon how the terms are defined. I keep getting emails about this. My colleague Andrew Le Page is a stickler for terminology in the same way I am, with his most trenchant comments being reserved for too facile use of the term ‘habitable.’
So we’ve got to be careful in this burgeoning field. Exoplanet researchers are aware of the need to establish the meaning of ‘Sun-like’ carefully. The fact that the public’s interest in exoplanets is growing means, however, that in public utterances like news releases, scientists need to clarify what they’re talking about. It’s the same thing that makes the term ‘Earth-like’ so ambiguous. A planet as massive as the Earth? A planet that is rocky as opposed to gaseous? A planet in the habitable zone of its star? Is a planet on a wildly elliptical orbit crossing in and out of the habitable zone of an F-class host Earth-like?
Let’s watch our terms so we don’t confuse everybody who is not in the business of studying exoplanets full-time. The interface between professional journals and public venues like websites and newspapers is important because it can have effects on funding, which in today’s climate is a highly charged issue. A confused public is less likely to support studies in areas it does not understand.
Protoplanetary Disks Are Smaller Than Expected
In astronomy, the first thing you see may be the least typical. A case in point: ‘Hot Jupiters.’ A few prescient souls, among them Buzz Aldrin and John Barnes in their novel Encounter with Tiber, speculated about gas giants that survived incredibly tight orbits around their star, and when asked about this in the 1990s, Greg Matloff ran the numbers and confirmed to his surprise that there was a theoretical case for their existence. Let me quote Matloff on this:
Although I was initially very skeptical since then-standard models of solar system formation seemed to rule out such a possibility, I searched through the literature and located the appropriate equation (Jastrow and Rasool, 1965)….To my amazement, Buzz was correct. The planet’s atmosphere is stable for billions of years. Since I was at the time working as a consultant and adjunct professor, I did not challenge the existing physical paradigm by submitting my results to a mainstream journal. Since “Hot Jupiters” were discovered shortly before the novel was published, I am now credited with predicting the existence of such worlds.
51 Pegasi b made the case for ‘hot Jupiters’ convincingly in 1995, and we also know that the reason we began finding these planets was that our best methods for detecting exoplanets in the mid-1990s involved radial velocity and transit measurements, both of which are most effective when dealing with large planets in close orbits. As best I can determine, current estimates find a scant ~0.5–1% of these planets around F-, G-, or K-class stars. The rarity of hot Jupiters like TOI-4201b underlines how rare these objects are around M-dwarfs.
Observational bias can be a tricky thing. So is something of the same phenomenon occurring in our study of how planets form? Protoplanetary disks are extensively studied, some of them showing what appear to be planets in their early stages of formation. As I’ve read studies about disks like those at TW Hydrae and HL Tauri, I’ve always assumed that they were massive. In some cases they extend far beyond our Solar System’s outer reaches. But now we have a new study making a strong case that disks like these are not typical.
The work in question involved data from ALMA, the Atacama Large Millimeter/submillimeter Array. It comprises data from observations from 2023 and 2024 at resolutions of 0.030 arcsecond, along with somewhat higher resolution archival data to produce a high resolution survey of an entire star-forming region. The imaging takes in all known protoplanetary disks around young stars in the Lupus star forming region, some 400 light years from Earth.
The results indicate how much we have to learn about protoplanetary disks. The smallest disk comes to only 0.6 astronomical units in radius, which means it is smaller than the Earth’s orbit. The researchers also learned that the kind of features that point to planets in formation do not occur in smaller disks. Thus lead author Osmar M. Guerra-Alvarado (Leiden University, The Netherlands):
“These results completely change our view of what a ‘typical’ protoplanetary disc looks like. Only the brightest discs which are the easiest to observe show large-scale gaps, whereas compact discs without such substructures are actually much more common.”
Image: Images of 73 protoplanetary discs in the Lupus star forming region (two of the images contain binary stars). Only a fraction of the discs extend beyond the orbit of Neptune, when compared to our own Solar System. Most of the observed discs are small and show no structures like gaps and rings. (c) Guerra-Alvarado et al.
Adds Leiden colleague Nienke van der Marel:
“The discovery that the majority of the small discs do not show gaps, implies that the majority of stars do not host giant planets. This is consistent with what we see in exoplanet populations around full-grown stars. These observations link the disc population directly to the exoplanet population.”
The smaller protoplanetary disks are found around stars massing between 10 and 50 percent of the mass of the Sun. While disks like these do not appear optimum for the formation of gas giants, they may be fertile breeding ground for super-Earths, with most of the dust located close to the host star. Given this context, our own Solar System may have been formed from a more massive protoplanetary disk large enough to produce a Jupiter and a Saturn.
Image: This is an artist’s impression of a young star surrounded by a protoplanetary disc in which planets are forming. Using ALMA’s 15-kilometre baseline astronomers were able to make the first detailed image of a protoplanetary disc, which revealed the complex structure of the disc. Concentric rings of gas, with gaps indicating planet formation, are visible in this artist’s impression and were predicted by computer simulations. Now these structures have been observed by ALMA for the first time. Note that the planets are not shown to scale. Credit: ESO/L. Calçada.
From the paper:
The fact that we observe numerous compact disks in the Lupus region, both with and without substructures, aligns with the scenario from van der Marel & Mulders (2021), in which such disks, undergoing significant radial drift, can supply enough dust material to form multiple super-Earths (Sanchez et al. 2024). This offers a potential explanation for the origin of the exoplanet populations observed around M-stars to this day. In addition, the fact that we are still observing these disks may require something to halt dust drift and trapping particles. Instead of forming a single Jupiter or Saturn-mass planet, several super-Earths or smaller planets may be forming in the inner regions of these disks, collectively stopping the drift (Huang et al. 2024b).
So we have to look at the question of observation bias once again. As with hot Jupiters, our best methods continue to improve, and as they do, we gradually form a broader picture of the phenomena being studied. In the case of the latest ALMA work, we are delving into higher-resolution measurements that can begin to tease out information about small disks that have previously been studied only in terms of brightness as opposed to size. A ‘typical’ disk turns out to be something other than the large, bright disks first revealed by ALMA.
The paper is Guerra-Alvarado et al., “A high-resolution survey of protoplanetary discs in Lupus and the nature of compact discs,” Accepted for publication in Astronomy & Astrophyiscs (preprint).
A Multiwavelength Look at Proxima Centauri’s Flares
The problem of flares in red dwarf planetary systems is stark. With their habitable zones relatively near to the star, planets that might support life are exposed to huge outbursts of particles and radiation that can strip their atmospheres. We can see that in nearby M-dwarfs like Proxima Centauri, which is extremely active not only in visible light but also in radio and millimeter wavelengths. New work at the Atacama Large Millimeter/submillimeter Array (ALMA) digs into the millimeter-wavelength activity. The results do nothing to ease the concern that systems like this may be barren of life.
Small M-dwarf stars are a problem because they operate through convection as energy from fusion at the core is transferred to the surface. A convective structure is one in which hot material from below moves constantly upward, a process that can be likened to what we see in a boiling cauldron of water. Larger stars like the Sun show a mix of radiative transfer – photons being absorbed and reabsorbed as they make their way to the surface – and convection. That enhances M-dwarf flare activity as their plasma is twisted and rotated, producing magnetic fields that snap open only to reconnect. Powerful flares and outbursts of particles are the result.
For a world in an otherwise habitable region around the star, that spells danger. Meredith MacGregor (Johns Hopkins University), who worked with Kiana Burton on the flaring at Proxima Centauri, explains:
“Our Sun’s activity doesn’t remove Earth’s atmosphere and instead causes beautiful auroras, because we have a thick atmosphere and a strong magnetic field to protect our planet. But Proxima Centauri’s flares are much more powerful, and we know it has rocky planets in the habitable zone. What are these flares doing to their atmospheres? Is there such a large flux of radiation and particles that the atmosphere is getting chemically modified, or perhaps completely eroded?”
Image: Artist’s concept of a stellar flare from Proxima Centauri. Credit: NSF/AUI/NSF NRAO/S. Dagnello.
MacGregor and Burton have been working on what they describe as the first multi-wavelength study using millimeter observations to probe into the physics of these flares. At their disposal are 50 hours of ALMA observations, covering some 463 flare events at energies between 1024 to 1027 erg. Most of these flares end quickly, ranging in duration from 3 to 16 seconds. The operative term in the study is flare frequency distribution (FFD), which maps the number of flares against energy levels. A power law function as at optical wavelengths would mean that lower-energy flares would be expected to occur more frequently than flares of higher energy, but the team found many flares within each energy range because of the high flare activity at Proxima.
Adds MacGregor:
“The millimeter flaring seems to be much more frequent–it’s a different power law than we see at the optical wavelengths. So if we only look in optical wavelengths, we’re missing critical information. ALMA is the only millimeter interferometer sensitive enough for these measurements.”
The point is significant, and I want to dig into the paper on this:
Proxima Cen has been observed frequently at optical wavelengths, with a much shallower FFD power-law index of 1.88 ± 0.06. This significant difference could indicate a disconnect between sources of optical and millimeter emission during flares. Since optical observations of stellar flares are more readily available and often used to infer the flaring flux at other wavelengths, this result underlines the need for further multiwavelength campaigns to constrain scaling relations. In particular, the higher rate of millimeter flares compared to optical flares and the tight correlation between FUV and millimeter emission observed by M. A. MacGregor et al. (2021) may suggest that the extreme-UV radiation environment of Proxima b due to small flares is also higher than predicted from the optical flare rate.
So the flare activity at Proxima Centauri is more complicated and perhaps more dangerous than we thought. As we learn more about flaring at this star, we have to hope that Proxima Centauri b has a strong magnetic field that can mitigate the effects of this incoming stream of energy and particles. The prospect of an atmosphere being stripped of ozone and water, for example, makes modification or erosion of its gases a strong possibility. Instruments like the Square Kilometer Array may one day be capable of detecting the interactions between such a magnetic field and the star’s stellar wind. But for now, we can only wait for further data.
The paper is Burton et al., “The Proxima Centauri Campaign — First Constraints On Millimeter Flare Rates from ALMA,” Astrophysical Journal Vol 982, Number 1 (17 March 2025), 43. Preprint / Abstract.
Can an Interstellar Generation Ship Maintain a Population on a 250-year Trip to a Habitable Exoplanet?
I grew up on generation ship stories. I’m thinking of tales like Heinlein’s Orphans of the Sky, Aldiss’ Non-Stop, and (from my collection of old science fiction magazines) Don Wilcox’s “The Voyage that Lasted 600 Years.” The latter, from 1940, may have been the first generation ship story ever written. The idea grows out of the realization that travel to the stars may take centuries, even millennia, and that one way to envision it is to take large crews who live out their lives on the journey, their descendants becoming the ones who will walk on a new world. The problems are immense, and as Alex Tolley reminds us in today’s essay, we may not be fully considering some of the most obvious issues, especially closed life support systems. Project Hyperion is a game attempt to design a generation ship, zestfully tangling with the issues involved. The Initiative for Interstellar Studies is pushing the limits with this one. Read on.
by Alex Tolley
“Space,” it says, “is big. Really big. You just won’t believe how vastly, hugely, mindbogglingly big it is. I mean, you may think it’s a long way down the road to the chemist, but that’s just peanuts to space..” The Hitchhikers Guide to the Galaxy – Douglas Adams.
Detail of book cover of Tau Zero. Artist: Manchu.
Introduction
Science fiction stories, and most notably TV and movies, avoid the fact of the physical universe being so vast by various methods to reduce travel time. Hyperspace (e.g. Asimov’s galaxy novels), warp speed (Star Trek), and wormholes compress travel time to be like that on a planet in the modern era. Relativity will compress time for people at the cost of external time (e.g. Poul Anderson’s Tau Zero). But the energy cost of high-speed travel has a preferred slow-speed method – the space ark. Giancarlo Genta classifies this method of human crewed travel: H1 [1]. a space-ark type, where travel to the stars will take centuries. Either the crew will be preserved with cryosleep on the journey (e.g. Star Trek TOS S01E22, “Space Seed”) or generations will live and die in their ship (e.g. Aldiss Non-Stop) [2].
This last is now the focus of the Project Hyperion competition by the Initiative for Interstellar Studies (i4is) where teams are invited to design a generation ship within various constraints. This is similar to the Mars Society’s Design a City State competition for a self-supporting Mars city for 1 million people, with prizes awarded to the winners at the society’s conference.
Prior design work for an interstellar ship was carried out between 1973 and 1978, by the British Interplanetary Society (BIS). It was for an uncrewed probe to fly by Barnard’s star 5.9 lightyears distant within 50 years – their Project Daedalus.[3] Their more recent attempt at a redesign, the ironically named Project Icarus, failed to achieve a complete design, although there was progress on some technologies [4] Project Hyperion is far more ambitious based on the project constraints including human crews and a greater flight duration .[5].
So what are the constraints or boundary conditions for the competition design? Seven are given:
1. The mission duration is 250 years. In a generation ship that means about 10 generations. [Modern people can barely understand what it was like to live a quarter of a millennium ago, yet the ship must maintain an educated crew that will maintain the ship in working order over that time – AT].
2. The destination is a rocky planet that will have been prepared for colonization by an earlier [robotic? – AT] ship or directed panspermia. Conditions will not require any biological modifications of the crew.
3. The habitat section will provide 1g by rotation.
4. The atmosphere must be similar to Earth’s. [Presumably, gas ratios and partial pressures must be similar too. There does not appear to be any restriction on altitude, so presumably, the atmospheric pressure on the Tibetan plateau is acceptable. – AT]
5. The ship must provide radiation protection from galactic cosmic rays (GCR).
6. The ship must provide protection from impacts.
7. The crew or passengers will number 1000 +/- 500.
The entering team must have at least:
- One architectural designer
- One engineer
- One social scientist (sociologist, anthropologist, etc.)
Designing such a ship is not trivial, especially as unlike a Lunar or Martian city, there is no rescue possible for a lone interstellar vehicle traveling for a quarter of a millennium at a speed of at least 1.5% of lightspeed to Proxima, faster for more distant stars. If the internal life support system fails and cannot be repaired, it is curtains for the crew. As the requirements are that at least 500 people arrive to live on the destination planet, any fewer survivors, perhaps indulging in cannibalism (c.f. Neal Stephenson’s Seveneves), means this design would be a failure.
Figure 1. Technology Readiness Levels. Source NASA.
The design competition constraints allow for technologies at the technology readiness level of 2 (TRL 2) which is in the early R&D stage. In other words, the technologies are unproven and may not prove workable. Therefore the designers can flex their imaginations somewhat, from propulsion (onboard or external) to shielding (mass and/or magnetic), to the all-important life support system.
Obviously, the life support system is critical. After Gerry O’Neil designed his space colonies with interiors that looked rather like semi-rural places on Earth, it was apparent to some that if these environments were stable, like managed, miniature biospheres, then they could become generation ships with added propulsion. [6]. The Hyperion project’s 500-1500 people constraint is a scaled-down version of such a ship. But given that life support is so critical, why are the teams not required to have members with biological expertise in the design? Moreover, the design evaluators also have no such member. Why would that be?
Firstly, perhaps such an Environmental Control and. Life Support System (ECLSS) is not anticipated to be a biosphere:
The requirements matrix states:
ECLSS: The habitat shall provide environmental control and life support: How are essential physical needs of the population provided? Food, water, air, waste recycling. How far is closure ensured?
Ecosystem: The ecosystem in which humans are living shall be defined at different levels: animals, plants, microbiomes.
This implies that there is no requirement for a fully natural, self-regulating, stable, biosphere, but rather an engineered solution that mixes technology with biology. The image gallery on the Hyperion website [5] seems to suggest that the ECLSS will be more technological in nature, with plants as decorative elements like that of a business lobby or shopping mall, providing the anticipated need for reminders of Earth and possibly an agricultural facility.
Another possibility is that prior work is now considered sufficient to allow it to be grafted into the design. It seems the the problem of ECLSS for interstellar worldships is almost taken for granted despite the very great number of unsolved problems.
As Soilleux [7] states
“To date, most of the thinking about interstellar worldships has focused, not unreasonably, on the still unsolved problems of developing suitably large and powerful propulsion systems. In contrast, far less effort has been given to the equally essential life support systems which are often assumed to be relatively straightforward to construct. In reality, the seductive assumption that it is possible to build an artificial ecosystem capable of sustaining a significant population for generations is beset with formidable obstacles.”
It should be noted that no actual ECLSS has proven to work for any length of time even for interplanetary flight, let alone centuries for interstellar flight. Gerry O’Neill did not make much effort beyond handwaving for the stability of the more biospheric ECLSS of his 1970s-era space colony. Prior work from Bios 3, Biosphere II, MeLISSA, and other experiments, has demonstrated very short-term recycling and sustainability, geared more towards longer duration interplanetary spaceflight and bases. Multi-year biospheres inevitably lose material due to imperfect recycling and life support imbalances that must be corrected via venting. The well-known Biosphere II project couldn’t maintain a stable biosphere to support an 8-person crew for 2 years, accompanied by a 10% per year atmosphere loss.
On a paper design basis, the British Interplanetary Society (BIS) recently worked on a more detailed design for a post-O’Neill space colony called Avalon. It supported a living interior for 10,000 people, the same as the O’Neill Island One design, providing a 1g level and separate food growing, carbon fixing, and oxygen (O2) generating areas. However, the authors did not suggest it would be suitable for interstellar flight as it did require some inputs and technological support from Earth. Therefore it remains an idea, and as with the hardware experiments to date, such ECLSSs remain at low TRL values. While the BIS articles are behind paywalls, there is a very nice review of the project in the i4is Principium magazine [7].
Table 1. Inputs to support a person in Space. Source: Second International Workshop on Closed Ecological Systems, Krasnoyarsk, Siberia, 1989 [8]
Baseline data indicates that a person requires 2 metric tonnes [MT] of consumables per year, of which food comprises 0.2 MT and oxygen (O2) 0.3 MT. For 1000 people over 250 years that is 500,000 MT. To put this mass in context, it is about the greatest mass of cargo that can be carried by ships today. Hence recycling to both save mass and to manage potentially unlimited time periods makes sense. Only very short-duration trips away from Earth allow for the consumables to be brought along and discarded after use, like the Apollo missions to the Moon. Orbiting space stations are regularly resupplied with consumables, although the ISS is using water recycling to save on the resupply of these consumables for drinking, washing, and bathing.
If an ECLSS could manage 100% recycling, then if the annual amount is the buffer to allow for annual crop growth, the ship could support the crew over the time period with just over 2000 MT, but of course with the added requirement for power and infrastructure and replacement parts to maintain the ECLSS.
Figure 2. Conceptual ECLSS process flow diagram for the BIS Avalon project.
A review of the Avalon project’s ECLSS [7] stated:
“A fully closed ECLSS is probably impossible. At a minimum, hydrogen, oxygen, nitrogen, and carbon lost from airlocks must be replaced as well as any nutrients lost to the ecosystem as products (eg soil, wood, fibres, chemicals). This presents a resource-management challenge for any fully autonomous spacecraft and an economic challenge even if only partially autonomous…Recycling does not provide a primary source of fixed nitrogen for new plant growth and significant quantities are lost to the air as N2 because of uncontrollable denitrification processes. The total loss per day for Avalon has been estimated at 43 kg to be replaced from the atmosphere.”
[That last loss figure would translate to 390 MT for the 1/10 scaled crewed Hyperion starship over 250 years.]
Hempsell evaluated the ECLSS technologies for star flight based on the Avalon project and found them unsuited to a long-duration voyage [9]. It was noted that the Avalon design still required inputs from Earth to be maintained as it was not fully recycled:
“So effective is this transport link that it is unlikely Earth orbiting colonies will have any need for the self-sufficiency assumed by the original thinking by O’Neill. It is argued that a colony’s ability to operate without support is likely to be comparable to the transport infrastructure’s journey times, which is three to four months. This is three orders of magnitude less than the requirements for a world ship. It is therefore concluded that in many critical respects the gap between colonies and world ships is much larger than previous work assumed.”
We can argue whether this will be solved eventually, but clearly, any failure is lethal for the crew.
One issue is not mentioned, despite the journey duration. Over the quarter millennium voyage, there will be evolution as the organisms adapt to the ship’s environment. Data from the ISS has shown that bacteria may mutate into more virulent pathogens. A population living in close quarters will encourage pandemics. Ionizing radiation from the sun and secondaries from the hull of a structure damages cells including their DNA. 250 years of exposure to residual GCR and secondaries will damage DNA of all life on the starship.
However, even without this direct effect on DNA, the conditions will result in organisms evolving as they adapt to the conditions on the starship, especially the small populations, increasing genetic drift. This evolution, even of complex life, can be quite fast, as the continuing monitoring of the Galápagos island finches observed by Darwin attests. Of particular concern is the creation of pathogens that will impact both humans and the food supply.
In the 1970s, the concept of a microbiome in humans, animals, and some plants was unknown, although bacteria were part of nutrient cycling. Now we know much more about the need for humans to maintain a microbiome, as well as some food crops. This could become a source of pathogens. While a space habitat can just flush out the agricultural infrastructure and replace it, no such possibility exists for the starship. Crops would need to be kept in isolated compartments to prevent a disease outbreak from destroying all the crops in the ECLSS.
If all this wasn’t difficult enough, the competition asks that the target generation population find a ready-made terrestrial habitat/terraformed environment to slip into on arrival. This presumably was prebuilt by a robotic system that arrived ahead of the crewed starship to build the infrastructure and create the environment ready for the human crew. It is the Mars agricultural problem writ large [10], with no supervision from humans to correct mistakes. If robots could do this on an exoplanet, couldn’t they make terrestrial habitats throughout the solar system?
Heppenheimer assumed that generation ships based on O’Neill habitats would not colonize planets, but rather use the resources of distant stars to build more space habitats, the environment the populations had adapted to [6]. This would maintain the logic of the idea of space habitats being the best place for humanity as well as avoiding all the problems of trying to adapt to living on a living, alien planet. Rather than building space habitats in other star systems, the Hyperion Project relies on the older idea of settling planets, and ensuring at least part of the planet is habitable by the human crew on arrival. This requires some other means of constructing an inhabitable home, whether a limited surface city or terraforming the planet.
Perhaps the main reason why both the teams and evaluators have put more emphasis on the design of the starship as a human social environment is that little work has looked at the problems of maintaining a small population in an artificial environment over 250 years. SciFi novels of generation ships tend to be dystopias. The population fails to maintain its technical skills and falls back to a simpler, almost pretechnological, lifestyle. However, this cannot be allowed in a ship that is not fully self-repairing as hardware will fail. Apart from simple technologies, no Industrial Revolution technology that continues to work has operated without the repair of parts. Our largely solid-state space probes like the Voyagers, now over 45 years old and in interstellar space, would not work reliably for 250 years, even if their failing RTGs were replaced. Mechanical moving parts fail even faster.
While thinking about how we might manage to send crewed colony ships to the stars with the envisaged projections of technologies that may be developed, it seems to me that it is rather premature given the likely start date of such a mission. Our technology will be very different even within decades, obsoleting ideas used in the design. The crews may not be human 1.0, requiring very different resources for the journey. At best, the design ideas for the technology may be like Leonardo Da Vinci’s ideas for flying machines. Da Vinci could not conceive of the technologies we use to avoid moving our meat bodies through space just to communicate with others around the planet.
Why is the population constraint 1000 +/- 500?
Gerry O’Neill’s Island One space colony was designed for 10,000 people, a small town. The BIS Avalon was designed for the same number, using a different configuration that provided the full 1g with lower Coriolis forces. The Hyperion starship constrains the population to 1/10th that number. What is the reason? It is based on the paper by Smith [14] on the genetics and estimated safe population size for a 5-generation starship [125 years at 25 years per generation?]. This was calculated on the expected lower limit of a small population’s genetic diversity with current social structures, potential diseases, etc. This is well below the number generally considered for long-term viable animal populations. However, the calculated size of the genetic bottleneck in our human past is about that size, suggesting that in extremis such a small population can recover, just as we escaped extinction. Therefore it should be sufficient for a multi-generation interstellar journey.
From Hein[14]:
While the smallest figures may work biologically, they are rather precarious for some generations before the population has been allowed to grow. We therefore currently suggest figures with Earth-departing (D1) figures on the order of 1,000 persons.
But do we need this population size? Without resorting to a full seed ship approach with the attendant issues of machine carers raising “tank babies”, can we postulate an approach without assuming future TRL9 ECLSS technology?
As James T Kirk might say: “I changed the conditions of the test competition!”
Suppose we only need a single woman who will bear a child and then age until death, perhaps before 100. The population then would be a newborn baby, the 25-year-old mother, the 50-year-old grandmother, the 75-year-old great-grandmother, and the deceased great-great-grandmother. 3 adults and a newborn. At 20, the child will count as an adult, with a 45-year-old mother, a 70-year-old grandmother, and possibly a 95-year-old great-grandmother, 4 adults. Genetic diversity would be solved from a frozen egg, sperm, and embryo bank of perhaps millions of individual selected genomes. The mother would give birth only to preselected, compatible females using implantation (or only female sperm) to gain pregnancy, to maintain the line of women.
This can easily be done today with minimal genetic engineering and a cell-sorting machine [11]. At the destination, population renewal can begin. The youngest mature woman could bear 10 infants, each separated by 2 years. The fertile female population would rapidly increase 10-fold each generation, creating a million women in 6 generations, about 150 years. Each child would be selected randomly from the stored genetic bank of female sperm or embryos. At some point, males can be phased in to restore the 50:50 ratio. Because we are only supporting 4 women, almost all the recycling for food and air can be discarded, just recycling water. Water recycling for urine and sweat has reached 98% on the ISS, so we can assume that this can be extended to the total water consumption. Using a conservative 98% recycle rate for water (sanitary, drinking, and metabolic), O2 replenishment from water by electrolysis (off-the-shelf kit) and an 80% recycle rate of metabolic carbon dioxide (CO2) to O2 and methane (CH4 by a Sabatier process), implies that the total water storage needs only be 5x the annual consumption. The water would be part of the hull as described by McConnell’s “Spacecoach” concept.
Based on the data in Table 1 above, the total life support would be about 125 MT. The methane could be vented or stored for the lander propulsion fuel. For safety through redundancy, the total number of women population might be raised 10x to 30-40, requiring only 1250 MT total consumables. All this does not require any new ECLSS technology, just enough self-repair, repairable, and recyclable machinery, to maintain the ship’s integrity and functioning. The food would be stored both as frozen or freeze-dried, with the low ambient temperature maintained by radiators exposed to interstellar space. This low mass requires no special nutrient recycling, just water recycling at the efficiency of current technology, O2 recycling using existing technology, and sufficient technology support to ensure any repairs and medical requirements can be handled by the small crew.
Neal Stephenson must have had something like this in mind for the early period of saving and rebuilding the human population in his novel Seveneves when the surviving female population in space was reduced to 7. Having said this, such a low population size even with carefully managed genetic diversity through sperm/ar embryo banks, genetic failure over the 10 generation flight may still occur. Once the population has the resources to increase, the genetic issues may be overcome by the proposed approach.
Table 2. ETHNOPOP simulation showing years to demographic extinction for closed human populations. Bands of people survive longer with larger starting sizes, but these closed populations all eventually become extinct due to demographic (age & sex structure) deficiencies. Source Hein [14]
From table 2 above, even a tiny, natural,human population may survive hundreds of years before extinction. Would the carefully managed starship population fare better, or worse?
Despite this maneuver, there is still a lot of handwavium about the creation of a terrestrial habitable environment at the destination. All the problems of building an ECLSS or biosphere in a zone on the planet or its entirety by terraforming, entirely by a non-human approach, are left to the imagination. As with humans, animals such as birds and mammals require parental nurturing, making any robotic seedship only partially viable for terraforming. If panspermia is the expected technology, this crewed journey may not start for many thousands of years.
But what about social stability?
I do applaud the Hyperion team for requiring a plan for maintaining a stable population for 250 years. The issue of damage to a space habitat in the O’Neill era was somewhat addressed by proponents, mainly through the scale of the habitat and repair times. Major malicious damage by groups was not addressed. There is no period in history that we know of that lasted this long without major wars. The US alone had a devastating Civil War just 164 years ago. The post-WWII peace in Europe guaranteed by the Pax Americana is perhaps the longest period of peace in Europe, albeit its guarantor nation was involved in serious wars in SE Asia, and the Middle East during this period. Humans seem disposed to violence between groups, even within families. I couldn’t say whether even a family of women would refrain from existential violence. Perhaps genetic modification or chemical inhibition in the water supply may be solutions to outbreaks of violence. This is speculation in a subject outside my field of expertise. I hope that the competition discovers a viable solution.
In summary, while it is worthwhile considering how we might tackle human star flight of such duration, there is no evidence that we are even close to solving the issues inherent in maintaining any sort of human star flight of such duration with the earlier-stage technologies and the constraints imposed on the designs by the competition. By the time we can contemplate such flights, technological change will likely have made all the constraints superfluous.
References and Further Reading
Genta, G Genta, “Interstellar Exploration: From Science Fiction to Actual Technology,”” Acta Astronautica Vol. 222 (September 2024), pp. 655-660 https://www.sciencedirect.com/science/article/pii/S0094576524003655?via%3Dihub
Gilster P, (2024) “Science Fiction and the Interstellar Imagination,” Web https://www.centauri-dreams.org/2024/07/17/science-fiction-and-the-interstellar-imagination/
BIS “Project Daedalus” https://en.wikipedia.org/wiki/Project_Daedalus
BIS “Project Icarus” https://en.wikipedia.org/wiki/Project_Icarus_(interstellar)
i4is “Project Hyperion Design Competition” https://www.projecthyperion.org/
Heppenheimer, T. A. (1977). Colonies in space: A Comprehensive and Factual Account of the Prospects for Human Colonization of Space. Harrisburg, PA: Stackpole Books, 1977
Soilleux, R. (2020) “Implications for an Interstellar World-Ship in findings from the BIS SPACE Project” – Principium #30 pp5-15 https://i4is.org/wp-content/uploads/2020/08/Principium30-print-2008311001opt.pdf
Nelson, M., Pechurkin, N. S., Allen, J. P., Somova, L. A., & Gitelson, J. I. (2010). “Closed ecological systems, space life support and biospherics.” In Humana Press eBooks (pp. 517–565). https://doi.org/10.1007/978-1-60327-140-0_11
Hempsell, M (2020), “Colonies and World Ships” JBIS, 73, pp.28-36 (abstract) https://www.researchgate.net/publication/377407466_Colonies_and_World_Ships
Tolley, A. (2023) “MaRMIE: The Martian Regolith Microbiome Inoculation Experiment.” Web https://www.centauri-dreams.org/?s=Marmie
Bio-Rad “S3e Cell Sorter” https://www.bio-rad.com/en-uk/category/cell-sorters?ID=OK1GR1KSY
Allen, J., Nelson, M., & Alling, A. (2003). “The legacy of biosphere 2 for the study of biospherics and closed ecological systems.” Advances in Space Research, 31(7), 1629–1639. https://doi.org/10.1016/s0273-1177(03)00103-0
Smith, C. M. (2014). “Estimation of a genetically viable population for multigenerational interstellar voyaging: Review and data for project Hyperion.” Acta Astronautica, 97, 16–29. https://www.sciencedirect.com/science/article/abs/pii/S0094576513004669
Hein A.M,, et al (2020). “World ships: Feasibility and Rationale” Acta Futura, 12, 75-104, 2020 Web https://arxiv.org/abs/2005.04100