
- Key Takeaways
- The Scale of the Problem
- What "AI" Actually Means in an Observatory
- Exoplanet Discovery and the Neural Network That Confirmed Two Planets
- Gravitational Waves and the Speed of Detection
- Radio Astronomy and Fast Radio Bursts
- Seeing the Unseeable: The Event Horizon Telescope
- AI Applications Across Astronomy
- Galaxy Classification at Scale
- Simulating the Universe
- The Search for Extraterrestrial Intelligence
- The James Webb Space Telescope and Spectral Analysis
- AI-Driven Telescope Operations
- New Institutional Investment
- What AI Gets Wrong
- The Changing Role of the Astronomer
- Near-Earth Objects and Planetary Defense
- Cosmological Parameter Estimation
- Future Missions and Embedded AI
- Summary
- Appendix: Top 10 Questions Answered in This Article
Key Takeaways
- AI now processes telescope data at scales no human team could ever replicate manually
- Neural networks have confirmed exoplanets, classified billions of galaxies, and detected gravitational waves in milliseconds
- AI is reshaping which scientific questions are worth asking, not just how fast they’re answered
The Scale of the Problem
Roughly 2,000 photographic glass plates sit in storage at the Palomar Observatory in California, some exposed in the 1950s and still not fully analyzed. This wasn’t negligence. It was arithmetic. There were never enough trained eyes to work through everything the telescopes captured.
That problem has grown several orders of magnitude more complex since then. The Vera C. Rubin Observatory in Chile released its first full-system images in June 2025, using a 3,200-megapixel camera – the largest digital camera ever built. When its Legacy Survey of Space and Time begins full science operations in early 2026, it will generate approximately 20 terabytes of raw data every observing night and discover an estimated 2,000 new supernova explosions every single night. To put that in context, every telescope on Earth currently combined finds about 40,000 supernovae per year. Rubin will surpass that in roughly three weeks, every three weeks, for a decade.
No team of human astronomers can process that. The choice isn’t between human analysis and automated analysis. At this scale, it’s between artificial intelligence and leaving most of the data unexamined.
This reality has pushed AI from an experimental tool at the edges of astrophysics into the operational center of the discipline. The shift has been rapid, most of the significant applications appearing since 2015, and it’s still accelerating. The SkAI Institute at Northwestern University has been developing new AI models specifically in support of astronomical surveys, and its director Adam Miller has stated plainly that there is absolutely no way any research team could look at the tens of billions of sources that Rubin will regularly monitor.
What “AI” Actually Means in an Observatory
The term “artificial intelligence” gets used loosely, and that looseness matters when trying to understand what’s actually changing in astronomy. Most of the AI currently reshaping the field is specifically machine learning – and within that, a subset called deep learning, which uses layered neural networks loosely inspired by the structure of the human brain.
These networks don’t follow explicit rules written by programmers. They’re trained on labeled examples. Show a neural network a million galaxy images and tell it which ones are spirals, ellipticals, or merging pairs, and it will eventually learn the visual patterns that distinguish them. After training, it can classify a new, unseen galaxy in milliseconds – and keep doing so without fatigue, indefinitely.
This matters for astronomy because the field is now flooded with exactly the kinds of pattern-recognition tasks that neural networks handle well: identifying object types in images, detecting periodic signals in time-series data, flagging anomalies that differ from expected behavior, and matching observed spectra against libraries of known chemical signatures. The power isn’t that AI is smarter than a trained astronomer. It’s that AI is far faster, can run continuously, and can be applied to problems where scale makes human analysis impractical.
A 2025 study published in Nature Astronomy, co-led by researchers from the University of Oxford and Google Cloud, illustrated just how far this has come. The team demonstrated that Google’s Gemini large language model could classify real astronomical transient events – exploding stars, tidal disruption events, fast-moving asteroids – with approximately 93% accuracy using just 15 training images and a plain-English instruction prompt. By using a self-correction loop to refine its initial examples, the system improved its classification performance from 93.4% to 96.7% on one dataset, showing how AI can learn and improve in partnership with human experts. That’s not a specialist model trained for months on millions of labeled images. That’s a general-purpose AI given fifteen examples and told what to look for.
Exoplanet Discovery and the Neural Network That Confirmed Two Planets
The Kepler Space Telescope spent nine years staring at roughly 150,000 stars, watching for the tiny dip in starlight that signals a planet passing across its host star’s face. By the time Kepler ran out of fuel in 2018, it had collected such an enormous volume of light curves that astronomers were still working through the data years after the mission ended.
In 2017, researchers at Google, working with astronomers at the SETI Institute and NASA, trained a convolutional neural network on previously classified Kepler signals. The network learned to distinguish genuine planetary transits from false positives caused by eclipsing binary stars, instrument artifacts, and other noise sources. Applied to a set of 670 star systems already known to host multiple planets – places where additional planets were considered statistically more likely – the neural network identified two previously missed candidates. Both were confirmed as genuine planets: Kepler-90i and Kepler-80g.
Kepler-90i gave the Kepler-90 system eight confirmed planets, tying our solar system for the most known planets around any single star at that time. That discovery came not from new observations but from reanalyzing data that already existed, overlooked because the signal was too subtle for earlier detection pipelines to catch. It’s a useful illustration of what AI-based reanalysis means in practice: the discovery was sitting in the archive for years, waiting for a tool capable of finding it.
TESS, Kepler’s successor mission, is producing even larger datasets, and AI-based detection pipelines are now embedded directly into how its data gets processed. MIT researchers developed the AstroNet-K2 network and later adapted it for TESS data, and similar tools are now standard across most planet-hunting pipelines. The expectation, well grounded in what’s already happened, is that many of the best exoplanet discoveries in the coming decade will come from AI reanalysis of archival data rather than from new observations alone.
Gravitational Waves and the Speed of Detection
When the LIGO detectors in Hanford, Washington and Livingston, Louisiana recorded the first confirmed gravitational wave signal in September 2015 – the event designated GW150914, produced by two black holes merging 1.3 billion light-years away – the detection relied on a matched-filter technique that compared incoming signals against a bank of precomputed waveform templates. It worked. It also had significant limitations: computationally expensive, it struggled with certain signal types and could only identify events whose shapes resembled templates already in the library.
Neural networks have changed that. A 2018 paper by a group at the California Institute of Technology demonstrated that a deep learning model could analyze LIGO data and detect gravitational wave signals in approximately one millisecond – orders of magnitude faster than traditional matched-filter methods – while maintaining comparable accuracy. Speed matters here because gravitational wave detections trigger rapid follow-up observations by telescopes around the world, and the shorter the delay, the better the chance of catching the optical counterpart before it fades.
The detection of GW170817 in August 2017, the first gravitational wave signal from two merging neutron stars, produced an electromagnetic counterpart observed by dozens of telescopes in the hours and days that followed. That event demonstrated why rapid classification matters. As LIGO and its partner detector Virgo grow more sensitive, and as new detectors come online in Japan and India, the expected rate of detections will increase substantially. Neural network-based classifiers are now integrated into the alert pipelines, helping astronomers prioritize which events deserve the fastest follow-up response.
There’s also a more speculative possibility. Some researchers believe AI may eventually detect gravitational wave signals that don’t match any known template – potentially signals from exotic objects or processes that physicists haven’t yet modeled. Whether that happens remains unclear. Neural networks can, in principle, find patterns that no human thought to look for, and that’s a genuinely different kind of scientific tool than anything available before.
Radio Astronomy and Fast Radio Bursts
Fast radio bursts are among the most puzzling phenomena in modern astrophysics. They’re intense millisecond-duration pulses of radio emission arriving from cosmological distances, and their origin remains incompletely understood despite years of study. The first confirmed fast radio burst was identified in 2007 in archival data from the Parkes Observatory in Australia, and for years afterward only a handful were known.
That changed when the Canadian Hydrogen Intensity Mapping Experiment – CHIME – came online at the Dominion Radio Astrophysical Observatory in British Columbia. CHIME doesn’t look like a traditional radio telescope. It has no moving parts and consists of four enormous cylindrical reflectors, each 100 meters long, oriented north-to-south. It surveys the entire northern sky every day and generates data volumes that can only be handled with real-time AI processing.
By 2023, the CHIME team had published a catalog of over 500 fast radio bursts – a number that would have been unthinkable a decade earlier. That catalog was built on an automated detection pipeline using machine learning at every stage, from the initial identification of candidate signals to the classification of burst morphology to the removal of terrestrial radio interference.
The Breakthrough Listen program, in partnership with NVIDIA, recently deployed an AI system on the Allen Telescope Array in California that represents a step change in processing speed. The previous state-of-the-art pipeline at the Allen Telescope Array required approximately 59 seconds to process 16.3 seconds of observational data – running almost four times slower than real-time. The new end-to-end AI system processes the same data 600 times faster, enabling it to operate more than 160 times faster than real-time constraints. That’s not an incremental improvement. It means the telescope can analyze signals as they arrive and immediately flag candidates for follow-up, rather than storing data and processing it hours later.
The Square Kilometre Array Observatory – the SKA – will take this further. Under construction across sites in South Africa and Australia, the SKA is projected to begin science operations in the late 2020s. When completed, it will be the largest radio telescope ever built, with collecting area spread across thousands of individual antennas. Its data rates are expected to exceed a terabyte per second during peak operations. AI-based data reduction isn’t an add-on; it’s fundamental to the design.
Seeing the Unseeable: The Event Horizon Telescope
The image of the supermassive black hole at the center of the galaxy Messier 87, released in April 2019, was the first direct image of a black hole’s shadow. Getting there required collecting data from eight radio telescope facilities across four continents, coordinated to act as a single Earth-sized instrument through a technique called very long baseline interferometry. It also required reconstructing an image from sparse and incomplete data using algorithms that drew heavily on machine learning principles.
The Event Horizon Telescope collaboration used multiple independent imaging methods to ensure the result wasn’t an artifact of any single algorithm. One method, called SMILI, used sparse modeling techniques borrowed from compressed sensing to reconstruct the image from raw visibility data. The result – a bright ring of emission surrounding a central dark region – was consistent across all methods, giving the team confidence in its validity.
The second black hole the EHT imaged, Sagittarius A* at the center of the Milky Way, was harder. The material around Sgr A* orbits the black hole in minutes rather than thousands of years, meaning the image was blurring and shifting during the hours-long observations. Reconstruction required AI-based averaging and interpolation techniques to handle that time variability. The result, released in May 2022, showed the same ring structure seen in M87 – a consistency that would have been much harder to achieve without AI-based reconstruction methods.
AI Applications Across Astronomy
| Astronomy Area | AI Method | Key Example | Outcome |
|---|---|---|---|
| Exoplanet Detection | Convolutional Neural Network | Kepler-90i (NASA/Google, 2017) | Discovered planets missed by classical pipelines |
| Gravitational Waves | Deep Learning Classifier | LIGO / Caltech (2018) | Millisecond event detection vs. minutes |
| Radio Transients | Real-Time ML Pipeline | CHIME FRB Catalog 2023 | 500+ bursts catalogued automatically |
| Black Hole Imaging | Sparse Modeling / Compressed Sensing | Event Horizon Telescope (2019, 2022) | Images reconstructed from incomplete data |
| Galaxy Morphology | Convolutional Neural Network | Galaxy Zoo / Rubin Observatory pipeline | Billions of galaxies classified at speed |
| Cosmological Simulation | Neural Emulators / Generative Models | CAMELS Project / IllustrisTNG | Simulation speed increased 10,000-fold |
| SETI Signal Analysis | Anomaly Detection / Deep Learning | Breakthrough Listen / Allen Telescope Array | 600x speed gain, 10x fewer false positives |
| Transient Classification | Large Language Model (Few-Shot) | Oxford / Google Cloud Gemini study (2025) | 93-96.7% accuracy from 15 training images |
Galaxy Classification at Scale
In 2007, a project called Galaxy Zoo launched with a straightforward premise: show members of the public images of galaxies from the Sloan Digital Sky Survey and ask them to classify the shapes. Within 24 hours of launch, the site was receiving 70,000 classifications per hour. Within a year, more than 150,000 volunteers had classified nearly a million galaxies through the Zooniverse citizen science platform.
It worked remarkably well, and it revealed the limits of human classification at scale at the same time. Getting five independent human classifications for every galaxy in the Rubin Observatory’s eventual catalog – roughly 20 billion objects – would require something like 100 billion individual human judgments. That’s not achievable, not at any realistic rate of volunteer participation.
Neural networks trained on Galaxy Zoo’s human-generated labels have largely solved this problem. A convolutional neural network trained on those morphological labels can replicate human judgments at a rate of thousands of galaxies per second. The model isn’t perfect – merging galaxies, edge-on spirals, and objects with unusual features can fool any automated system – but for the majority of standard galaxy types, neural network classifiers now perform at a level comparable to trained human astronomers.
This matters beyond counting shapes. Galaxy morphology reflects physics: how a galaxy looks tells scientists about its star formation history, its interactions with neighbors, and the dark matter halo it lives in. Classifying galaxy shapes at the scale of billions of objects opens statistical studies that would otherwise be unreachable. The questions become possible only because the classification becomes possible.
The current version of Galaxy Zoo, Galaxy Zoo DECaLS, uses a hybrid approach: a neural network pre-classifies galaxies, routing the most uncertain cases – objects near the decision boundaries between categories – preferentially to human classifiers. Human attention becomes concentrated at exactly the cases where automated systems are least reliable. A global team at over a dozen institutions released the Multimodal Universe dataset in 2024, a 100-terabyte collection bringing together hundreds of millions of astronomical observations specifically formatted for machine learning research, designed to accelerate new developments in both astronomy and machine learning.
Simulating the Universe
Cosmological simulations are among the most computationally expensive activities in science. A modern simulation like IllustrisTNG, which models the formation and evolution of galaxies across cosmic time using gravity, hydrodynamics, star formation, and black hole feedback, requires millions of CPU hours to run. Adjusting parameters and re-running to test a different model of dark matter or a different prescription for black hole feedback means spending millions more.
This creates a bottleneck. Astronomers want to compare many different theoretical models against observations, but each comparison requires a full simulation, and simulations are slow. Testing thousands of variations is practically out of reach.
Neural network emulators offer a way around this. The basic idea is to train a network on the outputs of a smaller set of full simulations, then use it to predict what a new simulation would produce without running it. The CAMELS project – Cosmology and Astrophysics with Machine Learning Simulations – built a database of over 4,000 IllustrisTNG and SIMBA simulations with systematically varied parameters, specifically designed to train emulators of exactly this kind. An emulator trained on CAMELS can predict the large-scale structure of a simulated universe in seconds rather than hours, and can do so with enough accuracy that the resulting statistics match full simulations within acceptable error margins.
Researchers at Carnegie Mellon University have used this approach to infer cosmological parameters directly from galaxy survey data. If a neural network has been trained on thousands of simulations spanning a range of parameters, it can look at an observed galaxy survey and estimate which combination of parameters best matches what’s observed – without the computationally expensive Monte Carlo sampling that traditional methods require.
The Search for Extraterrestrial Intelligence
The SETI Institute and the Breakthrough Listen program at the University of California, Berkeley are generating radio and optical survey data at rates that dwarf everything that came before. Breakthrough Listen, funded by investor Yuri Milner with a $100 million ten-year commitment announced in 2015, has been systematically observing one million nearby stars and a hundred nearby galaxies.
The challenge is that nobody knows exactly what an artificial signal from another civilization would look like. Classic SETI searches looked for narrowband radio signals – the assumption being that a technological civilization would transmit in a deliberately narrow frequency range to make itself distinguishable from natural sources. But this assumption may be wrong. A civilization using spread-spectrum communication, for instance, would produce signals that look like noise by design.
Machine learning approaches, particularly unsupervised methods that look for anomalies rather than specific signal types, open up the search in ways that matched-filter methods can’t. In 2020, a team applying neural networks to Breakthrough Listen data from the Green Bank Telescope identified 72 previously undetected bursts from the repeating fast radio burst source FRB 121102 – signals that had already been processed by conventional methods and missed. The AI found what human-designed pipelines overlooked.
A more recent AI system deployed on the Allen Telescope Array achieves 7% better accuracy than existing pipelines while reducing false positives by nearly ten-fold, processing data over 160 times faster than real-time. Whether any detected signal will ever be traced to an artificial origin is genuinely beyond what current evidence can say. AI doesn’t change that fundamental uncertainty. What it changes is the fraction of the parameter space that can be examined at all, and that fraction was previously very small.
The James Webb Space Telescope and Spectral Analysis
The James Webb Space Telescope launched in December 2021 and reached its operational orbit at L2 in January 2022. Its primary instruments collect infrared light at wavelengths invisible to human eyes, and the spectroscopic data it produces arrives in forms that require sophisticated analysis to interpret.
One of JWST’s key science goals is analyzing the atmospheres of exoplanets. When a planet passes in front of its star, some starlight filters through the planet’s atmosphere and picks up absorption features characteristic of specific molecules – water, carbon dioxide, methane, ozone. Identifying those features against the noise of real observations is a pattern-matching problem well suited to machine learning.
Researchers at the Space Telescope Science Institute and partner institutions have been developing AI-based spectral retrieval tools specifically for JWST data. Traditional retrieval methods use nested sampling algorithms that can take days to converge on a solution. AI-based retrievals can produce comparable results in minutes. As JWST accumulates more atmospheric targets, that speed difference becomes significant – the difference between analyzing a handful of atmospheres per year and studying dozens.
JWST’s imaging data is also powering galaxy morphology studies at high redshift, looking at galaxies as they existed when the universe was only one or two billion years old. Convolutional neural networks are classifying these distant objects and identifying merger signatures in data that would otherwise require years of visual inspection.
AI-Driven Telescope Operations
A less visible but impactful application is in how observatories actually run. Large facilities manage complex scheduling problems: balancing the scientific priorities of hundreds of approved programs against weather constraints, instrumental limitations, and the movement of targets across the sky. At facilities like the W. M. Keck Observatory in Hawaii and the European Southern Observatory in Chile, scheduling has traditionally involved a combination of automated systems and human expertise.
Reinforcement learning systems – where an AI agent learns by taking actions and receiving feedback – have been explored for telescope scheduling, and in some configurations they outperform traditional optimization algorithms. A night at an 8-meter telescope costs tens of thousands of dollars. Rare observing conditions can disappear within hours. Squeezing more science out of that time has concrete value.
Adaptive optics, the technology that corrects for atmospheric turbulence in real time and allows ground-based telescopes to approach the resolution of space-based instruments, is another area where AI is making measurable progress. Traditional adaptive optics systems use deformable mirrors controlled by algorithms that respond to wavefront sensor measurements. Neural network controllers can learn the complex, nonlinear relationships between sensor measurements and mirror commands, and in some configurations they outperform classical control algorithms, particularly under challenging atmospheric conditions.
The Subaru Telescope in Hawaii and the Very Large Telescope in Chile have both explored AI-enhanced adaptive optics. The Extremely Large Telescope currently under construction in Chile – which will have a primary mirror 39 meters across – will incorporate neural network-based control systems from the start. At that scale, with thousands of deformable mirror actuators to coordinate in real time, classical control algorithms reach the edge of what’s feasible. AI-based control isn’t optional; it’s the path forward.
New Institutional Investment
The scale of institutional investment in AI for astronomy has grown substantially. The National Science Foundation and the Simons Foundation jointly established the NSF-Simons CosmicAI Institute at the University of Texas at Austin with $20 million in funding over five years. CosmicAI will develop AI approaches to efficiently process large astronomical datasets, explore the nature of dark matter, and model prebiotic molecules that are key to life in the universe, and the institute plans to develop a powerful AI-based astronomy co-pilot to streamline the scientific method in astrophysics and assist researchers with visualizing data and producing statistical analyses.
That’s one institute among several launched in the past few years specifically targeting the intersection of AI and astronomy. The SkAI Institute at Northwestern University, the Polymathic AI collaboration spread across the Flatiron Institute and multiple universities, and a growing number of industrial partnerships between observatories and technology companies all reflect the same basic recognition: AI is no longer a peripheral tool in astronomy. It’s infrastructure.
What AI Gets Wrong
The capabilities described above are real and documented, but AI systems in astronomy also fail in ways that matter. Neural networks trained on one dataset often don’t generalize well to data from different instruments, different observing conditions, or different parts of the sky. A galaxy classifier trained on images from the Sloan Digital Sky Survey may perform poorly on JWST images, which have different resolution, depth, and noise characteristics. Retraining is necessary each time the observational context changes significantly.
There’s also a subtler problem with rare objects. Neural networks learn from examples, and if a class of object is rare in the training set – a particular type of binary star, a very young galaxy, an unusual quasar – the network will tend to misclassify it or miss it. The things that neural networks are worst at identifying are often exactly the things that are most scientifically interesting, because unusual objects are unusual for a reason.
Astronomers working with AI tools are aware of this. The standard practice is to use AI systems for bulk classification and initial filtering, then apply human expertise to the objects that emerge as uncertain, anomalous, or scientifically high-priority. The division of labor is one where AI handles enormous volumes of routine analysis, freeing human researchers to focus on interpretation and follow-up. That’s a reasonable working arrangement. Whether it stays that way as AI capabilities improve is less certain.
Large language models have shown some ability to assist with literature synthesis and hypothesis generation, but their usefulness for actual data analysis remains limited compared to specialized domain models trained on astronomical data. Where the boundary between AI assistance and AI-led discovery ends up in ten or twenty years, nobody can say with any confidence. That uncertainty is real and it should probably be stated plainly rather than papered over with optimistic projections.
The Changing Role of the Astronomer
When the Rubin Observatory begins generating millions of alerts per night, professional astronomers receiving those alerts won’t be sitting at terminals reviewing each one manually. They’ll be setting up filters, training classifiers, writing code to query databases, and studying outputs that AI systems flag as interesting. The observational skill that defined the astronomer for most of the discipline’s history – the ability to recognize a subtle feature in a photographic plate or a spectrum – is becoming less central to the job.
What’s replacing it is partly computational skill and partly interpretive skill. The ability to understand what an AI system is actually doing, where it’s likely to fail, and how to design experiments that test what the AI can’t test for itself – these are growing more valuable. So is the ability to ask questions that require AI to work at all, questions that would be meaningless to pose if the answer involved sifting through ten billion objects by hand.
This is probably net positive for the science, and that’s a position worth stating. Astronomy’s history is full of examples where the bottleneck wasn’t the quality of observations but the inability to analyze what had already been collected. The Palomar plates are still there. AI-based analysis tools are working through them now, and they’re finding things that sat undiscovered for 70 years.
At the same time, there’s a genuine loss in the transition. Many important discoveries in astronomical history came from someone staring at data long enough to notice something unexpected. Henrietta Swan Leavitt spent years measuring the brightness of variable stars on photographic plates and recognized the period-luminosity relationship that later became the foundation for measuring cosmic distances. That kind of slow, careful attention to data is hard to replicate in an automated pipeline optimized for throughput.
The AI systems being deployed today are designed to find what they were trained to find. Novel discoveries that fall outside the training distribution – objects or phenomena that nobody anticipated – are not what these systems are built for. Keeping space in the scientific workflow for unexpected findings is something the field is actively thinking about, and it’s a harder problem than building the AI systems themselves.
Near-Earth Objects and Planetary Defense
One area where AI has moved from research tool to operational system is the detection and characterization of near-Earth asteroids. The impact hazard from asteroids isn’t speculative – Earth has been hit by large impactors multiple times in its history, and the 2013 Chelyabinsk airburst over Russia, which injured around 1,500 people, demonstrated that even relatively small objects can cause significant damage with no warning.
The NASA Center for Near Earth Object Studies and partner organizations track hundreds of thousands of known near-Earth objects, and the pipeline for identifying new ones from survey data is AI-driven from the start. When the Rubin Observatory’s LSST begins full operations, it’s expected to discover hundreds of thousands of new near-Earth objects that aren’t currently known. Early results from the Rubin commissioning campaign have already seen new asteroids reported to the Minor Planet Center, and observations of the third known interstellar object, designated 3I/ATLAS.
Processing the detection and orbital determination of that many new objects requires neural networks at every stage. An asteroid’s orbital parameters need to be refined quickly after discovery to determine whether it poses any risk and to ensure the object isn’t lost before the next observation. AI-based orbit determination tools can process new detections faster than any traditional pipeline, which matters when a newly discovered object’s orbital solution is still uncertain.
Cosmological Parameter Estimation
One of the more technically demanding applications of AI in astronomy is in inferring the fundamental properties of the universe from observational data. The standard cosmological model has roughly six free parameters describing the overall composition and geometry of the universe – the matter density, dark energy density, amplitude and slope of the primordial power spectrum, optical depth to reionization, and the Hubble constant.
Traditionally, these parameters are estimated by comparing observed quantities against theoretical predictions using Markov Chain Monte Carlo sampling, a method that is computationally expensive. Neural network-based inference approaches, sometimes called simulation-based inference, sidestep this by training a network to estimate the posterior probability of the model parameters directly from simulated data. Once trained, the inference runs nearly instantly. Researchers at Princeton University and New York University have demonstrated this for cosmological parameter estimation using weak gravitational lensing data, achieving competitive constraints without the traditional computational overhead.
The Hubble tension – the persistent disagreement between the Hubble constant measured from the early-universe cosmic microwave background and the value measured from local distance indicators – remains one of the most discussed problems in cosmology. Several groups are using AI methods to search for systematic errors in the measurements or to test whether modified cosmological models can resolve the discrepancy. Whether AI will reveal new physics here or confirm that the tension is genuine remains open.
Future Missions and Embedded AI
The next generation of space missions is being designed with AI embedded at the instrument level. The Nancy Grace Roman Space Telescope, scheduled to launch by May 2027, carries a 300-megapixel camera with a field of view roughly 100 times wider than the Hubble Space Telescope’s primary camera. It will generate data volumes that require automated processing pipelines at every stage.
ESA‘s Euclid telescope, which launched in July 2023 and began science observations in 2024, is mapping the geometry of the dark universe by imaging billions of galaxies and measuring their shapes with extreme precision. Weak gravitational lensing – the subtle distortion of galaxy shapes caused by intervening dark matter – requires the shapes of millions of galaxies to be measured accurately, demanding AI-based shape measurement algorithms. The science literally couldn’t be done without them.
Further ahead, the proposed Habitable Worlds Observatory – a large UV/optical/infrared space telescope concept that would directly image Earth-like planets around nearby stars – involves AI at every level, from coronagraph control to spectral analysis of potential biosignatures. If such a mission were ever to detect an oxygen-rich atmosphere on a planet in the habitable zone of a nearby star, AI would have been part of how that detection was made and interpreted.
Summary
The core shift in astronomy over the past decade isn’t that AI is replacing astronomers. It’s that the volume of data has grown so large that the science was either going to become AI-dependent or become impossible. The Rubin Observatory’s first images in June 2025, an AI system at the Allen Telescope Array processing data 600 times faster than its predecessor, and a general-purpose language model classifying supernova events with 96.7% accuracy from fifteen training images – these aren’t milestones pointing toward some future state. They’re descriptions of where the field is right now.
What hasn’t received enough attention is the question of discovery style. Every AI system deployed in astronomy today is, at some level, looking for what it was trained to look for. The classification networks find known classes of objects. The matched-filter analogs find signals resembling known signals. The emulators predict outputs consistent with the parameter spaces they were trained on. There is a version of AI-dominated astronomy where the field becomes extraordinarily efficient at confirming and characterizing the phenomena it already knows about, while growing systematically less likely to stumble across something that doesn’t fit.
The pulsars discovered by Jocelyn Bell Burnell in 1967 were initially dismissed as instrumental artifacts precisely because nobody expected a natural source of such regular radio pulses. A pipeline trained to remove radio frequency interference might have silently discarded them. Building AI systems that are genuinely open to the unprecedented – that flag, rather than filter, the deeply anomalous – may turn out to be the most important technical challenge in the field, and it’s one that pure engineering won’t resolve on its own.
Appendix: Top 10 Questions Answered in This Article
How much data will the Vera C. Rubin Observatory produce?
The Vera C. Rubin Observatory will generate approximately 20 terabytes of raw data every observing night during its Legacy Survey of Space and Time, which began commissioning in 2025 with full science operations starting in early 2026. Over its ten-year survey lifetime, it will catalog roughly 20 billion galaxies and detect an estimated 2,000 new supernova explosions per night.
How did AI help discover the exoplanet Kepler-90i?
In 2017, a Google team working with NASA and SETI Institute astronomers trained a convolutional neural network on labeled Kepler light curves to distinguish genuine planetary transits from false positives. The network identified a previously missed transit signal in the Kepler-90 system, leading to the confirmation of Kepler-90i and giving that star system eight known planets – matching our solar system’s count at the time.
Can AI detect gravitational waves faster than traditional methods?
A deep learning model demonstrated by California Institute of Technology researchers in 2018 can detect gravitational wave signals in LIGO data within approximately one millisecond, compared to the minutes or longer required by traditional matched-filter methods. That speed advantage is operationally important because gravitational wave events trigger rapid follow-up observations by other telescopes around the world.
What role did AI play in imaging the black hole in Messier 87?
The Event Horizon Telescope used multiple image reconstruction algorithms, including AI-informed sparse modeling techniques, to reconstruct a black hole image from incomplete interferometric data collected by eight radio observatories worldwide. The consistency of results across different reconstruction methods provided confidence in the validity of the resulting ring-shaped image released in April 2019.
How is CHIME using AI to detect fast radio bursts?
The CHIME radio telescope uses a real-time machine learning pipeline to identify fast radio burst candidates from its continuous all-sky survey data, handling the classification, morphology analysis, and interference removal automatically. By 2023, this approach had produced a catalog of more than 500 fast radio bursts, most detected without real-time human review.
What is the Breakthrough Listen AI achievement at the Allen Telescope Array?
Breakthrough Listen, in partnership with NVIDIA, deployed an end-to-end deep learning system on the Allen Telescope Array in California that processes astronomical data 600 times faster than the previous state-of-the-art pipeline. The system achieves 7% better accuracy than existing methods while reducing false positives by nearly ten-fold, enabling real-time analysis of data streams previously too fast to examine as they arrived.
What is simulation-based inference in cosmology?
Simulation-based inference is a method where a neural network is trained on outputs from many cosmological simulations to learn the relationship between model parameters and observable statistics. Once trained, the network can estimate cosmological parameters from real observational data nearly instantly, without the computationally expensive Markov Chain Monte Carlo sampling that traditional methods require.
What is the CAMELS project?
CAMELS – Cosmology and Astrophysics with Machine Learning Simulations – is a database of over 4,000 cosmological simulations with systematically varied parameters, built specifically to train neural network emulators. These emulators can predict what a new simulation would produce in seconds rather than the hours a full simulation requires, allowing researchers to explore vastly larger theoretical parameter spaces.
How does AI affect the search for extraterrestrial intelligence?
Machine learning, particularly anomaly detection approaches, allows SETI researchers to search for unusual signals across large radio survey datasets without specifying in advance what an artificial signal would look like. In 2020, a neural network applied to Breakthrough Listen archival data found 72 fast radio burst events from FRB 121102 that conventional analysis had missed, demonstrating that AI can find signals in data already processed by traditional methods.
Will AI replace professional astronomers?
AI is taking over routine classification, pattern recognition, and data-processing tasks that would otherwise consume most of an astronomer’s working time. Human astronomers remain essential for hypothesis generation, experimental design, and interpreting results that fall outside what AI systems were trained to expect – which are precisely the results most likely to represent new physics.

