← Back to Timeline

Detection and Parameter Estimation of Gravitational Waves from Binary Neutron-Star Mergers in Real LIGO Data using Deep Learning

Astrophysics

Authors

Plamen G. Krastev, Kiranjyot Gill, V. Ashley Villar, Edo Berger

Abstract

One of the key challenges of real-time detection and parameter estimation of gravitational waves from compact binary mergers is the computational cost of conventional matched-filtering and Bayesian inference approaches. In particular, the application of these methods to the full signal parameter space available to the gravitational-wave detectors, and/or real-time parameter estimation is computationally prohibitive. On the other hand, rapid detection and inference are critical for prompt follow-up of the electromagnetic and astro-particle counterparts accompanying important transients, such as binary neutron-star and black-hole neutron-star mergers. Training deep neural networks to identify specific signals and learn a computationally efficient representation of the mapping between gravitational-wave signals and their parameters allows both detection and inference to be done quickly and reliably, with high sensitivity and accuracy. In this work we apply a deep-learning approach to rapidly identify and characterize transient gravitational-wave signals from binary neutron-star mergers in real LIGO data. We show for the first time that artificial neural networks can promptly detect and characterize binary neutron star gravitational-wave signals in real LIGO data, and distinguish them from noise and signals from coalescing black-hole binaries. We illustrate this key result by demonstrating that our deep-learning framework classifies correctly all gravitational-wave events from the Gravitational-Wave Transient Catalog, GWTC-1 [Phys. Rev. X 9 (2019), 031040]. These results emphasize the importance of using realistic gravitational-wave detector data in machine learning approaches, and represent a step towards achieving real-time detection and inference of gravitational waves.

Concepts

gravitational waves convolutional networks signal detection classification real-time gravitational-wave inference posterior estimation bayesian inference inverse problems simulation-based inference feature extraction anomaly detection

The Big Picture

Imagine trying to hear a whisper in a hurricane. That’s roughly the challenge facing gravitational-wave astronomers every time they hunt for signals from colliding neutron stars buried in the relentless, crackling noise of LIGO’s detectors. When two neutron stars spiral into each other, they send ripples through spacetime, ripples so faint that LIGO’s mirrors move less than one-thousandth the diameter of a proton. Finding those whispers quickly enough to alert telescopes around the world is a problem of both physics and time.

The standard approach, matched filtering, works by sliding thousands of pre-computed signal templates across the raw detector output and looking for a fit, like leafing through a catalog of fingerprints to find a match. It works, but it’s slow. Working out what actually collided (the masses, spins, and distance of the merging objects) through Bayesian inference, a statistical method for systematically narrowing down possibilities from noisy data, can take hours to days of supercomputer time. In multi-messenger astronomy, hours might as well be forever: the electromagnetic fireworks from a neutron-star merger can fade in minutes, and you need to know where to point your telescopes right now.

A team at Harvard has now shown that a deep neural network can do in seconds what conventional pipelines take hours to accomplish, and do it in real, messy LIGO data rather than idealized simulations.

Key Insight: For the first time, a deep learning framework correctly identifies all binary neutron-star and binary black-hole events from LIGO’s first observing catalog (GWTC-1) directly from real detector data, distinguishing them from each other and from noise.

How It Works

The researchers built a Convolutional Neural Network (CNN), an architecture originally developed for image recognition that learns to spot patterns in visual data without being told exactly what to look for. Here, instead of identifying cats or faces, the CNN reads detector output formatted as a spectrogram: a picture of how the signal’s frequency content shifts over time.

Training on simulated noise is easy. Training to survive real LIGO noise is the hard part. Real detector data is non-stationary and non-Gaussian, full of glitches, instrumental artifacts, and environmental disturbances that no textbook noise model anticipates. What makes this work different is that the team trained on real LIGO background noise, not idealized approximations, forcing the network to learn the actual character of the instrument.

Figure 1

The pipeline works in three stages:

  1. Detection: The CNN classifies short (~one-second) windows of strain data into three categories: background noise, binary black-hole (BBH) signal, or binary neutron-star (BNS) signal. This runs in milliseconds on a GPU.
  2. Localization in time: Once a candidate event is flagged, the network pinpoints the merger time within the data segment.
  3. Parameter estimation: A second network component maps the detected signal directly to source parameters (specifically component masses) without running a single Bayesian sampler, the traditional statistical machinery that exhaustively explores all possible source properties to find the best fit.

Figure 2

The proof-of-concept result is striking. The framework correctly classified every event in GWTC-1: ten BBH mergers and the single BNS event, GW170817. That’s the real-world validation that prior deep learning papers couldn’t claim, since those earlier networks were tested on synthetic noise, not the instrument itself.

Figure 3

The network’s sensitivity scales well too. By training on injected signals spanning a range of signal-to-noise ratios, the model learns to detect faint events near the detection threshold, exactly the regime where prompt follow-up matters most.

Why It Matters

The immediate payoff is speed. Multi-messenger astronomy lives and dies by latency. When GW170817 merged in August 2017, it triggered one of the most coordinated observation campaigns in history; 70 telescopes across every wavelength watched the aftermath.

But the electromagnetic signal started fading within hours. If the gravitational-wave alert had come faster, astronomers could have caught the earliest, brightest moments. A deep-learning pipeline that flags a BNS event in seconds, with a rough mass estimate attached, dramatically compresses that timeline.

There’s a deeper point here too: neural networks can generalize to real instrument noise in a way that holds up against a ground-truth catalog. That’s not just good performance on held-out test data; it’s good performance on events we know happened. The framework also points toward covering the full 9-dimensional parameter space of compact binary signals, a regime where matched filtering becomes computationally prohibitive. A bigger parameter space means more training data, not more templates to compute at runtime.

Open questions remain. The network must generalize to a much larger catalog as O3 and O4 events accumulate, and BNS mergers are rare (the real-event training set is a single confirmed detection). The authors address this through careful simulation-to-reality transfer, but that challenge sharpens as detectors improve and detection rates climb.

Bottom Line: A convolutional neural network trained on real LIGO noise can detect and characterize binary neutron-star mergers in seconds, validating against every event in GWTC-1 and opening a credible path to real-time gravitational-wave alerts for multi-messenger follow-up.

IAIFI Research Highlights

Interdisciplinary Research Achievement
This work sits at the intersection of modern deep learning and observational gravitational-wave physics, showing that CNN architectures developed for image recognition can function as real-time astrophysical instruments.
Impact on Artificial Intelligence
The CNNs here generalize from simulated training environments to real, non-Gaussian, non-stationary sensor data, a transferability result with implications well beyond astrophysics.
Impact on Fundamental Interactions
Rapid parameter estimation of binary neutron-star mergers accelerates multi-messenger observations that probe nuclear matter at extreme densities and test General Relativity in the strong-field regime.
Outlook and References
Future work will extend the approach to the full 9-dimensional compact binary parameter space and incorporate data from additional detectors like KAGRA; the preprint is available at [arXiv:2012.13101](https://arxiv.org/abs/2012.13101).

Original Paper Details

Title
Detection and Parameter Estimation of Gravitational Waves from Binary Neutron-Star Mergers in Real LIGO Data using Deep Learning
arXiv ID
2012.13101
Authors
["Plamen G. Krastev", "Kiranjyot Gill", "V. Ashley Villar", "Edo Berger"]
Abstract
One of the key challenges of real-time detection and parameter estimation of gravitational waves from compact binary mergers is the computational cost of conventional matched-filtering and Bayesian inference approaches. In particular, the application of these methods to the full signal parameter space available to the gravitational-wave detectors, and/or real-time parameter estimation is computationally prohibitive. On the other hand, rapid detection and inference are critical for prompt follow-up of the electromagnetic and astro-particle counterparts accompanying important transients, such as binary neutron-star and black-hole neutron-star mergers. Training deep neural networks to identify specific signals and learn a computationally efficient representation of the mapping between gravitational-wave signals and their parameters allows both detection and inference to be done quickly and reliably, with high sensitivity and accuracy. In this work we apply a deep-learning approach to rapidly identify and characterize transient gravitational-wave signals from binary neutron-star mergers in real LIGO data. We show for the first time that artificial neural networks can promptly detect and characterize binary neutron star gravitational-wave signals in real LIGO data, and distinguish them from noise and signals from coalescing black-hole binaries. We illustrate this key result by demonstrating that our deep-learning framework classifies correctly all gravitational-wave events from the Gravitational-Wave Transient Catalog, GWTC-1 [Phys. Rev. X 9 (2019), 031040]. These results emphasize the importance of using realistic gravitational-wave detector data in machine learning approaches, and represent a step towards achieving real-time detection and inference of gravitational waves.