1 Introduction
In today’s rapidly evolving world, artificial intelligence (AI) stands as a beacon of transformation, seamlessly integrating into every facet of our lives. From the convenience of face recognition [
1] unlocking our devices to the magic of speech synthesis [
2] that brings virtual assistants to life, AI’s influence is omnipresent. Diving deeper into the world of language, large language models (LLMs) like ChatGPT (GPT-3.5 and GPT-4) [
3] and Llama 2 [
4] have revolutionized our interaction with digital content. These advanced language models are now being harnessed for creative endeavors, such as co-writing novels, generating art-inspired poetry [
5], and even composing music [
6]. They are also playing pivotal roles in bridging communication gaps, offering real-time translations for diplomats and global travelers, and assisting in preserving endangered languages by generating rich linguistic content [
3]. Beyond these linguistic marvels, AI’s imprint extends to personalized healthcare, where it crafts treatments tailored to individual genetic profiles [
7], and to our cities, optimizing the ebb and flow of traffic [
8]. Venturing into the domain of scientific inquiry, AI emerges as a powerful ally, reshaping our methodologies and accelerating the pace of discovery [
9]. Innovations like self-supervised learning are reimagining how we interpret complex datasets [
10], while generative AI techniques are forging new pathways in diverse fields, from drug design [
11] to the creation of advanced materials [
12]. Delving into specific achievements, tools like AlphaFold2 [
13] have unraveled the intricate puzzle of protein folding, offering unprecedented insights into the very fabric of biological life. Similarly, in the vast cosmos, AI assists astrophysicists in deciphering the myriad signals from the universe, unveiling celestial phenomena that were once shrouded in mystery [
14]. As we journey further into this decade, AI promises to be more than just a technological marvel — it is poised to be the compass guiding our quest for knowledge, reshaping our world, and expanding the horizons of what’s possible [
15].
In 2015, the detection of gravitational waves (GWs) provided a monumental breakthrough in astrophysics [
16], validating Einstein’s century-old theoretical prediction [
17] and introducing a new window to probe the universe’s mysteries [
18,
19]. gravitational wave data analysis (GWDA) is a complex endeavor that consist of many stages. Given the sensitivity of detectors like the Laser Interferemeter Gravitational Wave Observatory (LIGO) [
20], Virgo [
21] and KAGRA [
22], it is imperative to differentiate genuine GW signals from terrestrial interference [
23]. This involves rigorous data quality labeling [
24], glitch classification to categorize transient noise events [
25], and noise suppression techniques to enhance the clarity of potential signals [
26]. In addition to these ground-based efforts, upcoming space-based observatories like Laser Interferometer Space Antenna (LISA) [
27], Taiji [
28,
29], and TianQin [
30] will open another detection window by targeting sources such as massive black hole binary (MBHB), extreme-mass-ratio inspiral (EMRI), galactic binary (GB), and stochastic gravitational wave background (SGWB). With cleaner data in hand, the focus then shifts to signal detection. For example, template-based methods rely on accurate waveform modeling to predict gravitational-wave signals from sources such as black hole mergers [
31] and employ matched filtering to detect these signals [
32]. However, signal detection also encompasses template-free approaches — such as excess power searches and coherent analyses for SGWBs and unmodeled signals — that identify statistically significant features in the data without relying on precomputed templates [
33]. Following a successful detection, the pipeline culminates in parameter estimation, where the goal is to decipher the astrophysical properties of the GW sources [
34]. Each stage of the GWDA process, from data pre-processing to scientific discoveries, presents its unique set of challenges. Real-time analysis is essential to enable timely communication of detections to the external astronomical community, ensuring that follow-up observations and coordinated multi-messenger campaigns can be initiated promptly. Real-time data analysis necessitates maintaining high data quality because it allows for robust and reliable prompt detections [
35,
36]. When it comes to signal detection, the non-stationary and non-Gaussian nature of the noise poses significant hurdle, as traditional matched filters are optimized for Gaussian noise [
37]. To adopt matched filtering in non-Gaussian cases, advanced techniques like adaptive power spectral density (PSD) estimation need to be employed [
38]. The generation of waveform templates, crucial for signal searching and parameter estimation, is computationally expensive and time-consuming when relying on classical methods [
39,
40], resulting in full parameter estimation taking several hours [
41,
42].
GWDA is confronted with challenges stemming from the high-dimensional parameter space and the presence of non-Gaussian, non-stationary artifacts in the interferometer background [
43]. The introduction of AI, particularly deep learning (DL) algorithms, offers a promising avenue to address these challenges [
14,
44] (Fig.1). These algorithms are characterized by their computational efficiency, leveraging accelerated hardware for rapid solutions [
45]. Their scalability ensures they can handle extensive datasets, providing reliable model performance estimates [
46]. The modular nature of these algorithms facilitates adaptability, allowing for the streamlined incorporation of new methodologies [
47]. Furthermore, the generalization capabilities of DL models ensure consistent performance across varied GWDA scenarios [
48]. Several recent reviews have explored the applications of AI in GWDA. Notably, Refs. [
14,
44,
45] focus primarily on ground-based, detectors, while Ref. [
49] offers a more comprehensive overview of the field without providing an overall statistical analysis of the literature (Fig.2). In light of these advancements, our review undertakes:
● The paper reviews various AI-driven approaches to GWDA, covering every stage of the entire compact binary coalescences search pipeline, from waveform modeling to scientific discoveries.
● We offer the review of AI’s application in gravitational waveform modeling, showcasing novel methodologies and their accuracy.
● We also present the review of AI’s role in dramatically accelerating parameter estimation within gravitational wave studies.
● This paper presents the most comprehensive review to date, seamlessly integrating the latest advancements in AI with their practical applications in the field of GWDA.
● Our study culminates in a meta-analysis, synthesizing insights and trends from diverse AI applications in GWDA, discussing the fusion of AI and GWDA, and opening new avenues for insights and future research directions.
With this exploration, we aim to offer a panoramic view, intertwining the intricacies of AI with the mysteries of GWs, underscoring the transformative potential of their collaboration.
The remainder of this review is structured to provide a comprehensive overview of the intersection between AI and GWDAs (Fig.3). In Section 2, we briefly introduce the development of DL. In Section 3, we delve into waveform modeling, discussing both traditional methods and the advancements brought about by DL. Section 4 is dedicated to data quality management, encompassing topics from data quality labeling to noise suppression. Signal detection, a pivotal stage in the analysis, is covered in Section 5, where we explore various methods and their implications. Parameter estimation, a complex yet crucial aspect, is dissected in Section 6. Moving beyond the technicalities, Section 7 sheds light on the broader scientific discoveries enabled by these methodologies. Lastly, in Section 8, we offer a meta-analysis of the literature, discuss futuristic insights, and chart potential directions for the field.
2 Deep learning
With AlexNet’s unprecedented performance in the ImageNet challenge, 2012 marked an important turning point in AI [
86]. This accomplishment demonstrated the potential of neural networks for complex visual tasks. Convolutional neural networks (CNNs) rapidly ascended as the benchmark in image processing, leveraging spatial hierarchies through convolutional layers and pooling operations [
87]. However, as network architectures deepened, the vanishing gradient problem became evident. Residual network (ResNet) addressed this, introducing skip connections to facilitate gradient flow [
88]. Concurrently, for sequential data, recurrent neural networks (RNNs) emerged as a promising approach [
89]. Their capacity to retain state was notable, but challenges with long-term dependencies persisted. WaveNet [
90], designed for raw audio generation, addressed this by employing dilated convolutions, expanding the receptive field. In the unsupervised learning domain, AutoEncoders [
91], with their encoder−decoder structure, excelled at deriving efficient data representations without relying on labels. Together, these innovations post-AlexNet have shaped the trajectory of DL, setting the stage for subsequent advancements.
Contrasted with discriminative models that have been the mainstay of DL, generative modeling has carved its own niche, offering unique capabilities [
92]. Generative adversarial networks (GANs) [
93] emerged as a groundbreaking approach, where two neural networks, a generator and a discriminator, engage in a game-theoretic framework [
94]. The generator crafts synthetic data, while the discriminator discerns between real and generated samples. This adversarial process results in the generator producing increasingly realistic data [
95]. Variational autoencoders (VAEs) [
96] offered another perspective, framing generative modeling as a probabilistic graphical model where the encoder and decoder networks are conditioned on latent variables. Normalizing flows [
97] and Diffusion models [
98] further enriched the generative landscape, providing mechanisms to transform simple distributions into complex data distributions. These generative models have been instrumental in the rise of artificial intelligencegenerated content (AIGC) [
99], enabling the synthesis of high-fidelity images [
100], videos [
101], and even art [
102]. The ability of these models to generate content that is often indistinguishable from real-world data has opened new avenues in digital media [
103], virtual reality [
104], and beyond.
As the field of DL progressed, the Transformer architecture emerged as a groundbreaking innovation, particularly for sequence-based tasks. Introduced by Vaswani
et al. [
105], the Transformer discarded the recurrent layers that characterized RNNs, instead relying on self-attention mechanisms to process input data. This self-attention allows the model to weight the significance of different parts of an input sequence, regardless of their positional distance, enabling it to capture long-range dependencies in the data with ease [
106]. Unlike CNNs, which uses fixed-size filters to process data in local receptive fields, the Transformer’s self-attention mechanism provides it with a dynamic receptive field, adjusting based on the content of the input [
107]. This flexibility allows Transformers to excel in tasks where the importance of data points varies contextually [
108]. Furthermore, while RNNs processes sequences step-by-step, leading to potential issues with long sequences due to vanishing and exploding gradients, transformers process all sequence elements in parallel. This parallel processing not only speeds up training but also alleviates the gradient issues associated with long sequences [
109]. Building on the foundational transformer architecture, LLMs like BERT [
110] and GPT [
111] have set new benchmarks in natural language processing. ChatGPT, encompassing iterations like GPT-3.5 and GPT-4 [
3], epitomizes the rapid evolution and capabilities of LLMs. These models, with their ability to understand, generate, and reason with text, have showcased the superiority of the transformer architecture over traditional CNNs and RNNs in handling sequence data [
112].
While supervised learning and generative models have made significant strides, another paradigm, reinforcement learning from human feedback (RLHF), has emerged as a potent tool in the AI arsenal [
113]. At its core, reinforcement learning is an approach where an agent learns to maximize rewards by interacting with an environment, much like how one might learn a game through trial and error [
114]. However, designing suitable reward functions for complex tasks can be challenging. RLHF, sidesteps these challenges by leveraging human feedback as a primary source of reward signals [
115]. This approach allows models to learn more complex behaviors without the need for explicit reward shaping. OpenAI’s fine-tuning of models like GPT-4 using RLHF is a testament to the power of this approach [
3]. By collecting comparison data, ranking different model responses, and using proximal policy optimization, models can be fine-tuned to produce safer and more useful outputs [
116]. Simultaneously, the field of multi-agent systems is undergoing substantial development [
117]. As AI models become more sophisticated, there is a growing interest in how they interact in shared environments [
118]. Multi-agent systems, where multiple AI entities collaborate or compete, offer insights into emergent behaviors, cooperation strategies, and even the evolution of communication [
119]. The fusion of LLMs with multi-agent setups is particularly intriguing [
120,
121]. Imagine agents equipped with the linguistic prowess of LLMs negotiating, strategizing, and evolving their communication protocols in real-time. Such advancements hint at a future where AI entities do not just operate in isolation but actively learn from, compete with, and collaborate with other AI entities, paving the way for more dynamic and adaptive AI ecosystems [
122].
3 Waveform modeling
GW astronomy has experienced significant progress in recent years, with the numerical modeling of compact binary coalescence — including BBH, BNS, and NSBH — which has been a key component in developing accurate waveform templates [
123]. The journey of these cosmic phenomena is typically segmented into three phases: the inspiral, merger, and ringdown, with the potential for tidal disruptions in systems involving neutron stars [
124]. Moreover, burst signals arising from events such as core-collapse supernovae or tidal disruption events contribute another class of short-duration waveforms, which, despite their less predictable morphology, are crucial for a complete understanding of GW sources. The accurate depiction of these stages through gravitational waveforms is vital for the detection of GW signals and the interpretation of the celestial narratives they unveil [
125,
126].
The creation of waveform templates is an essential aspect of GW data analysis. These templates act as blueprints for expected GW signals from diverse astrophysical origins and are crucial for the comparison with data collected by GW observatories. However, the production of these templates, particularly for intricate systems, poses significant computational challenges. The computational demands and the sparse nature of numerical relativity (NR) waveforms necessitate the use of more practical approaches. To this end, a variety of approximate waveform models have been developed, such as the
[
129–
131] and
[
132–
134] families, which are calibrated against NR simulations to ensure accuracy and efficiency in GW data analysis.
In parallel with these developments, the field of waveform modeling is undergoing a transformative phase with the introduction of DL techniques. DL offers a promising pathway to expedite the generation of waveform templates by employing neural networks that are potentially capable of quickly synthesizing the complex dynamics of compact binary mergers. This innovative approach could speed up the process of generating waveforms, potentially overcoming some of the computational bottlenecks of traditional methods. For a detailed overview of the advancements in waveform modeling facilitated by DL, refer to Tab.1.
Although the evolution of GW waveforms is fundamentally deterministic and governed by differential equations, forecasting in the deep learning context refers to training a model to implicitly learn the waveform evolution from data [
135]. This data-driven approach enables rapid predictions that are useful for real-time analysis [
51]. DL models, particularly transformers, have the potential to revolutionize waveform forecasting. Khan
et al. [
51] stands as a testament to this potential, showcasing the capabilities of DL in predicting the evolution of GW signals.
In GW astronomy, the imperative for swift and accurate waveform modeling is ever-present [
136]. DL presents a promising pathway, but the intricate parameter space of GW signals can be daunting [
137]. Surrogate models, trained on NR waveforms, integrate DL techniques to expedite traditional waveform calculations [
55]. These models stand as rapid waveform generators, adeptly spanning the vast GW parameter space [
128]. Typically, the speed-up factor achieved through surrogate models is greater than 100, making them significantly faster than traditional methods [
138]. Recent investigations indicate that surrogate models may substantially enhance computational efficiency, offering a promising approach to align traditional GW methodologies with the capabilities of AI [
139].
4 Data quality management
Ensuring high data quality is essential in GW astronomy, where detecting these subtle spacetime ripples depends on precise measurements. This section examines the critical aspects of data quality management, including accurate data labeling, robust glitch classification, and effective denoising. Each of these procedures is vital for maintaining data integrity and reliability, which are fundamental to successful GW detection.
4.1 Data quality labeling
The detection of GW is a complex task, primarily due to the challenge of distinguishing these faint cosmic signals from background noise. The sensitivity of GW detectors is such that they pick up a myriad of noises, ranging from seismic activities to instrumental glitches [
140]. This noise can often mimic or obscure the actual GW signals, making the task of identifying genuine events extremely challenging [
141,
142]. The intricate nature of GW signals, often buried deep within the noise, requires sophisticated methods to accurately separate signals from noise. This complexity is a fundamental difficulty in GW detection, necessitating advanced techniques for noise analysis and signal extraction [
61].
The quality of the data plays a crucial role in the successful detection and analysis of GW signals. High-quality data ensures that the signals extracted are accurate representations of astrophysical events. To achieve this, it is essential to label and categorize the data effectively [
24]. Traditional methods of data quality monitoring involve manual labeling, which, while thorough, can be time-consuming and subject to human error [
143]. The vast volume of data generated by GW detectors adds to the complexity, as each data segment needs to be meticulously analyzed to determine its quality. This process is vital as it directly impacts the reliability of signal detection and the subsequent astrophysical interpretations [
24].
DL has emerged as a powerful tool in addressing the challenges of data quality labeling in GW astronomy [
56,
59]. DL algorithms, particularly neural networks, have the capability to automate the process of data quality assessment, offering both efficiency and accuracy [
144]. These algorithms can process large volumes of data rapidly, identifying patterns and anomalies that may indicate issues with data quality [
145]. Several studies have demonstrated the effectiveness of DL in automating the data quality labeling process, significantly reducing the time and effort required while maintaining, if not improving, the accuracy of the labels [
36,
49,
146]. By ensuring that only high-quality data is fed into analysis pipelines, DL contributes significantly to the robustness and reliability of GW signal detection and analysis [
24].
4.2 Glitch classification
Glitches, or transient noise artifacts, are a recurring challenge in GW data analysis. These anomalies can significantly impede signal detection processes (see Fig.4), necessitating their accurate classification and understanding [
147]. The impact of glitches is twofold: they can either mimic authentic GW signals, leading to false positives, or they can obscure genuine signals, resulting in missed detections [
148]. This dual nature of glitches makes them a critical focus in GW data analysis, as their presence can skew the results and interpretations of GW observations.
Traditionally, the classification of glitches has relied on a variety of methods. Notable initiatives in this realm are the Gravity Spy and GWitchHunters project, which have been instrumental in tackling the glitch challenge [
57,
149]. This project has compiled a comprehensive dataset that serves as a crucial resource for the classification of glitches. Traditional methods often involve manual inspection or semi-automated techniques, where analysts use witness data—auxiliary information from the detectors — to help identify and categorize these glitches. These methods, while effective, can be time-consuming and may not always capture the subtle nuances of different glitch types [
150].
The advent of DL has introduced a new paradigm in glitch classification, offering more efficient approaches [
146]. Some of these techniques directly analyze time-series data from GW detectors, identifying patterns that are characteristic of specific types of glitches [
159]. Another innovative approach involves transforming time-series data into visual formats, allowing CNNs to classify glitches based on their visual signatures [
59,
146,
152,
153]. This method leverages the pattern recognition capabilities of CNNs to discern between different glitch types effectively. Looking ahead, the integration of DL in glitch classification is poised to play a pivotal role in the future of GW data analysis, potentially leading to more accurate detections and a deeper understanding of the cosmos [
149]. For a detailed exploration of these DL methods, refer to Tab.2.
4.3 Data denoising
In the realm of GW astronomy, noise is an unavoidable and challenging factor that significantly impacts signal detection [
148]. The presence of noise in the data result in a low signal-to-noise ratio (SNR), leading to difficulties in identifying genuine GW events [
32]. Effective denoising is therefore crucial, as it not only improves the SNR but also decreases the false alarm rate (FAR), making it easier to distinguish real signals from noise. Additionally, denoising plays a vital role in the convergence of Markov-chain Monte Carlo (MCMC) parameter estimation processes. By suppressing the noise, denoising methods help in faster and more accurate extraction of signal from the noisy data, thereby enhancing the overall efficiency and reliability of GW signal analysis [
61].
Traditional denoising methods in GW astronomy have focused on direct noise suppression techniques, encompassing a variety of approaches [
160–
162]. These include variational-based methods [
160], wavelet-based methods, and techniques utilizing the Hilbert-Huang transform [
163]. Each of these strategies aims to remove noise components while preserving the integrity of potential GW signals. For instance, variational-based methods apply mathematical optimization techniques to filter out noise [
164], while wavelet-based methods use wavelet transformations to reconstruct signals from data [
165,
166]. The Hilbert-Huang transform, a novel approach, is particularly effective for non-linear and non-stationary data [
163]. While effective in certain scenarios, these traditional approaches can be limited in their ability to handle the complex and dynamic nature of noise in GW data without the help of data from witness channel [
167].
Although DL-based denoising methods are not yet integrated into current detection pipelines, research has shown that they have advanced signal extraction techniques in GWDA. These include dictionary learning [
168], which employs basis waveforms for signal representation; WaveNet [
169] for its sequential data handling capabilities; denoising autoencoders [
170] for reconstructing signals from noisy data; and RNN [
171] for capturing temporal dependencies. Notably, the transformer-based model, as discussed in Refs. [
61,
62], has been applied to both ground-based and space-based GW denoising and detection (Fig.5), showcasing its potential in handling complex GW data, the transformer-based approach demonstrated measurable improvements in sensitivity and a reduction in false alarm rates compared to traditional methods, thereby underscoring its practical utility in real-world applications.
5 Signal detection
5.1 BBH
The detection of GWs, is a formidable task, primarily due to the subtlety of these signals against a backdrop of overwhelming instrumental and terrestrial noise. The advent of interferometric detectors, notably LIGO and Virgo, has revolutionized GW astronomy by enabling the first direct detection of GW event. However, the increased sensitivity and observational bandwidth of these instruments have also introduced intricate data analysis challenges. A cornerstone of GWDA has been matched filtering [
32], a technique that cross-correlates observed data with theoretical waveform templates. Given the high-dimensional parameter space associated with potential astrophysical sources, this method, while powerful, is computationally demanding, especially in the context of real-time analysis.
Machine learning, particularly deep learning, has emerged as a potential tool for addressing the computational challenges in GW detection [
63,
64]. The initial forays into integrating DL algorithms with mock LIGO data underscored the viability and promise of these approaches [
63,
64,
172–
174]. These models, trained on simulated waveforms, were subsequently tested on synthetic data. Impressively, their performance in discerning potential GW signatures from noise was on par with traditional matched filtering-based techniques, highlighting an important step forward in GWDA [
65], Tab.3 lists out some related works, and their area under the curve (AUC) for detailed performance comparison corresponding receiver operating characteristic (ROC) curve is depicted in Fig.6.
In light of the proven effectiveness of machine learning on simulated LIGO datasets, the subsequent logical step was its application to genuine observational data. This transition presented inherent complexities, primarily attributed to the non-Gaussian and non-stationary nature of the noise, which elevated false alarm rates. Addressing these challenges, Wang
et al. [
66] integrated a matched filtering sensing layer prior to the CNN, showcasing its performance on O1 data. Building upon foundational research [
60,
177], Huerta
et al. [
46] employed a hardware-accelerated WaveNet to probe for GW signals within a month-long span of LIGO data. Concurrently, Wang
et al. [
61] proposes WaveFormer, a deep learning-based data quality enhancement method for real observational gravitational wave processing, achieving state-of-the-art noise suppression performance, which paves a solid foundation for future strides in GW data processing and search [
178,
179]. See Tab.4 for a detailed performance comparison of FAR.
As the GW community approaches a transformative phase, the ongoing development of space-based and third-generation ground-based detectors heralds a new observational epoch. These advanced instruments, set to provide unparalleled observational capabilities, also bring forth intricate data analysis challenges, particularly due to overlapping GW signals [
180]. Recent endeavors on LISA [
27] and Einstein Telescope (ET) [
181] mock data have shown promising results [
62,
67,
182]. The ongoing advancement and incorporation of machine learning techniques remain crucial for effectively addressing these challenges in the forthcoming era.
5.2 EMRI
EMRIs are among the most intriguing astrophysical phenomena in the universe. These events occur when a stellar-mass compact object, such as a neutron star or a stellar-mass black hole, spirals into a supermassive black hole found at the center of galaxies [
186]. The detection of EMRIs holds significant scientific objectives [
27]. Firstly, they provide a unique probe into the spacetime geometry around supermassive black holes, allowing for precise tests of general relativity (GR) in strong-field regimes [
187]. Secondly, EMRIs can offer insights into the evolution and demographics of compact objects in galactic centers, shedding light on the formation and growth of supermassive black holes over cosmic time [
187].
From a technical perspective, detecting EMRIs poses substantial challenges. The GW signals produced by EMRIs are weak and buried in the noise of space-based GW detectors. Their waveform patterns are intricate due to the complex interplay of relativistic effects, making them difficult to model accurately [
188]. Furthermore, the long duration of EMRI signals demands efficient and robust data analysis techniques to comb through vast amounts of data. Advanced computational methods, including machine learning and deep neural networks, are being explored to enhance the detection capabilities, such as Zhang
et al. [
68], Yun
et al. [
70], Zhao
et al. [
189] detecting EMRI signals using CNNs, achieving rapid detection of EMRIs with high accuracy in time domain, frequency domain, and time-frequency domain. These advancements are poised to play a pivotal role in the forthcoming era of GW astronomy, where space-based detectors like LISA [
27], Taiji [
28,
29], and TianQin [
30] will offer unprecedented observational capabilities for EMRI detection.
5.3 GB
GW studies have revealed a range of astrophysical sources, each with distinct signal characteristics and associated detection challenges. Among these, continuous GWs emitted by GBs or isolated neutron stars—offer valuable insights into the dynamics of compact objects. In this section, we review the characteristics of continuous signals from galactic binaries, focusing on detection strategies and the innovative techniques employed to resolve these overlapping signals from noisy data.
Continuous GWs, unlike their transient counterparts, persist over extended observation periods [
190]. These waves, often emanating from sources like rapidly rotating neutron stars, present a distinct challenge due to their weak amplitude [
191]. Traditional data analysis methods often grapple with the intricacies of these continuous signals [
190]. However, the advent of DL has ushered in a new era of possibilities. Neural networks, with their ability to learn intricate patterns from vast amounts of data, have shown promise in detecting and analyzing these weak, continuous GW signals [
192]. Several DL models have been proposed, aiming to enhance the detection capabilities by leveraging the power of convolutional and recurrent architectures [
71,
72,
74,
192–
194]. Although current pipelines for continuous signals rely predominantly on traditional signal processing methods, early studies suggest that deep learning could enhance the speed of these searches [
73]. Ongoing research continues to assess the potential of these methods to complement traditional techniques in the pursuit of continuous GW detection [
195,
196].
LISA is set to revolutionize our understanding of GWs from space. However, one of the inherent challenges LISA faces is the confusion noise. This noise arises from the superposition of countless unresolved sources, primarily binary systems, creating a cacophony that can mask potential signals. Addressing this requires sophisticated data analysis strategies and a deep understanding of the galactic GW foreground.
Gaussian processes (GP) have emerged as a powerful tool in the realm of GW research, particularly for GB GW separation [
75]. By leveraging the non-parametric nature of GPs, researchers can model the statistical properties of GBs, effectively separating it from potential signals [
197]. This technique holds promise, especially in scenarios with overlapping signals or in the presence of non-Gaussian noise, offering a robust method to extract meaningful information from the data [
198].
6 Parameter estimation
The extraction of astrophysical source properties from observed GW signals, a process integral to GW astronomy, is known as parameter estimation (PE). At the heart of PE lies the MCMC algorithm, a statistical method that has been instrumental in the field [
199]. MCMC operates on the principle of building Markov chains, which are sequences of random samples that, over time, approximate the desired distribution of the parameters being estimated (see Fig.7 for detail). This method is particularly effective in exploring high-dimensional parameter spaces, making it widely used in GWDA [
200].
Despite the effectiveness of MCMC in providing detailed insights into complex data sets, it comes with its own set of challenges. The primary concerns are its computational intensity and occasional struggles with convergence, especially in scenarios involving intricate and voluminous data. Moreover, global fitting — which aims to jointly estimate parameters for all overlapping signals in extremely high-dimensional spaces — is particularly challenging for space-based missions and third-generation ground-based detectors, significantly increasing the computational demands and speed requirements for parameter estimation [
203]. These limitations become increasingly significant as the complexity of GW data escalates and the demand for rapid analysis grows [
200].
In response to these challenges, the field of gravitational-wave astronomy has begun to explore deep learning models as a complementary approach to traditional parameter estimation methods such as MCMC. Several studies have demonstrated that machine learning-based methods can offer potential improvements in processing large datasets and reducing computational time [
42,
78,
204]. These methods are still under active development, and further research is needed to validate and refine them for operational use.
6.1 Point estimation
Fast PE by MLP. One of the initial forays into DL for PE involved the use of Multi-Layer Perceptrons (MLP). These architectures, designed to provide a direct “point estimate” for each parameter, have the advantage of speed. Capable of delivering estimates in near-real-time. Although MLPs are not currently a standard component of the production pipelines for multimessenger searches, early studies suggest that they could potentially be valuable in scenarios where rapid response is required [
77].
Localization. Beyond the realm of intrinsic parameters, DL models have also been proposed to address extrinsic parameters. For instance, exploratory studies have suggested that DL-based approaches could be used to estimate both sky location and distance, potentially providing rapid source localization estimates for future multi-messenger observations [
205].
6.2 Posterior distribution
While point estimates are invaluable for quick insights, a deeper understanding of source properties necessitates the exploration of the full posterior distribution. To this end, several DL models have been proposed:
Flow-based models. Flow-based architectures, which are known for their ability to model complex distributions (Fig.7), have been a popular choice for density estimation and posteriors. Green
et al. [
206] and Shen
et al. [
207] first demonstrated the efficacy of these models in GW PE by proposing toy models for the low-dimensional parameter sub-space of binary black hole (BBH), then Green and Gair [
208] extended the work to the full parameter space of BBH and evaluated on GW150914. Subsequently, Wang
et al. [
209] accelerated the convergence speed by intergrating domain knowledge into prior distribution. Recently, Dax
et al. [
42,
78,
79], Wildberger
et al. [
210], Williams
et al. [
211,
212], Bhardwaj
et al. [
213] have further refined these models, showcasing their performance on a variety of GW events. Wong
et al. [
214–
216] developed a
-based [
217] framework for rapid PE. The ability of these models to capture the nuances of the posterior distribution, especially in scenarios with complex degeneracy, has been a key factor in their success. Furthermore, the computational efficiency of these models, coupled with their ability to generate samples, has potential in addressing the challenges of traditional PE methods.
Conditional variational autoencoders (CVAE). Another promising avenue is the use of CVAE. As demonstrated by Gabbard
et al. [
218] and Green
et al. [
206], CVAEs, by conditioning on observed data, can generate samples that align closely with the true posterior distribution, offering a probabilistic perspective on source parameters.
Gaussian processes (GP). For specialized sources such as GBs and EMRIs, GP have been employed, as seen in Refs. [
75,
219]. Tailored to the unique signatures of these sources, GPs provides refined posterior estimates, enriching our comprehension of these intriguing astrophysical systems.
7 AI for GW science
GW astronomy has opened a new window to the universe, allowing us to probe extreme astrophysical and cosmological phenomena. With the increasing complexity and volume of GW data, AI has emerged as a powerful tool to address various challenges and unlock new scientific potential in the field.
7.1 Fundamental physics
GWs, ripples in the fabric of spacetime, have emerged as a revolutionary tool for astrophysical exploration. These waves, produced by cataclysmic events such as the merger of black holes or neutron stars, provide a pristine medium to probe the universe’s most enigmatic phenomena [
223]. Unlike electromagnetic radiation, which can be obscured or altered by intervening matter, GWs travels undisturbed, offering a direct and unadulterated glimpse into their sources. The detection of GWs has opened a new avenue to explore the strong-field regime of GR [
224]. Traditional electromagnetic observations, while invaluable, often fall short in this domain, especially when it comes to events like black hole mergers. GW detections, on the other hand, allow scientists to test Einstein’s theory under extreme conditions, shedding light on the intricate dance of massive celestial bodies and the spacetime they warp [
224].
The promise of space-based observatories, particularly LISA, is immense. Positioned far from Earth’s noisy environment, LISA aims to access the low-frequency GW spectrum [
225]. This capability is anticipated to unveil events like supermassive black hole mergers, which have remained elusive to ground-based detectors, offering a new observational window into the dynamics of galactic centers [
226]. The GW community eagerly anticipates the advent of third-generation observatories, such as the ET. These state-of-the-art facilities promise enhanced sensitivity and a broader frequency range [
181]. With these advancements, scientists expect to detect a diverse array of sources, from the dramatic death throes of massive stars in core-collapse supernovae to potential signals from exotic compact objects. The enriched catalog of detections will undoubtedly deepen our understanding of the universe’s violent processes.
While GR has withstood a century of scrutiny, GW observations offer a unique platform to test its predictions and probe potential deviations. The precision of these observations might reveal subtle signatures hinting at alternative theories of gravity, pushing the boundaries of our current understanding and opening doors to new realms of physics (Fig.8). The synergy of advanced detector networks has been instrumental in refining parameter estimation techniques. As detectors’ sensitivity improves, so does the accuracy with which we can determine the properties of GW sources. This precision not only allows for a more detailed understanding of individual events but might also offer subtle hints of new physics lurking in the shadows. Among the most tantalizing prospects in GW science is the detection of the SGWB. This omnipresent hum, a relic from the early universe, holds the potential to offer insights into primordial processes and interactions, painting a picture of the cosmos’s infancy. Interdisciplinary efforts, merging the expertise of astrophysicists, general relativists, and data scientists, promise to usher in a new era of GW science. As technology advances and our observational capabilities expand, so too will our understanding of the cosmos, revealing its mysteries one GW at a time.
7.2 Cosmology
GWs offer a unique lens to probe the cosmos, and their potential in cosmological studies is gradually being realized. One pressing issue in cosmology is the Hubble tension—the discrepancy between the Hubble constant values derived from cosmic microwave background radiation and those from local distance ladder measurements [
227]. While traditional methods have presented conflicting results, AI-driven analyses of GW data could provide an independent and precise measurement of the Hubble constant [
228]. By analyzing the GW signals from binary mergers, AI can help refine our understanding of the universe’s expansion rate, offering a potential resolution to the Hubble tension. This synergy between GW astronomy and AI not only underscores the interdisciplinary nature of modern astrophysics but also promises to address some of the most perplexing challenges in the field [
82].
7.3 Astrophysics
GWs emanate from some of the most violent and energetic processes in the universe, making them invaluable tools for astrophysical studies [
229]. Neutron stars, the remnants of massive stellar explosions, are laboratories for extreme physics. Their equation of state (EoS) remains one of the outstanding puzzles in astrophysics [
230]. AI-driven methods can assist in classifying different EoS models based on the GW signals emitted during neutron star mergers [
83–
85,
231]. Furthermore, core-collapse supernovae (CCSN) are cataclysmic events marking the end of a massive star’s life. The EoS governing the processes inside these explosive events can be probed using GWs [
232]. AI can aid in deciphering the intricate signals from CCSN, providing insights into the underlying physics [
233]. As the number of GW detections grows, there’s an increasing interest in understanding the population properties of the sources, be it BBHs, neutron stars (BNSs), or other exotic objects [
234]. AI can assist in population synthesis studies, helping to unravel the formation and evolution histories of these compact objects [
215].
8 Discussion
As the application of DL techniques to GW analysis matures, several key themes and challenges emerge [
45]. This section delves into a meta-analysis of the current landscape, discusses the prospects of waveform forecasting [
235], highlights the importance of gap imputation [
236], and underscores the significance of multi-modality [
237] and interpretability [
51] in DL models.
8.1 Meta-analysis
The intersection of GW astronomy and DL has generated a growing body of research. Our meta-analysis, presented in Fig.1 and Fig.2, provides a quantitative overview of publication trends and the evolution of training dataset sizes within this field. Fig.1 illustrates two key trends: the increase in average training dataset size over time and the rising number of publications — both peer-reviewed articles and preprints on platforms like arXiv. This analysis highlights how the availability of large datasets is associated with a corresponding increase in research output, which in turn supports the development of DL models for GW applications. Each data point in these figures marks progress in the field, reflecting improvements in detector technology and data processing methods, as well as the expansion of the research community. In Fig.2, the breakdown into subdomains — such as waveform modeling, signal detection, and parameter estimation — demonstrates the diversity of research efforts and their interconnections within the GW and AI.
While our meta-analysis offers a retrospective view of the past decade, the trends observed may also help inform future research directions and resource allocation. We acknowledge that these figures represent a limited perspective of a complex field; however, they serve as a useful overview of the evolving landscape at the intersection of GW astronomy and DL.
8.2 Waveform forecasting
GW astronomy has ushered in a new era of understanding the universe, with waveform modeling playing a pivotal role in deciphering the signals from astrophysical cataclysms. Waveform forecasting, a subset of this modeling, focuses on predicting the evolution of these waveforms (Fig.9), especially in scenarios where only a fragment of the waveform is known or computationally feasible to generate.
The importance of efficient waveform modeling cannot be overstated. Traditional waveform modeling, particularly for systems exhibiting higher modes or precession, demands significant computational resources [
134]. The ability to forecast a waveform, rather than compute it in its entirety, offers a promising avenue to mitigate these computational challenges [
238]. As the GW community moves towards real-time detections, the need for rapid template generation becomes essential [
239]. Waveform forecasting can expedite this process, ensuring that potential GW events are not missed during live observations. Some astrophysical systems present intricate waveforms that are computationally intensive to model [
240]. Forecasting provides a mechanism to capture the essence of these systems, offering a balance between accuracy and computational feasibility.
While traditional methods have laid the groundwork for waveform forecasting, they come with inherent limitations, especially when dealing with complex astrophysical systems [
241]. Enter DL. Drawing inspiration from its successes in other domains, particularly sequence forecasting, DL holds the promise of revolutionizing waveform forecasting in GW astronomy [
242]. The potential benefits include enhanced accuracy, reduced computational overhead, and the ability to handle a broader range of astrophysical systems.
However, the journey of integrating DL into waveform forecasting is not without challenges. The complexity of GW signals, the omnipresent detector noise, and the need for vast training datasets to train robust models are but a few of the hurdles to overcome [
235].
In conclusion, as we stand on the cusp of a new frontier in GW astronomy, the symbiosis between DL and waveform forecasting is poised to play a transformative role. Collaborative efforts between the AI and GW communities will be instrumental in navigating this exciting journey ahead.
8.3 Gap imputation
In the realm of time series data analysis, especially when dealing with time series data, the presence of gaps or missing data points can pose significant challenges [
243]. These gaps can arise due to a myriad of reasons, from instrumental downtime to environmental disturbances or even data transmission issues [
244]. GW data, with their intricate patterns and crucial reliance on continuity, are no exception to this challenge [
245].
GW observations, by their nature, require continuous and uninterrupted data streams for accurate data analysis [
246]. Any gaps in this data can lead to biases in parameter estimation (Fig.10) [
236]. Traditional methods employed in other fields to address missing data, such as linear interpolation or mean imputation, often fall short when applied to the complexities of GW signals [
247]. These methods, while simple, may not capture the intricate patterns and dependencies inherent in GW data, leading to inaccuracies [
248].
DL is a paradigm that has shown remarkable success in various domains, including gap imputation in time series data [
249]. DL models, with their ability to learn and capture intricate patterns in data, hold significant promise for addressing the gap imputation challenge in GW data [
250]. The potential benefits are manifold, from improved accuracy in imputation to the capability to handle large gaps that traditional methods might struggle with. However, the application of DL to GW data for gap imputation is not without its challenges. The non-stationary nature of GW data, combined with the noise, necessitates careful consideration and design of DL models [
251]. Additionally, it is essential to ensure that the integration of imputed data does not lead to biases or inaccuracies in further analytical processes [
252].
In conclusion, while DL offers a promising avenue for gap imputation in GW data, the field is ripe for research and exploration. Collaborative efforts between the AI and GW communities could pave the way for innovative solutions, ensuring that our observations of the universe remain as accurate and uninterrupted as possible.
8.4 Multi-modality
The realm of AI, particularly in the domain of DL, has witnessed a surge in the exploration and application of multi-modal techniques [
253,
254]. These techniques harness information from multiple data sources or modalities, aiming to provide a more comprehensive understanding of the underlying phenomena. In many AI applications, multi-modality has proven to be transformative. For instance, in medical imaging, combining visual data from MRI scans with textual patient records has enhanced diagnostic accuracy and predictive modeling [
255]. Similarly, in natural language processing, the fusion of textual, auditory, and visual cues in multi-modal sentiment analysis models has led to more nuanced and context-aware interpretations [
256].
GW astronomy, while primarily reliant on signal data, has the potential to benefit from a multi-modal approach. As a starting point, one can combine raw time series data with their time-frequency representations — such as spectrograms — to capture complementary features. This initial idea, though relatively simple, forms the basis for more sophisticated multi-modal analyses (Fig.11). Consider the richness of information available: beyond the GW signals, there are electromagnetic signals, neutrino observations, and more [
257]. Each modality offers a unique perspective on astrophysical events. While current GW analyses have not yet integrated such diverse data streams, the success of multi-modal techniques in other AI domains suggests potential avenues for exploration. One can envision a future where GW detections are enhanced by concurrent observations from electromagnetic or neutrino observatories [
258]. DL models, adept at handling multi-modal data, could be trained to extract features from each modality and then fuse these features to improve event characterization, source localization, and parameter estimation [
259].
However, the integration of multi-modal data in GW analysis is not without challenges. The synchronization of data streams, the handling of disparate data resolutions and formats, and the development of fusion techniques tailored to the specificities of GW events are all areas that would require meticulous research and innovation [
237]. Yet, the potential rewards are significant. A successful multi-modal approach could provide a more holistic view of astrophysical events, bridging gaps in our understanding and opening new frontiers in GW astronomy [
259].
8.5 Interpretability
Interpretability in machine learning has garnered significant attention [
260], especially when models are applied to complex scientific domains [
261]. In the realm of GW analysis, understanding the decision-making process of models is not just a luxury but a necessity. This ensures that the predictions and insights derived are both reliable and scientifically meaningful. GW signals, with their intricate patterns and the profound astrophysical phenomena they represent, pose a unique challenge for DL models [
76]. While these models can achieve impressive performance metrics, deciphering their decision-making process in the context of such complex signals remains a formidable task [
51].
Recent works have made strides in bridging this interpretability gap. For instance, Khan
et al. [
51] delved into creating models that not only detect but also elucidate the characteristics of GW signals. Similarly, the studies by Shi
et al. [
52] and Zhao
et al. [
62] have contributed to enhancing the transparency of models, especially Transformers, ensuring that their predictions can be traced back to understandable reasoning. An interpretable model in GW analysis offers more than just reliable predictions (see Fig.12). It provides a window into the astrophysical processes that generate these ripples in spacetime. By understanding how a model discerns between different types of signals, researchers can gain deeper insights into the astrophysical events behind these waves, potentially unlocking new facets of our universe.
As the field progresses, there is a palpable need for more research dedicated to enhancing interpretability. The fusion of traditional astrophysical knowledge with transparent machine learning models holds the promise of richer, more profound insights into the cosmos.