AI and Deep Learning for Terahertz Ultra-Massive MIMO: From Model-Driven Approaches to Foundation Models

Wentao Yu , Hengtao He , Shenghui Song , Jun Zhang , Linglong Dai , Lizhong Zheng , Khaled B. Letaief

Engineering ›› 2026, Vol. 56 ›› Issue (1) : 14 -33.

PDF (2397KB)
Engineering ›› 2026, Vol. 56 ›› Issue (1) :14 -33. DOI: 10.1016/j.eng.2025.07.032
Research
research-article

AI and Deep Learning for Terahertz Ultra-Massive MIMO: From Model-Driven Approaches to Foundation Models

Author information +
History +
PDF (2397KB)

Abstract

This study explored the transformative potential of artificial intelligence (AI) in addressing the challenges posed by terahertz ultra-massive multiple-input multiple-output (UM-MIMO) systems. It begins by outlining the characteristics of terahertz UM-MIMO systems and identifies three primary challenges for transceiver design: computational complexity, modeling difficulty, and measurement limitations. The study posits that AI provides a promising solution to these challenges. Three systematic research roadmaps are proposed for developing AI algorithms tailored to terahertz UM-MIMO systems. The first roadmap, model-driven deep learning (DL), emphasizes the importance of leveraging available domain knowledge and advocates the adoption of AI only to enhance bottleneck modules within an established signal processing or optimization framework. Four essential steps are discussed: algorithmic frameworks, basis algorithms, loss function design, and neural architecture design. The second roadmap presents channel state information (CSI) foundation models, aimed at unifying the design of different transceiver modules by focusing on their shared foundation, that is, the wireless channel. The training of a single compact foundation model is proposed to estimate the score function of wireless channels, which serve as a versatile prior for designing a wide variety of transceiver modules. Four essential steps are outlined: general frameworks, conditioning, site-specific adaptation, and the joint design of CSI foundation models and model-driven DL. The third roadmap aims to explore potential directions for applying pretrained large language models (LLMs) to terahertz UM-MIMO systems. Several application scenarios are envisioned, including LLM-based estimation, optimization, search, network management, and protocol understanding. Finally, the study highlights open problems and future research directions.

Graphical abstract

Keywords

Terahertz communications / Ultra-massive multiple-input multiple-output / Model-driven deep learning / Foundation models / Large language models

Highlight

Cite this article

Download citation ▾
Wentao Yu, Hengtao He, Shenghui Song, Jun Zhang, Linglong Dai, Lizhong Zheng, Khaled B. Letaief. AI and Deep Learning for Terahertz Ultra-Massive MIMO: From Model-Driven Approaches to Foundation Models. Engineering, 2026, 56(1): 14-33 DOI:10.1016/j.eng.2025.07.032

登录浏览全文

4963

注册一个新账户 忘记密码

1. Introduction

1.1. Background

Our society is undergoing a digital revolution marked by a drastic increase in both connectivity and throughput [1]. Media transmission has expanded significantly, encompassing images, videos, and, imminently, augmented and virtual reality streams [2,3]. The advent of fifth-generation (5G) mobile communication technology has introduced transformative benefits through the concept of “connected things,” while global efforts in sixth-generation (6G) mobile network research and development are expected to usher in a new era of “connected intelligence” with groundbreaking applications and services [4], as shown in Fig. 1. Prominent applications such as the artificial intelligence of things (AIoT), autonomous driving, smart manufacturing, and edge artificial intelligence (AI) are anticipated to play vital roles in 6G and beyond systems. To support these developments, innovative technologies are required to manage the exponential increase in mobile data traffic and its varied applications. Future communication systems must meet demanding requirements in terms of throughput, scalability, latency, and complexity [5,6]. Anticipated performance targets include high data rates of up to 1 terabits per second (Tbps), extremely low end-to-end latencies of less than 100 µs, high spectral efficiency of approximately 100 bits per second per Hertz (bits·s−1·Hz−1), ultra-wide bands of up to 3 THz, and numbers of connections reaching at least 108 devices per square kilometers. These requirements mandate transformative advances in wireless technology.

Several white papers and technical reports from the International Telecommunication Union (ITU) [7], 5G Americas [8], and China’s IMT-2030 (6G) Promotion Group [9] have emphasized the importance of exploring unexplored higher frequency bands for 6G and beyond systems. Among these, the terahertz band, that is, the spectrum spanning 100 GHz to 10 THz, remains largely underutilized and presents opportunities to fulfill the ever-increasing demand for wireless links [10]. The US Federal Communications Commission (FCC) has already allocated 95 GHz-3 THz spectrum for 6G, positioning the United States as a leader in the 6G race [11]. Furthermore, the Institute of Electrical and Electronics Engineers (IEEE) 802.15.3d standard has initiated preliminary standardization for terahertz-band communications [12]. Along with the quest for higher frequency bands, a natural trend exists toward deploying more antennas at base stations (BSs). These developments have evolved from small-scale multiple-input multiple-output (MIMO) systems with only a few antennas in third generation (3G) to large MIMO systems in fourth-generation (4G)/5G systems [13].

1.2. Ultra-massive MIMO systems in the terahertz band

Looking ahead to future wireless networks, ultra-massive MIMO (UM-MIMO) arrays with more than 1024 antennas are expected to be deployed in the terahertz frequency band [[14], [15], [16]]. This transformative technology can address significant path loss and molecular absorption loss by employing highly directional beamforming to enhance coverage [17] and leveraging the abundant bandwidth to achieve high spectral and energy efficiencies [18]. Additionally, terahertz UM-MIMO offers higher localization accuracy with reduced transmission power and a smaller footprint compared to its millimeter-wave (mmWave) counterparts [19], while supporting the integration of sensing and communication functionalities [20,21]. Another key application is in nanocommunications, such as on-chip communications and in-body networks, which benefit from the compact size of terahertz arrays built using nano- and meta-materials [22,23]. In addition, line-of-sight (LoS) terahertz UM-MIMO arrays have the potential to replace copper or fiber point-to-point links in data center networks and provide high-capacity links for aerial and space networks, enabled by spherical-wave spatial multiplexing in the near-field of the array [24,25]. Terahertz-band communications are also poised to improve connectivity and physical-layer security in 6G and beyond systems [26].

Nevertheless, the study of terahertz UM-MIMO systems is still in its early stages and several unique challenges must be addressed [27]. These challenges can be summarized by three “hard-to” problems from a signal processing perspective. First, the massive system scale and short coherence times in terahertz UM-MIMO systems make many traditional model-based designs too complex, resulting in a “hard-to-compute” problem. This necessitates low-complexity, real-time algorithms capable of handling high-dimensional signal processing and optimization tasks efficiently. Second, the complex channel characteristics in the terahertz band give rise to several new phenomena, such as hybrid far- and near-field effects (also known as the hybrid-field effect [28,29]), spatial non-stationary effect [30,31], and wideband beam-squint effect [32]. Together, these phenomena present a “hard-to-model” challenge in transceiver design, as classical optimization and analysis methods often rely on precise system and channel modeling. Overcoming this requires innovative approaches that can learn from and adapt to complex environments without relying on analytical models. Third, there is the “hard-to-measure” problem, specifically referring to channel measurement. Although “hard-to-measure” may also imply difficulties in evaluating the performance of AI-based communication systems using traditional metrics, in this study, the phrase refers solely to the challenges in channel measurement. The array-of-subarray (AoSA) architecture commonly used in terahertz UM-MIMO systems results in far fewer radio frequency (RF) chains than the number of antennas [33,34]. Combined with hardware impairments [[35], [36], [37]], this architecture produces incomplete and corrupted measurements during channel estimation, significantly complicating the acquisition of accurate channel state information (CSI) [38]. To overcome these limitations and to enhance the performance of CSI acquisition and channel-dependent tasks, novel solutions should be studied to leverage historical data and overcome the drawback of corrupted measurements.

Notably, UM-MIMO has also been investigated in the literature using other terminologies, such as extremely/extra large-scale MIMO (XL-MIMO) systems [31, [39], [40], [41]], extremely large aperture array (ELAA) [42,43], and gigantic MIMO (gMIMO) [44] for different frequency bands. In this study, given the focus on terahertz communications, we adhere to the terminology terahertz UM-MIMO, originally proposed in Ref. [15], as it was the first to introduce this concept specifically for the terahertz band.

1.3. AI for communications

Since 2017 [45], AI, particularly deep learning (DL), has regained attention in the wireless communications community and has become an indispensable tool for the design and optimization of large-scale multi-antenna systems [46,47]. The applications of AI can be categorized according to the three “hard-to” challenges discussed previously. Firstly, one potential application of AI is to exploit its capability to facilitate the solution of “hard-to-compute” problems, those traditionally considered intractable owing to challenges such as high-dimensionality, high-cardinality, or non-convexity [48]. AI models can be trained to directly and implicitly approximate the desired solution or function by replacing bottleneck modules in established optimization or signal processing algorithms with learnable components [49]. These two paradigms are called data-driven and model-driven approaches in wireless communication literature [50,51]. In addition, AI is also helpful in tackling “hard-to-model” problems by learning complex and non-linear relationships from data, eliminating the need for accurate analytical models. This has been particularly successful in high-dimensional problems at the physical layer and model-free problems at the medium access control (MAC) layer and above [46,47]. Finally, “hard-to-measure” problems may be addressed using generative AI [52] and foundation models. These approaches learn channel distributions from historical data and generate high-fidelity synthetic channels for data augmentation, thereby enhancing channel-dependent tasks. Furthermore, they can serve as an open-ended prior for solving inverse problems such as channel estimation, prediction, and tracking, in the physical layer using incomplete and corrupt measurements [52].

After nearly eight years of exploring AI for communication, a few shared perspectives have gradually emerged in this field. The first is that AI is especially valuable when analytical models fail to deliver optimal solutions [46,48,53,54], whether due to excessive complexity (i.e., the “hard-to-compute” problem) or the lack of an accurate model (i.e., the “hard-to-model” problem). This understanding has given rise to the model-driven DL paradigm, where AI augments traditional model-based methods by replacing their bottleneck components with learnable modules, thereby integrating expert knowledge with learning capabilities [50,55]. The second perspective is that data will probably be a major bottleneck for training AI models in wireless communications [[56], [57], [58]]. Large-scale data collection and channel measurement are required to characterize site-specific channel and user distributions, which are both expensive and time-consuming (i.e., the “hard-to-measure” problem). The third perspective, based on our own understanding, is that wireless transceivers share a common foundation—the wireless channel. This raises the question of whether a unified foundation model can be developed to serve various transceiver modules, rather than designing separate AI models for each problem. A unified model can reduce both training and deployment costs. These last two perspectives inspired us to propose a new concept: the CSI foundation model. Based on the available data, we aim to train a single, compact generative AI model capable of capturing site-specific channel characteristics for data augmentation and serving as a versatile prior for designing a wide range of transceiver modules.

1.4. Contributions and organization

Although terahertz UM-MIMO is envisioned as a promising candidate for 6G and beyond wireless systems [59], research in this field is still in its early stages, with significant potential for future development. As the field continues to evolve, researchers interested in terahertz UM-MIMO systems may wish to apply AI to solve various emerging challenges but lack guidance on where to begin. Conversely, experts in AI for communications may not be familiar with the distinct features of terahertz systems, limiting their ability to contribute. This study aims to bridge these gaps by introducing the key system and channel characteristics of terahertz UM-MIMO, identifying the associated challenges, and illustrating how AI can be leveraged to address them. We propose two systematic research roadmaps—model-driven DL and CSI foundation models—and provide a step-by-step guide for navigating each. We highlight key frameworks and techniques essential for developing AI-based solutions for terahertz UM-MIMO systems. Through this study, we seek to foster collaboration between experts in AI and terahertz communications to advance this exciting interdisciplinary field [60,61].

The organization of this paper is illustrated in Fig. 2. Section 1 provides the introduction. Section 2 discusses the preliminaries of terahertz UM-MIMO communications and highlights key system and channel characteristics. Section 3 examines how the three “hard-to” challenges manifest in terahertz UM-MIMO systems and explain why AI is well-suited to tackle these challenges. Section 4 presents research roadmaps for developing model-driven DL and CSI foundation models for terahertz UM-MIMO systems, illustrating their design principles, key components, and representative case studies. Finally, Section 5 concludes the paper.

2. Terahertz UM-MIMO communications

In this section, we introduce the key features and channel models of terahertz UM-MIMO systems and explain the challenges they present in transceiver design. Because an overview paper [18] already offers a comprehensive introduction to the mathematical models of terahertz UM-MIMO, we avoid redundancy and direct readers to Section 2 in Ref. [18] for more details. Open-source codes for terahertz UM-MIMO systems are available in Ref. [38].

2.1. Key system characteristics

Terahertz-band waves experience severe propagation path loss due to molecular absorption at terahertz frequencies [62,63]. UM-MIMO antenna systems, comprising thousands of antennas, have recently emerged as promising solutions to overcome this limitation [15,16]. The small wavelength in the terahertz band allows the deployment of a large number of antennas with a small footprint. Consequently, super-narrow beams are formed, which help mitigate path losses and expand communication range. The practical implementation of terahertz UM-MIMO has become feasible with the development of new plasmonic materials, such as graphene and nanomaterials, that enable the construction of nanoantennas and transceivers for terahertz band communication [23].

2.1.1. AoSA architecture

Terahertz UM-MIMO systems face constraints owing to the high-complexity and power-consuming demands of terahertz hardware. These costly components preclude traditional designs used in lower-frequency bands. A fully digital architecture with one RF chain per antenna is infeasible because of its high cost. Similarly, the fully connected hybrid analog-digital architecture for mmWave bands [33,64], where a limited number of RF chains drive the entire antenna array, is also inefficient in the terahertz regime due to transmit power and circuit feeding limitations. In this setup, each RF chain must be connected to all antennas via phase shifters to perform analog beamforming or combining [34]. This requires a significant number of phase shifters in the terahertz band, making it expensive to implement.

Consequently, the prevailing choice for terahertz UM-MIMO arrays is a simplified architecture called AoSA [18,34]. In this architecture, a UM-MIMO array is grouped into various nonoverlapping subarrays (SAs), with each RF chain connecting to and powering only its corresponding SA, as illustrated in Fig. 3(a) [29]. This is similar to the partially connected architecture used in mmWave bands [33,65]. AoSA significantly reduces the number of required phase shifters in the analog beamformer (or combiner) and lowers energy consumption.

Fig. 3(b) [29] illustrates a magnified view of a part of the terahertz UM-MIMO array. The spacing between a pair of adjacent antenna elements (AEs) is denoted as da, and that between adjacent SAs is represented by ${{d}_{\text{sub}}}\ \triangleq \ \omega {{d}_{\text{a}}}\ \left( \omega \ \gg \ 1 \right)$, where ω is a constant. Here, da is typically small owing to the small wavelength in the terahertz band. Conversely, the SAs should be separated by a much larger distance (i.e., ω≫1), because closely integrating too many AEs can create difficulties in the control and cooling of circuits [18,66]. As a result, terahertz UM-MIMO arrays are often nonuniform, differing from conventional uniform arrays. This non-uniformity necessitates specialized considerations in algorithm design [38,67].

2.1.2. Beam-squint effect

The AoSA architecture illustrated in Fig. 3(a) [29] is used for narrowband terahertz UM-MIMO systems. However, in wideband terahertz systems, a new challenge known as beam squint arises, requiring modifications to the original narrowband architecture [32]. In wideband terahertz UM-MIMO systems, an ultrawide bandwidth and a large number of antennas can result in a non-negligible propagation delay across the antenna array, which may exceed the sampling period. This leads to variations in the angles of departure (AoDs) or the angles of arrival (AoAs) across different subcarriers during transmission or reception, making the array gain frequency selective. Given this characteristic, if frequency-flat phase shifters are adopted, the resultant beams will be dispersed and will point towards various angles, as shown in Fig. 3(c) [29]. To address the beam-squint effect, various studies have adopted true-time-delay (TTD) modules to compensate for angle variations [68,69], such as delay-phase precoding [70] and joint phase-time arrays [71]. TTD modules are typically deployed between each RF chain and the associated phase shifters. The number of TTD modules is usually smaller than that of the phase shifters because of their relatively higher costs [68]. In addition, the beam-squint effect can be exploited to accelerate wideband terahertz beam tracking by controlling the degree of the squint to scan multiple angular directions simultaneously [72,73].

2.2. Key channel characteristics

Measurement campaigns are currently underway worldwide to further the understanding of terahertz propagation characteristics. For comprehensive surveys, readers are referred to Refs. [74,75]. As this study focuses on physical layer signal processing in terahertz UM-MIMO systems, we summarize only the most relevant channel features here.

2.2.1. Hybrid-field effect

Near- and far-field phenomena are crucial at terahertz frequencies [76]. In the far-field region, the wavefront can be assumed to be approximately planar. However, when signals originate in the near-field region, this assumption no longer holds [77]. An exactly spherical wavefront model must be considered when modeling the terahertz UM-MIMO channels with near-field multipath components, as illustrated in Fig. 3(c) [29].

The boundary between the far-field and the (radiating) near-field regions is called the Rayleigh distance, defined by $\frac{2{{D}^{2}}}{{{\lambda }_{\text{c}}}}$, where D is the array aperture and ${{\lambda }_{\text{c}}}$ denotes the carrier wavelength. While the Rayleigh distance appears to increase linearly with frequency for a fixed aperture, this relationship can be misleading. In practice, the array aperture D is also a function of the carrier wavelength, as the AE spacing is typically set to half the wavelength. Consider a planar AoSA with $\sqrt{S}\times \sqrt{S}$ SAs with dsub≜ωdaω≫1, where each SA is a uniform planar array with S¯×S¯ AEs, S and $\bar{S}$ are both constants. In this case, the Rayleigh distance is given by ${{\left\{ \sqrt{S}\left( \sqrt{{\bar{S}}}-1 \right)+\left( \sqrt{S}-1 \right)\omega \right\}}^{2}}{{\lambda }_{\text{c}}}$ [38]. This shows that the boundary between the far- and near-field regions is jointly determined by the carrier wavelength and array geometry and should be analyzed on a case-by-case basis.

In a general scenario, sources and scatterers can be positioned in both the far- and near-field regions of a UM-MIMO array. Consequently, different multipath components can arrive at the array as spherical or planar wavefronts. Hence, terahertz UM-MIMO channels typically consist of a dynamic mixture of both, which have been identified in the context of channel estimation as the hybrid-field effect [29,38,78]. In the literature, this phenomenon is referred to as the cross-field effect [28,79], although both terms convey a similar concept.

Spherical wavefronts provide important benefits such as enhanced spatial multiplexing, high-precision localization, and transverse velocity sensing. However, they complicate the representation, acquisition, and exploitation of channels, particularly when mixed with far-field planar wavefronts, due to the hybrid-field effect.

2.2.2. Multipath components

Terahertz channels are typically sparse, with a limited number of resolvable paths, owing to high scattering and diffraction losses in the terahertz band. For example, Yan et al. [80] reported that at 300 GHz, the number of multipaths was only five for a 256×256 UM-MIMO array, which was 32.5% less than its counterpart at 60 GHz. Intuitively, the terahertz-band channel exhibits a higher K-factor than the lower-frequency bands owing to greater reflection and diffraction losses [75]. Although we emphasize LoS-dominant propagation, multipath components should still be considered, particularly in indoor terahertz communications. Existing measurement campaigns are mostly conducted within the sub-terahertz band (i.e., 100-300 GHz). Further investigation is needed for the full terahertz band.

2.2.3. Molecular absorption effect

When terahertz electromagnetic waves propagate through a nonvacuum medium, they can trigger resonances in certain molecules along their paths, resulting in notable energy loss at certain frequencies. From a communications perspective, this causes strong frequency-selective channel gains at different subcarriers in a wideband terahertz system. This phenomenon is known as the molecular absorption effect. The absorption strength can vary with environmental conditions, such as humidity and temperature, and depends on the propagation distance. A detailed frequency profile of these absorption peaks can be found in Ref. [63]. These peaks divide the terahertz band into multiple narrow spectral windows with relatively low absorption. Therefore, terahertz communication should only be placed within these spectral windows [63]. Conversely, molecular absorption loss can be leveraged for designing distance-adaptive absorption peak modulation to improve the covertness of terahertz communications [81].

2.2.4. Spatial non-stationary effect

The spatial non-stationary effect occurs when the terminals or scattering clusters are visible only from a portion of the UM-MIMO array [31]. This phenomenon is more prevalent in linear arrays, where the array aperture is larger than in planar arrays with the same number of AEs. Additionally, the lower-frequency end of the terahertz spectrum is more susceptible to spatial non-stationarity in the sub-terahertz band, where the larger wavelengths result in increased AE and SA spacings, and hence larger array apertures. In contrast, planar arrays operating at higher terahertz frequencies exhibit less pronounced spatial non-stationarity owing to their smaller apertures. Therefore, a novel three-dimensional (3D) non-stationary geometry-based stochastic model has been proposed for terahertz UM-MIMO systems [30].

3. Integrating DL with terahertz UM-MIMO: Three key motivations

We believe that DL is especially effective in tackling three key challenges in wireless communications—that is, “hard-to-compute,” “hard-to-model,” and “hard-to-measure.” In this section, we discuss the manifestation of these challenges in terahertz UM-MIMO systems, this explains the motivation for integrating DL with terahertz UM-MIMO systems.

3.1. “Hard-to-compute” problems

In terahertz UM-MIMO networks, the system scales drastically in terms of network density, number of antennas, supportable users, and system bandwidth. This leads to high-dimensional signal processing and optimization problems and presents challenges in terms of computational complexity. Additionally, the channel coherence time in terahertz UM-MIMO networks is extremely short, resulting in rapid channel fluctuations. Conventional statistical and optimization-based approaches involving computationally intensive operations, such as singular value decomposition (SVD), bisection search, and matrix inversion, may struggle to meet latency requirements. These challenges highlight the importance of low-complexity methods.

Unlike traditional methods, DL excels at making fast approximations to avoid heavy computation [48]. It can learn intricate patterns and solve large-scale signal processing and optimization problems approximately in real time. Traditional algorithms treat every problem instance as completely new and solve each from scratch; however, DL can recognize similarities between incoming problem instances and find a shortcut [49]. When trained and fine-tuned in a site-specific manner, DL can efficiently learn the distribution of “problem instances” unique to that environment, allowing it to outperform general algorithms in terms of performance and efficiency [82]. In addition, DL models can operate in parallel across various wireless resource domains. For example, in wideband terahertz UM-MIMO systems, different SAs and subcarriers can be computed simultaneously. The industry is also actively developing AI-native radio access network (RAN) algorithms that leverage the powerful parallel computation capabilities of graphics processing units (GPUs) [83].

3.2. “Hard-to-model” problems

Modeling capability is also a key motivation for applying DL. The success of classical mathematical tools such as optimization and analysis highly depends on the accuracy of system and channel models. Because the network architecture, communication environment, and wireless channels in terahertz UM-MIMO systems are heterogeneous, complex, and non-linear, analytical models struggle to capture them precisely. In terms of network architecture, the limited coverage of terahertz systems necessitates further densification, which can cause complex interference issues. Additionally, narrow beams increase the possibility of misalignment. In terms of the communication environment, terahertz waves are highly susceptible to spatial non-stationarity, signal blockage, and frequency-selective molecular absorption loss. Real-world networks are significantly more complex than simplified system models due to these combined factors. Most importantly, as mentioned in Section 2.2, terahertz UM-MIMO channels may consist of a dynamic mixture of planar-wave and spherical-wave multipath components due to the hybrid-field effect [28,38]. Consequently, the widely used angle-domain sparsity properties or simplified prior distributions in traditional far-field systems are no longer applicable. This makes it difficult to estimate and track wireless channels [84,85].

3.3. “Hard-to-measure” problems

As discussed previously, terahertz UM-MIMO systems often utilize the AoSA architecture to enhance energy efficiency and reduce costs. Such architectures have far fewer RF chains than antennas [34]. The signals received in each timeslot provide only an incomplete measurement of the channels. Moreover, the presence of frequency-selective noise, hardware distortions, and impairments adds another layer of complexity to the measurement process [36,37,86]. Incomplete measurements and impairments significantly complicate the estimation and tracking of channels and environmental dynamics [[87], [88], [89], [90]].

Wireless channels are pivotal in transceiver design. Difficulties in acquiring CSI can lead to performance degradation in many downstream tasks that rely on channel statistics. To address these problems, we can pretrain generative AI models offline using site-specific historical data to serve as priors, compensating for missing information due to incomplete measurements [57,91,92]. Such AI-based priors, when incorporated with appropriate algorithmic frameworks, can solve versatile tasks including channel estimation [93,94], data detection [95,96], CSI compression and feedback [97], and RF source separation [98]. In addition, they can be directly sampled to perform data augmentation and generate synthetic channel samples to further facilitate the design of transceiver modules.

4. Research roadmaps: From model-driven DL to foundation models and LLMs

In this section, we present three systematic roadmaps for developing AI-enabled solutions to address the key challenges posed by terahertz UM-MIMO systems. These roadmaps include model-driven DL models, CSI foundation models, and large language model (LLM) applications. Our goal is to inspire readers to follow these roadmaps to design AI-enabled solutions for their own research in terahertz UM-MIMO systems. We outline the three roadmaps, consisting of essential steps and case studies, in Fig. 4. By following these steps, one should be able to develop effective AI-enabled algorithms for various problems in terahertz UM-MIMO. Additionally, we discuss how the roadmaps intersect and how they can be seamlessly integrated. Because we aim to structure this paper as a roadmap, we refer readers to the existing literature for in-depth discussions on these topics and provide detailed explanations only for those lacking adequate guidance. Although these roadmaps are primarily designed for the characteristics of terahertz UM-MIMO systems, most of them are also backward-compatible with traditional MIMO and massive MIMO systems in the sub-6 GHz, upper-mid, and mmWave bands. This is because these systems typically operate on a smaller scale and exhibit simpler channel characteristics.

For ease of reading, a summary of the AI/DL methods and frameworks is provided in Table 1 [38, 91, 92, 94, [99], [100], [101], [102], [103], [104], [105], [106], [107], [108], [109], [110], [111], [112], [113], [114]].

4.1. Roadmap 1: Model-driven DL

4.1.1. Overview

DL algorithms for communications can be categorized into two different paradigms: model-driven DL and (fully) data-driven DL. The data-driven paradigm trains neural networks, such as multilayer perceptrons (MLPs), to map system parameters directly to the desired outputs without relying on domain knowledge or existing algorithms. This approach leverages the universal approximation capabilities of neural networks to learn solely from data [115]. Early attempts at applying DL to communications mostly followed this paradigm. For example, an MLP was trained to map the received pilot and data blocks to the detected data symbols [116]. In another example [117], an MLP was trained to learn the direct mapping between the inputs and outputs of an optimization algorithm for interference management.

However, many physical-layer problems involve well-defined system models and established algorithms that already offer efficient solutions. A data-driven paradigm may overlook valuable domain knowledge of the system. While neural networks can, in principle, learn this knowledge from large amounts of data, such an approach is not sample-efficient and requires substantial training data—akin to reinventing the wheel. As training data in wireless systems are often difficult to collect, it is essential to improve the performance of DL tools using the limited available data. Additionally, unlike traditional signal processing and optimization algorithms that can directly adapt to different system parameters, such as the number of antennas, users, and signal-to-noise ratios (SNRs), data-driven DL models often lack generality and flexibility. The outputs of data-driven DL methods may not conform to the physical constraints defined by wireless system models. Finally, data-driven DL approaches often cannot provide the same level of theoretical guarantees as traditional algorithms.

Model-driven DL is a framework that integrates system domain knowledge into the design of DL models. In particular, model-driven DL replaces bottleneck modules within established signal processing or optimization algorithms with neural network-based learnable components. Often, we recognize that the overall algorithmic framework is optimal; however, its practical implementation is hindered by specific bottlenecks. These bottlenecks may stem from “hard-to-compute” problems, where certain parts of the algorithm are computationally intensive; “hard-to-model” problems, where no analytical solution exists for some components; or “hard-to-measure” problems, where it is difficult to obtain sufficient channel data for training. The earliest work on model-driven DL in communications dates to Ref. [118], where a denoising convolutional neural network (DnCNN) was incorporated as a non-linear denoiser within the framework of the approximate message passing (AMP) algorithm. Following this pioneering study, numerous works have been conducted based on model-driven principles [119].

In the following section, we break down the procedure for designing model-driven DL algorithms into four steps, as shown in Fig. 4: determining algorithmic frameworks, selecting basis algorithms, loss function design, and neural architecture design. By designing the neural architecture and loss function of the DL components, we enforce them to generalize across system scales and dynamic environments and reduce the requirement for clean channel data and optimal labels in the training process. In addition, when combined with CSI foundation models, the DL components in model-driven DL can be shared, which significantly reduces deployment costs. We guide readers through each of these steps and present two representative case studies that apply model-driven DL to terahertz UM-MIMO systems.

4.1.2. Determining algorithmic frameworks

This is the first step in designing model-driven DL algorithms. The problem at hand is analyzed, and the decision to employ iterative or non-iterative algorithms is taken depending on the computational and memory budgets. We propose two general algorithmic frameworks that are scalable, efficient, and theoretically guaranteed. For iterative algorithms, we advocate the fixed point networks (FPNs) framework [38,120] shown in Fig. 5(a) [85], whereas for non-iterative algorithms, we discuss the neural calibration (NC) framework in Fig. 5(b) [99].

The idea of FPNs was inspired by the fact that many widely used iterative algorithms in wireless communications, such as AMP and proximal algorithms, can be regarded as fixed-point iterations of a contractive operator [121]. A contractive operator is characterized by a Lipschitz constant of less than 1, ensuring linear convergence to a unique fixed point through a fixed-point iteration. To design FPNs, we first identify the contractive operator associated with the target iterative algorithm. We then replace the bottleneck modules with a neural network and construct a learnable operator. The learnable operator is trained to be contractive, with a fixed point that closely approximates the desired solution. This framework is illustrated in Fig. 5(a) [85].

The training process for FPNs aims to achieve two key objectives. First, the learnable mapping must be contractive to ensure fast and monotonic convergence, which can be achieved using regularization techniques [29,122]. The Lipschitz constants of compositions of different mappings equal the product of their respective Lipschitz constants. Thus, we can calculate the necessary Lipschitz constant for the neural network and use regularization techniques to ensure that the learnable mapping is contractive. Second, the fixed point of the learnable operator must closely approximate the desired solution. Training can be conducted in two distinct modes: end-to-end (E2E) and plug-and-play (PnP).

E2E mode: In this approach, we train the learnable components of the contractive operator by directly optimizing the overall performance of the iterative algorithm. For instance, Yu et al. [38] trained a learnable operator to optimize the final estimation accuracy.

PnP mode: In this mode, the training process begins by identifying the specific role that the learnable module assumes within the iterative algorithm, as well as the Lipschitz constant required for convergence. For instance, if a minimum mean-square-error (MMSE) denoiser is required in an iterative algorithm, we can independently train a module that fulfills this role. Once trained, this module can be “plugged” into certain iterative algorithms to enhance its performance, as demonstrated previously in Ref. [101]. One advantage of the PnP scheme is that the learnable module can be separated from a specific algorithm; hence, it can be reused in various algorithms.

The contractive-mapping property offers several advantages for FPNs, including but not limited to the following:

Convergence: The Banach fixed-point theorem guarantees that each iteration brings the solution closer to a fixed point, thereby ensuring monotonic convergence.

Adaptive performance-complexity tradeoff: Owing to monotonic convergence, more iterations correspond to a closer distance to the fixed point and better performance. The number of iterations can be adaptively controlled according to the computational budget.

Scalability and low complexity: Based on the implicit function theorem, the gradient in E2E training can be computed with constant complexity regardless of the number of iterations [122]. The computation relies only on a fixed point and does not require storage of intermediate states during the iteration. In the PnP mode, training is independent of the iterative algorithms and avoids the complexity caused by iterations. Its low complexity makes it favorable for large-scale systems.

Deep unfolding/unrolling networks (DUNs) discussed in literature refer to concepts similar to FPNs [100,123]. Both DUNs and FPNs aim to replace bottleneck modules in existing iterative algorithms with neural networks. The difference lies in the training process; DUNs do not impose contractive constraints. Consequently, they lack several advantages that are unique to FPNs. For instance, DUNs generally do not guarantee convergence, their computational complexity is not adaptive, and their training process requires the storage of all intermediate states, resulting in high computational and memory demands [85]. Most DUNs can be transformed into FPNs by enforcing contractive constraints. We recommend FPNs over DUNs because they are easy to implement and offer several advantages.

Subsequently, we introduce the NC framework for non-iterative algorithms [99]. This can be considered a degenerate version of the FPNs with only one iteration. As shown in Fig. 5(b) [85], the NC framework consists of two components: a structured neural network and a time-efficient basis algorithm. We retain the backbone of the low-complexity method while integrating the neural networks to calibrate the inputs. The existence of calibration mappings to improve system performance was proved in Ref. [99]. The NC employs a structured neural network architecture that leverages the permutation equivariance (PE) property inherent to wireless networks. PE is a property whereby an algorithm’s output remains unchanged regardless of the order of its input elements. This phenomenon is common in wireless networks. Two representative PE properties are the uplink antenna-wise PE and downlink user-wise PE, where the order of antennas and users should not affect the algorithm’s performance. By leveraging this property, NC-based algorithms can be generalized to different numbers of antennas and users [99]. This is important for terahertz UM-MIMO systems, whose design may involve high-dimensional problems with varying system scales, such as the number of users [85].

4.1.3. Selecting basis algorithms

Once the general algorithmic framework has been established, the next step is to select a (near) optimal basis algorithm tailored to a specific problem. This process requires expert knowledge of wireless systems. After identifying the basis algorithm, it is crucial to analyze its bottleneck modules and determine whether the bottleneck is due to the “hard-to-compute” problem, such as a large-scale matrix inversion, or the “hard-to-model” problem, such as the lack of prior distribution or optimal step size schedule.

Upon identifying the bottleneck module, it should be separated from the other components of the basis algorithm. Only the bottlenecks should be replaced with suitable neural networks. We analyze a few examples from existing research on model-driven DL for terahertz UM-MIMO.

Example 1: The orthogonal AMP (OAMP) algorithm serves as a near-optimal basis algorithm to solve compressive channel estimation problems using FPNs. Each iteration consists of a linear estimator (LE) and a non-linear estimator (NLE). The LE of OAMP utilizes information from the system model to decouple the original compressive sensing problem into equivalent additive white Gaussian noise (AWGN) denoising problems for the NLE module to solve [124], and it can be derived in closed form. The bottleneck of the OAMP algorithm for channel estimation is the prior distribution required by the MMSE-optimal denoiser in the NLE module, which is challenging to obtain due to the complex channel characteristics of terahertz UM-MIMO. After identifying the NLE as the bottleneck module, the LE is kept unchanged in each iteration, and the NLE, which depends on the prior distribution, is replaced with a neural network [38].

Example 2: The near-field multi-user beam-focusing problem presents “hard-to-compute” challenges due to the computational complexity of the weighted MMSE (WMMSE) algorithm in terahertz UM-MIMO [125]. Consequently, a low-complexity zero-forcing (ZF) scheme can be used as the basis algorithm. However, the bottleneck of ZF is its suboptimal performance when serving many users. Hence, we resort to the NC framework and employ neural networks to calibrate the input of the ZF beamformer, aiming to achieve performance comparable to that of iterative WMMSE algorithms but with lower complexity [85,99].

Hardware imperfections in the terahertz band mainly include mutual coupling, phase noise, antenna failures (e.g., gain and phase errors), and power amplifier (PA) non-linearity [18,74,126], among others. Currently, most studies on hardware-impaired terahertz systems are based on simulations, and different studies have considered different combinations. Hardware limitations can be grouped into three categories based on whether they influence the wireless channel h, the measurement matrix M, and the noise n. Impairments affecting the channels, such as mutual coupling and antenna failures, can be absorbed into the channels, as the goal is usually to estimate the effective channel. Impairments affecting the measurement process, such as phase noise and PA non-linearity, can be considered in model-driven DL, as the LE is designed to decouple the measurements. However, these impairments can render the measurements a non-linear transformation of $y=Mh+n$, where y denotes the received pilot signals. Hence, a more complicated LE should be utilized for decoupling. Basis algorithms compatible with non-linear the measurement process should be considered, such as those in Ref. [127]. Impairments affecting the noise term are often modeled as spatially correlated noise [18]. In these cases, the NLE of the model-driven DL is responsible for handling more complex noise types. If the noise statistics are known, whitening can be applied to simplify the problem, or the noise statistics can be incorporated into the neural network design to improve performance. Additionally, adversarial loss and test-time training [38] can be incorporated to enhance robustness to unexpected changes in noise statistics during online deployment.

4.1.4. Loss function design

Once the algorithmic framework and basis algorithm are determined, the next step is to design a loss function. The simplest approach is to use standard classification or regression losses based on relevant input-label pairs, referred to as direct loss functions. However, direct loss may be inefficient in some scenarios. This has motivated the development of improved indirect loss functions, including task-oriented and empirical Bayesian losses. Representative cases are discussed below.

Task-oriented losses: Given the extremely large scale of terahertz UM-MIMO systems, obtaining optimal labels for standard loss functions is often computationally challenging or even impossible. Task-oriented losses can be utilized to tackle the “hard-to-compute” problem, especially for resource allocation problems. In an early work, Sun et al. [117] used the iterative WMMSE algorithm to create labels for training a neural network for interference management; however, this process may be prohibitively complex for large-scale systems. Instead, losses should be designed to bypass the expensive label generation step and directly train neural networks to solve the target task. These are referred to as task-oriented losses. One of the earliest examples of this approach is in Ref. [128]. For the sum-rate maximization problem in downlink multiuser beamforming, Huang et al. [128] proposed using the negative sum-rate directly as a task-oriented loss. This approach avoided the cumbersome label-generation process and achieved performance comparable to the WMMSE algorithm. Similar concepts have been widely applied in model-driven DL frameworks for terahertz UM-MIMO. For instance, Nguyen et al. [100] emphasized the importance of using an unsupervised task-oriented loss to directly minimize the objective function in hybrid beamforming design.

Empirical Bayesian losses: Another important consideration in loss function design is the “hard-to-measure” problem. In terahertz UM-MIMO systems, obtaining a large channel dataset to train neural networks is difficult, particularly for channel estimation problems. This creates a chicken-and-egg problem: Accurate training data relies on channel estimation, but enhanced channel estimation is difficult without sufficient data. In channel estimation problems, the received pilot signals y are typically modeled as $y=h+n$ or $y=Mh+n$ [94]. In practice, only y are available, not the ground-truth h. Therefore, the mean-square-error (MSE) loss adopted in many previous studies cannot be applied, that is, $\mathbb{E}\|\hat{h}-h\|_{2}^{2}$, where $\hat{h}$ denotes the estimated channel and $\mathbb{E}$ denotes the expectation over wireless channels. To circumvent the dependence on ground-truth $h$, empirical Bayesian methods may be adopted. For linear models, $y=h+n$, Stein’s unbiased risk estimator (SURE) is a good surrogate for the MSE loss [129]. Furthermore, for the generalized linear model $y=Mh+n$, the generalized SURE (GSURE) can be adopted as an extension [130]. For high-dimensional problems in terahertz UM-MIMO, SURE and GSURE can be efficiently computed using only a few Monte Carlo trials, as reported in the literature, because of the high dimensionality of terahertz UM-MIMO channels [131]. Both SURE and GSURE have been successfully applied to UM-MIMO systems to achieve (near) MMSE optimal channel estimation performance in unknown environments [101,132]. The extension from white noise to spatially correlated Gaussian noise is discussed in Ref. [94], while recent results extend them to other exponential families of noise distributions [133], covering most noise types encountered in wireless communication. Empirical Bayesian losses facilitate unsupervised learning and adaptation in unknown environments without requiring prior knowledge or access to clean channels.

4.1.5. Neural architecture design

The neural architecture is also pivotal in enhancing the generalization and efficiency of model-driven DL. While the system and channel conditions in terahertz UM-MIMO systems may change rapidly owing to blockage, severe path attenuation, and hybrid-field effect, enhancing the generalization capability of the designed neural networks to accommodate these variations is crucial, including the number of users, far- and near-field multipaths, and SNR levels. Therefore, we introduce graph neural networks (GNNs) and hypernetworks as two candidate tools. Furthermore, because the limited number of multipaths in terahertz UM-MIMO channels suggests that the effective channel dimensions are significantly smaller than the number of antennas, it is feasible to work with a reduced-dimensional channel representation to lower the complexity of DL algorithms. Accordingly, we propose the neural RF radiance field (NeRF2), Gaussian splatting (GS), and learning in the transform domain as potential tools.

Graph neural networks: A promising direction for improving scalability and generalization is the design of neural architectures tailored to wireless networks. One important finding is that GNNs perfectly align with the PE property of wireless networks [134], making them exceptionally effective in tackling large-scale resource management [135] and data detection problems [102,136]. Unlike traditional DL models, such as convolutional neural networks (CNNs) and MLPs, which often struggle with large-scale networks and new system settings, GNNs effectively utilize graph topology and the permutation invariance property inherent in wireless communications. In addition, in GNNs, the input-output dimensions of the neural network in each node are invariant with the number of users. This allows them to generalize across different system scales. For example, for the beamforming problem in an interference channel, a GNN trained on a small-scale network with 50 users can achieve near-optimal performance in a much larger-scale network with 1000 users [134]. Furthermore, GNNs are computationally efficient owing to their parallel execution capability and, to date, represent the only method capable of finding near-optimal beamformers for thousands of users within milliseconds [134]. This is particularly suitable for UM-MIMO networks with an extremely large system scale and a varying number of users. Please refer to Ref. [137] for a detailed overview of GNN applications in large-scale wireless networks.

Hypernetworks: Owing to hybrid-field effects and blockage at terahertz frequencies, network parameters such as the number of far- and near-field paths, as well as SNR levels, can vary rapidly. When deploying AI-based algorithms in practical systems, a major challenge is ensuring that neural networks can adapt to such dynamics. Hypernetworks provide an effective solution [138]. A hypernetwork generates the weights for another neural network, known as the target network. Instead of fixed weights, the target network weights are dynamically produced by the hypernetwork based on its input. By setting parameters such as SNRs, number of paths, and speed of users as inputs to the hypernetwork, different weights can be generated for the target network under different system conditions. This allows the model-driven DL to seamlessly adapt to varying system conditions in the terahertz bands. Existing works in this direction include channel estimation [139], prediction [103], and joint source-channel coding [140].

Learning on the transform domain: Owing to high path loss and limited diffraction in the terahertz band, terahertz UM-MIMO channels can be represented as a superposition of a limited number of paths. This characteristic implies a lower-dimensional representation of the terahertz channels compared to an ultra-massive number of antennas. Working on the transform domain of a channel with sparse features can reduce difficulty in training AI models for terahertz UM-MIMO channels. For channels where all multipaths lie in the far-field region, the array response vectors depend only on the angles of arrival, rendering the channel sparse in the angular domain. In such cases, the discrete Fourier transform (DFT) matrix is a suitable dictionary for sparse transformation. For near-field channels, Cui and Dai [141] first introduced an angle-distance domain dictionary, which was later enhanced with Newtonized dictionaries [90] and discrete prolate spheroidal sequence (DPSS) dictionaries [89] for improved performance. For generic hybrid-field channels, dictionary learning techniques can be used to obtain site-specific dictionaries [104]. A comparison between the results of Refs. [38] and [131] shows that applying learning in the sparse transform domain yields both improved performance and faster convergence for the same model-driven DL network (e.g., FPN-OAMP) used in channel estimation. Furthermore, learning in the transform domain can significantly reduce the number of training samples required to reach a performance level comparable to that achieved in the original non-sparse domain.

NeRF2 and GS: In a given propagation environment, once the positions of all transmitters are determined, the channel response at any position in that environment can be determined. Based on this principle, Google introduced the neural radiance field (NeRF) to model light field ray tracing. This concept was later extended to RF signals in Ref. [142], where NeRF2 was proposed to model RF signal propagation as a continuous volumetric scene function. After training on limited measurements, NeRF2 can provide accurate location-to-channel mapping, enabling prediction of the channel response at any spatial position within the environment. This enhances both the accuracy of channel modeling and the dimensionality reduction of high-dimensional wireless channels. Such mapping can be used in downstream tasks such as channel estimation and prediction. Although the original experiments were conducted using a 5G massive MIMO array at sub-6 GHz frequencies, extending this idea to terahertz UM-MIMO systems holds considerable promise. However, a major limitation of NeRF2 is its computational complexity [105]. In a typical setup, NeRF2 requires approximately 200 ms to synthesize the channel characteristics for a given scene. This exceeds the requirements of most latency-sensitive applications. Wen et al. [105] recently proposed the utilization of 3D GS to accelerate the environmental aware channel modeling, called wireless radiation field GS (WRF-GS). Under the same setup, WRF-GS required only 5 ms for channel synthesis, which was significantly more efficient than NeRF2.

4.1.6. Case study 1: UM-MIMO beam-focusing

We consider downlink beamforming (beam-focusing) [143] to maximize the sum-rates in a multiuser UM-MIMO system operating in the near-field region [85]. We explain how to walk through the four essential steps to design an effective model-driven DL algorithm for this problem based on the NC framework.

Step 1. Determining algorithmic frameworks. In the first step, we determine the type of algorithmic framework to be applied. Although the WMMSE algorithm offers near-optimal results, its complexity makes it difficult to implement in UM-MIMO systems [125]. Therefore, we focus on non-iterative linear algorithms because of their lower computational complexity, and to enhance their performance using the NC framework to achieve a performance comparable to that of the WMMSE.

Step 2. Selecting basis algorithms. We select a ZF beamformer as the basis algorithm because of its good tradeoff between complexity and performance. The bottleneck of the ZF scheme is its suboptimality when the number of users is large. Hence, we designed a structured neural network module to calibrate the input to the ZF beamformer and enhance its performance, as shown in Fig. 5 [85].

Step 3. Loss function design. Using the direct loss function requires generating optimal labels, which requires repeatedly running high-complexity iterative WMMSE algorithms and is very expensive to implement. To avoid a cumbersome label generation process, we choose a task-oriented loss to directly minimize the negative sum-rates of the considered system without labels. This significantly reduces the complexity of the training process.

Step 4. Neural architecture design. We design a structured neural network based on the PE property of wireless networks. We model the wireless network as a directed graph, with nodes representing the BS and users, and edges representing transmission links. The beam-focusing task is then framed as an optimization problem over this graph, where the PE property ensures that the order of user presentation does not affect system performance. Leveraging this property, we can deploy identical CNNs on every user, each sharing a common set of parameters, which dramatically lowers the total trainable weights and the training overhead. In addition, as identical CNNs use the same parameters, they can be applied to any number of users. Simulation results demonstrate that this approach can effectively generalize across a varying number of users.

Fig. 6 [85] shows the advantages of the proposed NC-based ZF beamformer. The performance upper bound is an iterative WMMSE algorithm with a significantly high computational complexity. The ZF and maximum ratio transmission (MRT) schemes were also compared. In the “matched” case, the number of users is the same during training and inference, while in the “mismatched” case, the training stage contains 75 users, but more users (i.e., 75-200) appear during inference. We highlight the following observations from the simulation results: First, it is shown that the proposed NC-based ZF outperforms both the original ZF and MRT schemes, and its performance compares well with the iterative WMMSE algorithm. In addition, the “mismatched” case indicates that the NC-based scheme trained on a small-scale network can be generalized to a much larger one. Finally, the proposed NC-based beamformer has significant advantages in terms of the computational complexity. Despite a similar performance, the runtime of the NC-based ZF scheme is only 0.16 s as compared to 19 s for the iterative WMMSE algorithm when the number of users is 150, which means that it is nearly 118 times more efficient. In this case study, we assumed a fully digital UM-MIMO system. It will be interesting to extend the NC framework further to the AoSA architecture in the future [144,145].

4.1.7. Case study 2: UM-MIMO data detection

In the second case study, we switch our focus to the data detection problem, that is, detecting data symbols $x={{\left[ {{x}_{1}},\ {{x}_{2}},...,{{x}_{N}} \right]}^{\text{T}}}$ from the received data signals ${{y}_{\text{data}}}={{H}_{\text{mul}}}x+n$, where N is the number of antennas and ${{H}_{\text{mul}}}$ is the multiuser UM-MIMO channel. In UM-MIMO systems, the antenna array is large. In this case study, we demonstrate how model-driven DL can significantly enhance the data detection performance while maintaining low complexity [102,136].

Step 1. Determining algorithmic frameworks. In this case study, we work on iterative detection algorithms because of their better performance than closed-form linear detectors, and we employ to FPNs to enhance them.

Step 2. Selecting basis algorithms. Iterative data detectors consist of a linear module (LM) and a non-linear module (NLM) in each iteration. The OAMP [146] and expectation propagation (EP) [147] detectors can offer a near-optimal performance close to that of the maximum likelihood (ML) detector but entail a high-complexity matrix inversion operation in the LM that incurs cubic computational complexity with respect to the number of antennas. This is prohibitive in UM-MIMO systems with thousands of antennas. Conversely, AMP-based detectors are free of matrix inversion in LE, and their complexity is dominated only by the matrix-vector product [148]. However, AMP generally exhibits inferior performance compared with OAMP and EP. Given that computational complexity is a major bottleneck in UM-MIMO detection, we select the AMP detector as our basis algorithm and apply model-driven DL techniques to enhance its inversion-free LM and boost its performance to a near-optimal level with low complexity.

Step 3. Loss function design. We employ a direct loss function, which is the MSE loss between the detected and true symbols, to train the networks without special designs, like most previous works in this direction [102]. Efforts have been made to design improved loss functions based on the general framework of neural feature learning [149], which can train detectors to generalize to include different fading scenarios without the requirement of online training [150,151].

Step 4. Neural architecture design. The core of the LM in AMP-based detectors is the iterative decoupling of the posterior probability $p\left( x|{{H}_{\text{mul}}},\ {{y}_{\text{data}}} \right)$ into a series of independent scalar probabilities $p\left( {{x}_{i}}|{{H}_{\text{mul}}},\ {{y}_{\text{data}}} \right)$, where $i=1,\ 2,...,\ N$ is the antenna index. In particular, $p\left( {{x}_{i}}|{{H}_{\text{mul}}},\ {{y}_{\text{data}}} \right)$ is assumed to be a Gaussian distribution independent of the other antenna indexes. The performance of the AMP detector is determined by the accuracy of this equivalent AWGN model, which is asymptotically true, but can be inaccurate when used with a finite number of antennas. Hence, He et al. 102] calibrated the LM of AMP using a neural network to learn an accurate approximation for the Kullback–Leibler divergence, enhance the accuracy of the AWGN model, and further improve the performance. Specifically, we chose GNNs because they obey the PE property of MIMO systems and can thus be trained more efficiently and generalized to different system scales.

In Fig. 7 [102], we illustrate the symbol error rate (SER) performance of different detectors. As illustrated, the inversion-free AMP-GNN detector [102] significantly improves the performance of the AMP detector and its performance matches closely with that of the near-optimal EP detector with a high-complexity matrix inversion. In the “matched” case, the training and inference are both carried out in a 32×24 MIMO system. In contrast, in the mismatched case, we trained the AMP-GNN using a mixture of 32×16 and 32×32 channels and tested it on a 32×24 system. The mismatch in the system scale causes little performance loss for AMP-GNN, which verifies its good scalability.

4.2. Roadmap 2: CSI foundation models

In addition to model-driven DL, we propose a visionary concept known as the CSI foundation model. The central idea is that the designs of various transceiver modules share a common basis, namely, the wireless channels. Knowledge of channel characteristics, such as distribution and second-order statistics, and features, such as sparsity and low rank, is fundamental to the design of all transceiver modules. Most existing studies train dedicated neural networks for different transceiver modules, such as channel estimation networks, CSI compression networks, and beamforming networks, each with its own task-oriented loss function. However, these separate networks can learn similar aspects related to the distribution of wireless channels. Training a dedicated neural network model for each transceiver module can cause significant redundancies in computational complexity, memory usage, and deployment costs. The existence of a common basis for the transceiver design, that is, a wireless channel, suggests the possibility of developing a unified foundation model. This model can provide essential information for designing a variety of downstream transceiver modules. This insight forms the basis for the CSI foundation models.

Specifically, the CSI foundation models aim to achieve two objectives. First, we train a neural network to learn the score function of wireless channels from limited channel data or directly from raw received signals. The inspiration for using the score function comes from the success of score-based generative models [152] and diffusion models [153]. A neural network that estimates the score function can serve as an open-ended prior for the design of various transceiver modules. Second, we develop model-driven DL frameworks that incorporate the physical layer foundation model as a flexible prior. This approach allows a single compact neural network model to function as a PnP prior, which enables the optimal design of many different transceiver modules.

We outline four essential steps for training and applying CSI foundation models and discuss an initial case study. A general research roadmap is shown in Fig. 4. The four steps include general frameworks, conditioning, site-specific adaptation, and joint design with model-driven DL. Finally, we present a case study that applies CSI-foundation models to UM-MIMO channel estimation based on Ref. [94].

Notably, a concurrent paper proposed a similar concept called “foundation models for wireless channels” [154], but with very different contents. Song et al. [155] focused on learning representations and proposed a large-scale transformer-based pre-trained model for wireless channels to generate rich, contextualized embeddings that can outperform raw channels for downstream tasks. Conversely, our work focuses on the common foundations of wireless transceiver design and discusses the training and deployment of a single compact generative prior that can support a wide variety of downstream tasks in transceiver design. The perspectives of these two studies are complementary. Their intersection provides inspiration for the future development of CSI foundation models.

4.2.1. General frameworks

As shown in Fig. 4, the general framework comprises two levels. At the first level, we explore methods for training a neural network to estimate the score function of wireless channels. For (vectorized) wireless channels $h\sim p\left( h \right)$, the prior score function is defined as the gradient of log-density function of the channels (i.e.,${{\nabla }_{h}}\log p\left( h \right))$. We discuss the training strategies in two scenarios: ① when clean channel data are available, and ② when only raw received signals are accessible, which can be both noisy and incomplete (due to the limited RF chains in the AoSA). Next, we discuss how to combine the neural score function estimator with different sampling strategies to design various transceiver modules. We first discuss the level one, training strategies.

Score function from channel data: When clean wireless channels h are available from measurement campaigns, we can train a neural network to approximate the score function—that is, learning ${{s}_{\theta }}\left( h \right)\approx {{\nabla }_{h}}\log p\left( h \right)$, where ${{s}_{\theta }}\left( h \right)$ is the neural score function estimator with parameter θ. We call such a network as “(prior) score network” for simplicity. The training of the score network has been extensively studied and can be achieved by denoising score matching [[155], [156], [157]]. If clean wireless channels are available, we can directly pre-train based on these methods to obtain a score network. In addition, as the effective dimensionality of high-frequency wireless channels is small because of the limited number of paths, it may be beneficial to first transform the channel into an appropriate sparse domain and then perform training on such a domain, which is similar to the idea behind latent diffusion models [158].

Score function from raw received signals: Depending on the system architecture, the received pilot signals can be either $y=h+n$ in fully digital systems [159] or $y=Mh+n$ in hybrid analog–digital systems (such as the AoSA architecture) [3,34]. In the former case, the (prior) score function ${{\nabla }_{h}}\log p\left( h \right)$ can be estimated using empirical Bayesian methods based on the SURE loss [129,160]. Meanwhile, for the AoSA architecture, the GSURE loss [130] can be adopted as an extension.

We then discuss the second level, that is, how to apply CSI foundation models as open-ended priors to enable a variety of downstream tasks in the transceiver design. We discuss four cases in the following paragraphs, including prior sampling, posterior sampling, sequential sampling, and joint sampling, which are certainly non-exhaustive. However, further interesting applications are yet to be discovered.

Prior sampling for data augmentation: Once the score network ${{s}_{\theta }}\left( h \right)$ has been pre-trained on the available channel data or the raw received signals, we can then sample from it by using score-based generative models to synthesize more channel data [152]. Synthetic channel data can be combined with measured channel data for data augmentation to create an expanded dataset that supports various downstream tasks in the transceiver design. Initial results have demonstrated the effectiveness of score-based data augmentation in enhancing the performance of DL-based channel compression, channel coding, uplink–downlink channel mapping, beam alignment, hardware impairment mitigation, constellation shaping, and wireless fidelity (Wi-Fi) sensing [91, 92, [161], [162], [163]].

Posterior sampling for inverse problems: Inverse problems, including channel estimation and tracking, are common in wireless transceiver design. For terahertz UM-MIMO systems, the system model for channel estimation can be simplified as $y=Mh+n$ after vectorization. In particular, M is the measurement matrix jointly determined by the pilot symbols and pilot combiners, and is perfectly known in the system, which is a fat matrix because the number of RF chains is significantly less than that of the antennas in the AoSA architecture. We are interested in estimating the wireless channel h from the received pilot signals y. Suppose that the channels follow a prior distribution $h\sim p\left( h \right)$, the channel estimation problem can be solved from a Bayesian perspective by posterior sampling based on diffusion models. The forward diffusion process follows a Markov chain with gradually added Gaussian noise from ${{h}_{0}}$ to ${{h}_{T}}$, where ${{h}_{t}}$ denotes the generated channels at time step $t\ \left( t=0,\ 1,\ 2,...,\ T \right)$, and T is the maximum number of time steps. The reverse sampling process follows from ${{h}_{T}}$ to ${{h}_{0}}$. To incorporate information from the system model, the posterior score is given by ${{\nabla }_{{{h}_{t}}}}\log p\left( {{h}_{t}}|y \right)={{\nabla }_{{{h}_{t}}}}\log p\left( {{h}_{t}} \right)+{{\nabla }_{{{h}_{t}}}}\log p(y|{{h}_{t}})$ by using the Bayes’ rule. In this equation, the prior score ${{\nabla }_{{{h}_{t}}}}\log p\left( {{h}_{t}} \right)$ is available from pre-trained CSI foundation models. We need to derive the likelihood score ${{\nabla }_{{{h}_{t}}}}\log p(y|{{h}_{t}})$ based on the system model $y=Mh+n$ to complete the picture. Meng and Kabashima [164] proposed a technique called noise-perturbed pseudo-likelihood score to approximate the true likelihood score with low computational complexity for the linear system model $y=Mh+n$. They further extend a similar idea to quantized inverse problems $y=Q\left( Mh+n \right)$ in Ref. [106], where Q(•) denotes a uniform quantizer. This technique has recently been applied to channel estimation in quantized MIMO systems with low-resolution analog-to-digital converters (ADCs) [165].

Sequential sampling for compression: CSI compression and feedback are fundamental components of frequency-division duplex (FDD) massive MIMO systems. In the downlink of such systems, users must estimate the channels and feed them back to the BSs. Previous compression algorithms are mostly based on discriminative rather than generative learning [120,166]. Generative CSI foundation models can also be used for the compression and decompression of CSI. The basic idea is sequential sampling, which uses a posterior sampler based on pre-trained CSI foundation models to iteratively gather informative channel measurements for compression via adaptive compressed sensing [107], and then decompress the codewords by reversing these steps to reconstruct the channels.

Joint sampling for cross-module design: In UM-MIMO systems, the pilot overhead required for accurate channel estimation is significant owing to the high dimensionality and limited number of RF chains. To address this issue, a joint channel estimation and data detection algorithm can be considered to reduce pilot overhead. Such a cross-module design can also be easily supported by CSI foundation models. To accomplish this goal, one can first establish a diffusion process that represents the joint distribution of the channels and symbols given the noisy received pilot signals and subsequently run the reverse denoising process to generate samples. In contrast to the continuous prior distribution of the channels, the prior distribution of the symbols is discrete. Thus, computing the score function of the symbols in the diffusion model is difficult. To address this, Zilberstein et al. [108] and Zilberstein et al. [167] proposed an annealed Langevin dynamics algorithm to incorporate the discrete nature of the constellation elements so that joint sampling can work. In future research, it would also be interesting to discover other possible cross-module designs, such as joint channel estimation and decoding.

4.2.2. Conditioning

In practical applications, it is important to enhance the generality of CSI foundation models for effective adaptation to various scenarios and environmental conditions without retraining. Conditional generation plays a crucial role in achieving this objective by incorporating a conditional label, denoted as c, into the score function, such as ${{s}_{\theta }}\left( h,c \right)$. This label specifies the particular type of the generated channel, enabling the model to tailor its output to specific conditions [168,169].

Several advantages can be harnessed using the conditioning technique for CSI foundation models. First, it enhances efficiency. Using conditioning, a single model can generate diverse types of channels by leveraging conditional generation, which eliminates the need to train multiple models and facilitates practical deployment. For example, when the conditional label is weather, the same model can serve as a score network and generate channels for sunny and rainy days. In addition, this increases generalization capability. Well-trained conditional models can generate a wide variety of channel samples with different characteristics, which enhances the capability of the model to generalize beyond the original training data and improves data augmentation performance. Commonly used conditional labels include the position in the environment and link status, such as uplink/downlink and LoS/non-LoS. Additionally, incorporating multimodal information into conditional labels is a promising direction. For instance, data from vision cameras can effectively assess the environmental status and serve as a potential conditional label. However, incorporating multimodal data for conditioning can be complex and expensive. It is advisable to borrow ideas from semantic [170] or task-oriented communication [171,172] to generate labels that capture the minimal sufficient information of the environment.

4.2.3. Site-specific adaptation

When deploying CSI foundation models at different sites, performance can degrade because of environmental differences. Therefore, it is important to design site-specific adaptation schemes to fine-tune the models. The vanilla scheme fine-tunes the entire model in a new environment. However, this approach can be prohibitive when the number of parameters is high. Considering the potential scale of the foundation model, it is important to design parameter-efficient fine-tuning (PEFT) schemes that retrain only a small portion of all parameters to adapt to the new environment. We may resort to existing fine-tuning schemes for LLMs for inspiration. The adapter tuning scheme adds a few linear layers inside the transformer for better results but can increase the inference delay [173]. BitFit methods only fine-tune the bias terms, which leads to a lower computational cost but degrades performance [174]. One of the most popular methods is the low-rank-adaptation (LoRA) scheme [175]. LoRA is a fine-tuning technique for transformer models that freezes pre-trained weights and introduces trainable low-rank matrices into each layer. This reduces computational complexity and preserves parallel processing capabilities. When multiple devices (such as BSs) participate in the fine-tuning of CSI foundation models, LoRA can be further extended to a federated LoRA over wireless networks to support collaborative PEFT [[109], [176], [177]] and simultaneously protect the privacy of raw data [177,178].

4.2.4. Joint design with model-driven DL

On the one hand, the sampling strategies mentioned above, combined with the learned score network, already belong to model-driven DL frameworks. However, the score function is related to the MMSE denoiser for various types of exponential noise families. Please refer to Ref. [133] for a detailed description. Hence, it is natural to integrate it with AMP-family algorithms [179] through the denoising AMP framework [124] to solve a variety of inverse problems in physical layer communications. The idea of “separation” is the core of the joint design of model-driven DL and CSI foundation models. Model-driven DL separates system-specific attributes (e.g., those typical of a certain system model) from system-agnostic characteristics shared across different system architectures and transceiver modules (e.g., the prior distribution of channels). Model-driven DL provides an interface for separation and considers system-specific factors. As a complement, CSI foundation models can provide prior knowledge regarding shared system-agnostic components and can be seamlessly combined with different model-driven DL frameworks to solve diverse problems.

4.2.5. Case study: CSI foundation models

We present an initial case study on training the CSI foundation model using denoising score matching with the raw pilot signals received and then apply it to UM-MIMO channel estimation. The simulation details are available in Ref. [94].

Step 1. Determining general framework. We consider training the score network from the raw received signals, $y=h+n$. This corresponds to a fully digitalized system model. For the AoSA architecture in UM-MIMO systems, such a system model can be achieved by stacking the received signals from multiple timeslots. Slightly different from the original physical layer foundation model described above, we train the score network sθh by using denoising auto-encoders based on y [94]. Based on the score network and Tweedie’s formula [180], the MMSE-optimal channel denoiser can be obtained in closed form for fully digital systems (i.e., $y=h+n)$. For the general hybrid analog–digital system model $y=Mh+n$, we chose to incorporate the denoiser as the NLE of the OAMP algorithm to jointly design it with the model-driven DL.

·•Step 2. Conditioning. In this initial case study, we did not consider conditioning and pretrained an unconditional score network, owing to limited time. Nevertheless, it is easy to follow the conditioning procedure introduced earlier to further enhance the model.

Step 3. Site-specific adaptation. Since here we train the score network ${{s}_{\theta }}\left( h \right)$ based on the received pilot signals y, the denoising score matching loss can be computed solely based on y. Hence, the model can be adapted whenever there are new received pilot signals arriving. We used Vanilla online learning to tune the model and found that it could quickly adapt to changes in the channel covariance matrix within a couple of pilot transmissions [94].

Step 4. Joint design with model-driven DL. The AoSA architecture results in fewer RF chains than antennas. Consequently, UM-MIMO channel estimation becomes a compressive sensing problem. We adopt the FPN framework and select OAMP as the basis algorithm. The LE of the OAMP can be derived in a closed form. It utilizes information from the system model to decouple the original compressive sensing problem into equivalent AWGN denoising problems for the NLE module to solve. The bottleneck of the OAMP lies in its unknown prior distribution. Bayes-optimal OAMP requires the NLE to be the MMSE channel denoiser [124,146], which, luckily, can be derived based on the physical layer foundation model by using Tweedie’s formula [94,180].

In Fig. 8 [94], we present the normalized MSE (NMSE) performance of compressive channel estimation in UM-MIMO systems as a function of the received SNR. We adopted a spherical wave model for the near-field UM-MIMO channel. The performance bound is the oracle MMSE method, which assumes that perfect knowledge of the channel distribution is available. As observed, the proposed physical layer foundation model performs close to the bound across different SNR levels and significantly outperforms both the least squares (LS) and sample MMSE estimators, verifying its effectiveness. In this case study, we did not consider the wideband beam-squint effect, and the proposed framework can be readily extended to handle such scenarios in a manner similar to that in Refs. [38,181].

4.3. Roadmap 3: Applications of LLMs

Recent breakthroughs in LLMs, such as GPT-4 [182] and DeepSeek-R1 [183], have transcended the original scope of natural language processing, demonstrating transformative capabilities in diverse fields, including code synthesis, scientific discovery, and autonomous system control [184]. Although LLMs have the potential to reshape wireless network designs, their applications in terahertz systems remain unexplored. To the best of our knowledge, no previous study has explicitly investigated LLMs in terahertz systems. This section summarizes existing works in the general area of LLM-enabled wireless networks and analyzes how they may be adapted to systems with terahertz-specific challenges. The discussion is divided into five subsections: LLMs for estimation, optimization, searching, network management, and protocol understanding. Because this is an emerging direction, the ideas and methods discussed here are certainly not exhaustive. In Section 5, we discuss future research directions and open issues for integrating LLMs with terahertz systems.

4.3.1. LLMs for estimation

Estimation problems, including channel estimation, prediction, tracking, localization, and sensing, play pivotal roles in physical-layer communication. We summarize the possible ways in which LLMs may play a role in these areas.

LLMs as backbones: LLMs have demonstrated powerful capabilities in cross-modal tasks. Liu et al. [110] proposed to freeze most parameters of a pretrained GPT-2 model and fine-tuning only a limited number of parameters for cross-modal knowledge transfer from the feature space of LLMs to the wireless channels. They then applied the fine-tuned model to the uplink-to-downlink channel prediction problem and achieved an improved performance. In Ref. [185], the model was further extended to handle general space-time-frequency wireless channels. By pre-training over a massive channel dataset, the model can be generalized for channel prediction in various CSI configurations. Yu et al. [186] incorporated multimodal data and proposed a ChannelGPT model for wireless networks. In Refs. [[187], [188], [189]], fine-tuning LLM models are proposed to serve multiple physical layer tasks, leveraging the multitask learning capabilities of large AI models.

·•LLMs as hypernetworks: Zhang et al. [190] investigated the use of LLMs in the context of the least absolute shrinkage and selection operator (LASSO) problem. Although their focus was not directly on wireless systems, many estimation problems in wireless systems can be similarly formulated as sparse reconstruction tasks, making this study highly relevant. The proposed LLM-LASSO framework introduces a novel paradigm in which pretrained LLMs dynamically generate context-aware regularization parameters or feature selection masks by interpreting problem-specific metadata such as hypernetworks. LASSO estimators are particularly well suited for terahertz UM-MIMO channels because they can be represented as high-dimensional sparse vectors. It is valuable to investigate how contextual information regarding wireless propagation can enhance the regulation parameter generation of terahertz UM-MIMO channel estimators.

4.3.2. LLMs for optimization

Recent advancements in LLMs have facilitated their integration into optimization tasks, particularly in wireless communication systems. LLMs, which leverage their pre-trained reasoning capabilities and multimodal data processing, can offer a paradigm shift from conventional optimization frameworks. Unlike traditional methods which rely on delicate problem formulation and modeling by human experts and training task-specific DL models, LLMs can automatically interpret and formulate the problem from language-based and possibly multimodal descriptions and provide zero- or few-shot solutions to different optimization problems, such as power control [191,192], traffic prediction [192], and spectrum sensing [111], and so forth, without further retraining. Iterative prompting and chain-of-thought allow LLMs to refine their optimization decisions through feedback information. In addition, LLMs can be combined with model-driven DL-based optimization algorithms. They can play a role similar to that of hypernetworks in generating context-specific network parameters based on multimodal descriptions of the environment. Nazar et al. [193] proposed ENWAR, an LLM framework for multisensory data that employs retrieval-augmented generation for the real-time environmental perception of wireless networks. It can serve as a feature extractor to work hand in hand with LLM-based optimization algorithms. Owing to the channel characteristics of terahertz UM-MIMO systems, radio signals are easily subjected to blockage and misalignment. LLMs can utilize multimodal sensory data to enhance situational awareness and make optimization decisions based on their own intelligence without the aid of human experts.

4.3.3. LLMs for searching

Compared with deploying LLMs online for estimation and optimization problems, which must consider computational and memory complexity issues, it seems more practical to leverage LLMs to solve offline problems. Examples include large-scale wireless network planning (e.g., searching for the optimal antenna downtilt angles), facility localization (e.g., searching for the optimal number and positions of BSs), and sparse array design (e.g., searching for the optimal antenna position for a given aperture and number of antennas) in the context of terahertz UM-MIMO systems. These problems have several common characteristics. First, they are large-scale combinatorial problems involving integer variables, making them particularly difficult to solve as the system dimensions increase. Therefore, efficient heuristic-search algorithms are required. Second, these problems are typically solved once and for all. When an optimal solution is obtained offline, continuous updates are not required.

For such offline tasks, LLMs are especially promising because of their extensive knowledge base and ability to inspire the design of novel heuristics. For instance, a FunSearch paper from Google illustrates how LLMs can facilitate mathematical discoveries by navigating several candidate algorithms through a program search to uncover effective heuristics that can exceed the performance of the best human experts [112]. A similar methodology was adopted for challenges in communication systems, in a recent study on the LLM-guided search for deletion-correcting codes, where LLMs assist in identifying efficient error-correcting codes [194]. Considering the scale of terahertz UM-MIMO systems, efficient heuristics are extremely important. Similarly, by harnessing the capabilities of LLMs, the efficiency of combinatorial search problems that are prevalent in the deployment of terahertz-band wireless systems can be enhanced.

4.3.4. LLMs for network management

Recent advances in LLM agents have underscored a transformative trend toward autonomous and intelligent network management. For instance, Shen et al. [113] introduced an autonomous edge AI system that considered a client-edge-cloud hierarchical architecture. A generative pre-trained transformer (GPT)-based LLM is deployed in the cloud to understand natural language inputs and then plan and generate code. The system is designed to dynamically coordinate distributed AI models at the network edge to fulfill diverse user requirements, such as triggering federated learning tasks to update the models, thus demonstrating the capability of autonomously managing edge AI systems. Tong et al. [195] proposed the WirelessAgent framework, in which LLMs are leveraged as AI agents to facilitate intelligent network slicing in 6G. LLM agents are designed to process multimodal data and use techniques such as retrieval-augmented generation to facilitate intent understanding and decision making. In addition to the development of LLMs for network management, transformative innovations in network architecture are required to support the deployment of large AI models. Researchers at China Telecom proposed a new concept called “AI Flow,” which introduced a novel framework designed to efficiently deploy multimodal LLMs across heterogeneous network infrastructures [196]. They proposed a dynamic distribution of inference tasks based on the available resources, network conditions, and task requirements. In terahertz systems, the substantial bandwidth enables high-rate and low-latency wireless communications capable of supporting LLMs with extensive data throughput. Nevertheless, the blockages and dynamic propagation environments prevalent in the terahertz band introduce additional uncertainty. The exploration of LLM-based autonomous network management is important for intelligent interference management and handover control. Investigating autonomous link adaptation and network slicing in response to the diverse computational demands of AI inference tasks is also meaningful.

4.3.5. LLMs for protocol understanding

Recent efforts have also demonstrated the significant potential of LLMs for understanding complicated communication protocols and standard documents. Nikbakht et al. [197] provided a comprehensive dataset spanning 3rd Generation Partnership Project (3GPP) releases from Releases 8 to 19, which enabled the fine-tuning of LLMs for questions and reasoning tasks in the telecommunication domain. Zou et al. [114] and Bariah et al. [198] illustrated the effectiveness of fine-tuning pre-trained LLMs for accurately classifying and extracting information from 3GPP technical documents. In addition, Bornea et al. [199] and Yilma et al. [200] leveraged retrieval-augmented generation by integrating telecommunication-specific knowledge bases into an LLM processing pipeline. Collectively, these studies highlight a growing body of literature that improves the understanding of protocols in telecommunications. In the future, it will be meaningful to develop domain-specific LLMs to understand the protocols in terahertz-band wireless systems.

5. Future directions and open issues

After discussing the three research roadmaps, we offer the following takeaway messages covering lessons learned, future directions, and open issues.

5.1. Remarks on roadmap 1: Model-driven DL

The essence of model-driven DL is to leverage the expert knowledge inherent in the system and allow AI to focus only on bottleneck modules that are either “hard-to-compute” or “hard-to-model.” It is important not to “reinvent the wheel” by attempting to learn something that is already available in the system model. It is preferable to design appropriate loss functions and neural architectures tailored to the specific features and requirements of the problem. Always keeping the role of the “model” in mind is fundamental to the success of model-driven DL.

For future directions, on the one hand, it is important to keep tracking the development of signal processing and optimization algorithms and leverage state-of-the-art techniques as the basis for model-driven DL. This requires a deep understanding of both the specific problem at hand and in-depth knowledge of classical model-based algorithms. Furthermore, the computational and hardware costs associated with the deployment of model-driven DL must be carefully considered. It is important to design hardware-friendly algorithms considering the scale of terahertz UM-MIMO systems.

However, it is important to strengthen the connections between model-driven DL and CSI foundation models. Model-driven DL incorporates algorithm-specific designs to improve the convergence and performance of traditional algorithms. As mentioned previously, many components of transceiver algorithms share a common foundation, the wireless channel, which highlights the potential for connecting model-driven DL with CSI foundation models. Such an approach would allow a shared knowledge base to be adapted for specific tasks, thereby enhancing the overall performance. Furthermore, establishing these connections is crucial for advancing the integration of theoretical insights and practical applications in wireless communication.

5.2. Remarks on roadmap 2: CSI foundation models

The essence of CSI foundation models is to identify and separate the common ground of diverse tasks and focus on learning these “foundations.” The goal is to enable a single compact foundation model to contribute to the design of a wide variety of transceiver modules. A few frameworks may be instrumental in separating the foundations of different problems. One such example is Bayesian inference, where the prior and likelihood can be separated [201]. Denoising-based AMP and posterior sampling are two important examples in this category. Both are related to the denoiser of the wireless channels under Gaussian noise, which is further connected to the score function of the wireless channels. Therefore, tracking the latest developments in signal denoising and score estimation is important. Another example is the information-theoretic framework of neural dependence decomposition, which separates feature learning from feature usages [149], and learns universal features for various downstream tasks. It is worthwhile to follow these directions and dig deeper into the core of the CSI foundation models while discovering more applications in this promising field.

In addition, it is crucial to investigate the neural network architectures underpinning CSI foundation models. Although not discussed in detail here, it is meaningful to draw insights from the advanced neural architectures developed in the broader machine learning community to strengthen our understanding of these models. It is also worthwhile to keep an eye on the state-of-the-art developments of LLMs, as scaling CSI foundation models for real-world applications, such as terahertz UM-MIMO systems with multiple BSs and a large number of users, will likely encounter similar scalability and computational challenges. Therefore, learning from the evolution of LLM architectures could be instrumental in the successful large-scale deployment of CSI foundation models.

5.3. Remarks on roadmap 3: Applications of LLMs

Despite the empirical performance gains provided by LLMs, the underlying reasons why pretrained LLMs achieve success in wireless tasks remain unclear. Moreover, a critical open question is whether the inherent complexity of LLMs is disproportionate for wireless channels, given their highly structured and often low-dimensional characteristics. Furthermore, the cost of deploying LLMs in wireless systems remains a challenge, necessitating rigorous analysis of their memory efficiency, energy efficiency, sample complexity, and inference latency. When it comes to terahertz UM-MIMO systems, due to the enormous scale of the system, we must be more careful about these practical concerns, especially the “hard-to-compute” problems mentioned earlier. Finally, LLMs require substantial data for both training and fine-tuning, which often necessitates extensive measurement efforts and poses “hard-to-measure” challenges. Additionally, some telecommunication data are subject to privacy concerns. One promising research direction is to explore fine-tuning telecommunication-specific LLMs in wireless digital twins, where data are abundant, before deploying them in real-world environments with minimal tuning. Another promising approach is to investigate distributed or federated fine-tuning, which enables cooperative development across many telecommunication users, while respecting data privacy.

6. Conclusions

In this paper, we presented a methodological study on the application of AI to address the complex challenges in terahertz UM-MIMO systems. Although both AI and terahertz technologies are recognized as crucial enablers of 6G and beyond systems, their intersection remains in its early stages. Therefore, this study aimed to bridge this gap. The first half of this paper focuses on the system and channel characteristics of terahertz UM-MIMO systems, illustrating how these challenges have naturally prompted the application of AI. The second half outlines three research roadmaps—model-driven DL, CSI foundation models, and applications of LLMs—which are promising directions for developing AI solutions for terahertz UM-MIMO systems. We outlined the essential steps of these roadmaps, combining high-level concepts with detailed explanations, and analyzed several representative case studies. Substantial progress has been made in this emerging area of interdisciplinary research. Finally, we presented our vision for future research and identified key open issues.

CRediT authorship contribution statement

Wentao Yu: Writing - review & editing, Writing - original draft, Visualization, Data curation, Conceptualization. Hengtao He: Writing - review & editing, Supervision. Shenghui Song: Writing - review & editing, Supervision. Jun Zhang: Writing - review & editing, Supervision. Linglong Dai: Writing - review & editing, Supervision. Lizhong Zheng: Writing - review & editing, Supervision. Khaled B. Letaief: Writing - review & editing, Visualization, Supervision, Resources, Project administration, Funding acquisition.

Declaration of competing interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Acknowledgment

This work was supported in part by the Hong Kong Research Grant Council (16209023).

References

[1]

Cisco. Cisco annual internet report (2018-2023). Report. San Jose: Cisco Systems, Inc.; 2023.

[2]

K.B. Letaief, W. Chen, Y. Shi, J. Zhang, Y.J.A. Zhang. The roadmap to 6G: AI empowered wireless networks. IEEE Commun Mag, 57 (8) (2019), pp. 84-90.

[3]

Preparing for a cloud VR/AR future. Shenzhen: Huawei Technologies Co., Ltd.; 2017.

[4]

K.B. Letaief, Y. Shi, J. Lu, J. Lu. Edge artificial intelligence for 6G: vision, enabling technologies, and applications. IEEE J Sel Areas Commun, 40 (1) (2021), pp. 5-36.

[5]

X. You, C.X. Wang, J. Huang, X. Gao, Z. Zhang, M. Wang, et al. Towards 6G wireless communication networks: vision, enabling technologies, and new paradigm shifts. Sci China Inf Sci, 64 (1) (2021), Article 110301.

[6]

S. Dang, O. Amin, B. Shihada, M.S. Alouini. What should 6G be?. Nat Electron, 3 (1) (2020), pp. 20-29.

[7]

Future technology trends of terrestrial International Mobile Telecommunications systems towards 2030 and beyond. Report. Geneva: International Telecommunication Union; 2022.

[8]

5G Americas. ITU’s IMT-2030 vision: navigating towards 6G in the Americas. Bellevue: 5G Americas; 2024.

[9]

IMT-2030 (6G) Promotion Group. White paper on 6G vision and candidate technologies. Report. Shenzhen: IMT-2030 (6G) Promotion Group; 2021.

[10]

T.S. Rappaport, Y. Xing, O. Kanhere, S. Ju, A. Madanayake, S. Mandal, et al. Wireless communications and applications above 100 GHz: opportunities and challenges for 6G and beyond. IEEE Access, 7 (2019), pp. 78729-78757.

[11]

Federal Communications Commission (FCC). FCC takes steps to open spectrum horizons for new services and technologies. Neil: Federal Communications Commission; 2018.

[12]

V. Petrov, T. Kurner, I. Hosako. IEEE 802.15.3d: first standardization efforts for sub-terahertz band communications toward 6G. IEEE Commun Mag, 58 (11) (2020), pp. 28-33.

[13]

T.L. Marzetta. Noncooperative cellular wireless with unlimited numbers of base station antennas. IEEE Trans Wirel Commun, 9 (11) (2010), pp. 3590-3600.

[14]

Björnson E, Chae CB, Heath Jr RW, Marzetta TL, Mezghani A, Sanguinetti L, et al. Towards 6G MIMO: massive spatial multiplexing, dense arrays, and interplay between electromagnetics and processing. 2024. arXiv:240102844.

[15]

I.F. Akyildiz, J.M. Jornet. Realizing ultra-massive MIMO (1024 × 1024) communication in the (0.06-10) terahertz band. Nano Commun Netw, 8 (2016), pp. 46-54.

[16]

A. Faisal, H. Sarieddeen, H. Dahrouj, T.Y. Al-Naffouri, M.S. Alouini. Ultramassive MIMO systems at terahertz bands: prospects and challenges. IEEE Veh Technol Mag, 15 (4) (2020), pp. 33-42.

[17]

B. Ning, Z. Tian, W. Mei, Z. Chen, C. Han, S. Li, et al. Beamforming technologies for ultra-massive MIMO in terahertz communications. IEEE Open J Commun Soc, 4 (2023), pp. 614-658.

[18]

H. Sarieddeen, M.S. Alouini, T.Y. Al-Naffouri. An overview of signal processing techniques for terahertz communications. Proc IEEE, 109 (10) (2021), pp. 1628-1665.

[19]

H. Chen, H. Sarieddeen, T. Ballal, H. Wymeersch, M.S. Alouini, T.Y. Al-Naffouri. A tutorial on terahertz-band localization for 6G communication systems. IEEE Commun Surv Tutor, 24 (3) (2022), pp. 1780-1815.

[20]

A.M. Elbir, K.V. Mishra, S. Chatzinotas, M. Bennis. Terahertz-band integrated sensing and communications: challenges and opportunities. IEEE Aerosp Electron Syst Mag, 39 (12) (2024), pp. 38-49.

[21]

C. Han, Y. Wu, Z. Chen, Y. Chen, G. Wang. THz ISAC: a physical-layer perspective of terahertz integrated sensing and communication. IEEE Commun Mag, 62 (2) (2024), pp. 102-108.

[22]

F. Lemic, S. Abadal, W. Tavernier, P. Stroobant, D. Colle, E. Alarcon, et al. Survey on terahertzx nanocommunication and networking: a top-down perspective. IEEE J Sel Areas Commun, 39 (6) (2021), pp. 1506-1543.

[23]

J.M. Jornet, I.F. Akyildiz. Graphene-based plasmonic nano-antenna for terahertz band communication in nanonetworks. IEEE J Sel Areas Commun, 31 (12) (2013), pp. 685-694.

[24]

H.J. Song, N. Lee. Terahertz communications: challenges in the next decade. IEEE Trans Terahertz Sci Technol, 12 (2) (2021), pp. 105-117.

[25]

H. Do, S. Cho, J. Park, H.J. Song, N. Lee, A. Lozano. Terahertz line-of-sight MIMO communication: theory and practical challenges. IEEE Commun Mag, 59 (3) (2021), pp. 104-109.

[26]

N. Yang, A. Shafie. Terahertz communications for massive connectivity and security in 6G and beyond era. IEEE Commun Mag, 62 (2) (2024), pp. 72-78.

[27]

A. Shafie, N. Yang, C. Han, J.M. Jornet, M. Juntti, T. Kürner. Terahertz communications for 6G and beyond wireless networks: challenges, key advancements, and opportunities. IEEE Netw, 37 (3) (2023), pp. 162-169.

[28]

C. Han, Y. Chen, L. Yan, Z. Chen, L. Dai. Cross far- and near-field wireless communications in terahertz ultra-large antenna array systems. IEEE Wirel Commun, 31 (3) (2024), pp. 148-154.

[29]

Yu W, Shen Y, He H, Yu X, Zhang J, Letaief KB. Hybrid far- and near-field channel estimation for THz ultra-massive MIMO via fixed point networks. In:Proceedings of IEEE Global Communications Conference; 2022 Dec 4-8; Rio de Janeiro, Brazil. New York City: IEEE; 2022. p. 5384-9.

[30]

J. Wang, C.X. Wang, J. Huang, H. Wang, X. Gao. A general 3D space-time-frequency non-stationary THz channel model for 6G ultra-massive MIMO wireless communication systems. IEEE J Sel Areas Commun, 39 (6) (2021), pp. 1576-1589.

[31]

E.D. Carvalho, A. Ali, A. Amiri, M. Angjelichinoski, R.W. Heath. Non-stationarities in extra-large-scale massive MIMO. IEEE Wirel Commun, 27 (4) (2020), pp. 74-80.

[32]

B. Wang, M. Jian, F. Gao, G.Y. Li, H. Lin. Beam squint and channel estimation for wideband mmWave massive MIMO-OFDM systems. IEEE Trans Signal Process, 67 (23) (2019), pp. 5893-5908.

[33]

J. Zhang, X. Yu, K.B. Letaief. Hybrid beamforming for 5G and beyond millimeter-wave systems: a holistic view. IEEE Open J Commun Soc, 1 (2020), pp. 77-91.

[34]

C. Lin, G.Y. Li. Terahertz communications: an array-of-subarrays solution. IEEE Commun Mag, 54 (12) (2016), pp. 124-131.

[35]

Wu Y, Koch JD, Vossiek M, Schober R, Gerstacker W. ML detection without CSI for constant-weight codes in THz communications with strong phase noise. In:Proceedings of IEEE Global Communications Conference; 2022 Dec 4-8; Rio de Janeiro, Brazil. New York City: IEEE; 2022. p. 831-6.

[36]

S. Bicais, J.B. Dore. Design of digital communications for strong phase noise channels. IEEE Open J Veh Technol, 1 (2020), pp. 227-243.

[37]

Z. Sha, Z. Wang. Channel estimation and equalization for terahertz receiver with RF impairments. IEEE J Sel Areas Commun, 39 (6) (2021), pp. 1621-1635.

[38]

W. Yu, Y. Shen, H. He, X. Yu, S. Song, J. Zhang, et al. An adaptive and robust deep learning framework for THz ultra-massive MIMO channel estimation. IEEE J Sel Top Signal Process, 17 (4) (2023), pp. 761-776.

[39]

Z. Wang, J. Zhang, H. Du, D. Niyato, S. Cui, B. Ai, et al. A tutorial on extremely large-scale MIMO for 6G: fundamentals, signal processing, and applications. IEEE Commun Surv Tutor, 26 (3) (2024), pp. 1560-1605.

[40]

E. Björnson, L. Sanguinetti, H. Wymeersch, J. Hoydis, T.L. Marzetta. Massive MIMO is a reality—what is next? Five promising research directions for antenna arrays. Digit Signal Process, 94 (2019), pp. 3-20.

[41]

J.C. Marinello, T. Abrão, A. Amiri, E. de Carvalho, P. Popovski. Antenna selection for improving energy efficiency in XL-MIMO systems. IEEE Trans Veh Technol, 69 (11) (2020), pp. 13305-13318.

[42]

S. Ye, M. Xiao, M.W. Kwan, Z. Ma, Y. Huang, G. Karagiannidis, et al. Extremely large aperture array (ELAA) communications: foundations, research advances and challenges. IEEE Open J Commun Soc, 5 (2024), pp. 7075-7120.

[43]

Amiri A, Angjelichinoski M, de Carvalho E, Heath RW. Extremely large aperture massive MIMO:low complexity receiver architectures. In:Proceedings of the 2018 IEEE Globecom Workshops; 2018 Dec 9-13; Abu Dhabi, United Arab Emirates. New York City: IEEE; 2018. p. 1-6.

[44]

Björnson E, Kara F, Kolomvakis N, Kosasih A, Ramezani P, Salman MB. Enabling 6G performance in the upper mid-band by transitioning from massive to gigantic MIMO. 2024. arXiv:240705630.

[45]

T. O’Shea, J. Hoydis. An introduction to deep learning for the physical layer. IEEE Trans Cogn Commun Netw, 3 (4) (2017), pp. 563-575.

[46]

N. Ye, S. Miao, J. Pan, Q. Ouyang, X. Li, X. Hou. Artificial intelligence for wireless physical-layer technologies (AI4PHY): a comprehensive survey. IEEE Trans Cogn Commun Netw, 10 (3) (2024), pp. 729-755.

[47]

Y. Shi, L. Lian, Y. Shi, Z. Wang, Y. Zhou, L. Fu. Machine learning for large-scale optimization in 6G wireless networks. IEEE Commun Surv Tutor, 25 (4) (2023), pp. 2088-2132.

[48]

E. Bjornson, P. Giselsson. Two applications of deep learning in the physical layer of communication systems. IEEE Signal Process Mag, 37 (5) (2020), pp. 134-140.

[49]

Y. Bengio, A. Lodi, A. Prouvost. Machine learning for combinatorial optimization: a methodological tour d’horizon. Eur J Oper Res, 290 (2) (2021), pp. 405-421.

[50]

H. He, S. Jin, C.K. Wen, F. Gao, G.Y. Li, Z. Xu. Model-driven deep learning for physical layer communications. IEEE Wirel Commun, 26 (5) (2019), pp. 77-83.

[51]

Z. Qin, H. Ye, G.Y. Li, B.H.F. Juang. Deep learning in physical layer communications. IEEE Wirel Commun, 26 (2) (2019), pp. 93-99.

[52]

N. Van Huynh, J. Wang, H. Du, D.T. Hoang, D. Niyato, D.N. Nguyen, et al. Generative AI for physical layer communications: a survey. IEEE Trans Cogn Commun Netw, 10 (3) (2024), pp. 706-728.

[53]

J.G. Andrews, T.E. Humphreys, T. Ji. 6G takes shape. IEEE BITS Inf Theory Mag, 4 (1) (2024), pp. 2-24.

[54]

W. Yu, F. Sohrabi, T. Jiang. Role of deep learning in wireless communications. IEEE BITS Inf Theory Mag, 2 (2) (2022), pp. 56-72.

[55]

H. He, C.K. Wen, S. Jin, G.Y. Li. Model-driven deep learning for MIMO detection. IEEE Trans Signal Process, 68 (2020), pp. 1702-1715.

[56]

R. Shafin, L. Liu, V. Chandrasekhar, H. Chen, J. Reed, J.C. Zhang. Artificial intelligence-enabled cellular networks: a critical path to beyond-5G and 6G. IEEE Wirel Commun, 27 (2) (2020), pp. 212-217.

[57]

T. Raviv, N. Shlezinger. Data augmentation for deep receivers. IEEE Trans Wirel Commun, 22 (11) (2023), pp. 8259-8274.

[58]

N. Soltani, K. Sankhe, J. Dy, S. Ioannidis, K. Chowdhury. More is better: data augmentation for channel-resilient RF fingerprinting. IEEE Commun Mag, 58 (10) (2020), pp. 66-72.

[59]

I.F. Akyildiz, C. Han, Z. Hu, S. Nie, J.M. Jornet. Terahertz band communication: an old problem revisited and research directions for the next decade. IEEE Trans Commun, 70 (6) (2022), pp. 4250-4285.

[60]

D. Bodet, J. Hall, A. Masihi, N. Thawdar, T. Melodia, F. Restuccia, et al. Data signals for deep learning applications in terahertz communications. Comput Netw, 254 (2024), Article 110800.

[61]

J. Hall, J.M. Jornet, N. Thawdar, T. Melodia, F. Restuccia. Deep learning at the physical layer for adaptive terahertz communications. IEEE Trans Terahertz Sci Technol, 13 (2) (2023), pp. 102-112.

[62]

S. Tarboush, H. Sarieddeen, H. Chen, M.H. Loukil, H. Jemaa, M.S. Alouini, et al. TeraMIMO: a channel simulator for wideband ultra-massive MIMO terahertz communications. IEEE Trans Veh Technol, 70 (12) (2021), pp. 12325-12341.

[63]

C. Han, W. Gao, N. Yang, J.M. Jornet. Molecular absorption effect: a double-edged sword of terahertz communications. IEEE Wirel Commun, 30 (4) (2022), pp. 140-146.

[64]

O.E. Ayach, S. Rajagopal, S. Abu-Surra, Z. Pi, R.W. Heath. Spatially sparse precoding in millimeter wave MIMO systems. IEEE Trans Wirel Commun, 13 (3) (2014), pp. 1499-1513.

[65]

X. Yu, J.C. Shen, J. Zhang, K.B. Letaief. Alternating minimization algorithms for hybrid precoding in millimeter wave MIMO systems. IEEE J Sel Top Signal Process, 10 (3) (2016), pp. 485-500.

[66]

Y. Huang, Y. Li, H. Ren, J. Lu, W. Zhang. Multi-panel MIMO in 5G. IEEE Commun Mag, 56 (3) (2018), pp. 56-61.

[67]

J. Dai, A. Liu, V.K.N. Lau. FDD massive MIMO channel estimation with arbitrary 2D-array geometry. IEEE Trans Signal Process, 66 (10) (2018), pp. 2584-2599.

[68]

F. Gao, B. Wang, C. Xing, J. An, G.Y. Li. Wideband beamforming for hybrid massive MIMO terahertz communications. IEEE J Sel Areas Commun, 39 (6) (2021), pp. 1725-1740.

[69]

Tan J, Dai L. Delay-phase precoding for THz massive MIMO with beam split. In: Proceedings of the 2019 IEEE Global Communications Conference; 2019 Dec 9-13; Big Island, HI, USA. New York City: IEEE; 2019. p. 1-6.

[70]

J. Tan, L. Dai. THz precoding for 6G: challenges, solutions, and opportunities. IEEE Wirel Commun, 30 (4) (2023), pp. 132-138.

[71]

V.V. Ratnam, J. Mo, A. Alammouri, B.L. Ng, J. Zhang, A.F. Molisch. Joint phase-time arrays: a paradigm for frequency-dependent analog beamforming in 6G. IEEE Access, 10 (2022), pp. 73364-73377.

[72]

T. Zheng, M. Cui, Z. Wu, L. Dai. Near-field wideband beam training based on distance-dependent beam split. IEEE Trans Wirel Commun, 24 (2) (2024), pp. 1-14.

[73]

J. Tan, L. Dai. Wideband beam tracking in THz massive MIMO systems. IEEE J Sel Areas Commun, 39 (6) (2021), pp. 1693-1710.

[74]

D. Serghiou, M. Khalily, T.W. Brown, R. Tafazolli. Terahertz channel propagation phenomena, measurement techniques and modeling for 6G wireless communication applications: a survey, open challenges and future research directions. IEEE Commun Surv Tutor, 24 (4) (2022), pp. 1957-1996.

[75]

C. Han, Y. Wang, Y. Li, Y. Chen, N.A. Abbasi, T. Kurner, et al. Terahertz wireless channels: a holistic survey on measurement, modeling, and analysis. IEEE Commun Surv Tutor, 24 (3) (2022), pp. 1670-1707.

[76]

Petrov V, Jornet JM, Singh A. Near-field 6G networks:why mobile terahertz communications MUST operate in the near field. In:Proceedings of IEEE Global Communications Conference; 2023 Dec 4-8; Kuala Lumpur, Malaysia. New York City: IEEE; 2023. p. 3983-9.

[77]

J.S. Jiang, M.A. Ingram. Spherical-wave model for short-range MIMO. IEEE Trans Commun, 53 (9) (2005), pp. 1534-1541.

[78]

X. Wei, L. Dai. Channel estimation for extremely large-scale massive MIMO: far-field, near-field, or hybrid-field?. IEEE Commun Lett, 26 (1) (2022), pp. 177-181.

[79]

S. Tarboush, A. Ali, T.Y. Al-Naffouri. Cross-field channel estimation for ultra massive-MIMO THz systems. IEEE Trans Wirel Commun, 23 (8) (2024), pp. 8619-8653.

[80]

L. Yan, C. Han, J. Yuan. A dynamic array-of-subarrays architecture and hybrid precoding algorithms for terahertz wireless communications. IEEE J Sel Areas Commun, 38 (9) (2020), pp. 2041-2056.

[81]

W. Gao, Y. Chen, C. Han, Z. Chen. Distance-adaptive absorption peak modulation (DA-APM) for terahertz covert communications. IEEE Trans Wirel Commun, 20 (3) (2021), pp. 2064-2077.

[82]

Y. Heng, J. Mo, J.G. Andrews. Learning site-specific probing beams for fast mmWave beam alignment. IEEE Trans Wirel Commun, 21 (8) (2022), pp. 5785-5800.

[83]

Hoydis J, Cammerer S, Aoudia FA, Vem A, Binder N, Marcus G, et al. Sionna: an open-source library for next-generation physical layer research. 2022. arXiv:220311854.

[84]

M. Cui, Z. Wu, Y. Lu, X. Wei, L. Dai. Near-field MIMO communications for 6G: fundamentals, challenges, potentials, and future directions. IEEE Commun Mag, 61 (1) (2023), pp. 40-46.

[85]

W. Yu, Y. Ma, H. He, S. Song, J. Zhang, K.B. Letaief. Deep learning for near-field XL-MIMO transceiver design: principles and techniques. IEEE Commun Mag, 63 (1) (2024), pp. 52-58.

[86]

Forsch C, Alrabadi O, Brueck S, Gerstacker W. Phase noise robust terahertz communications. In:Proceedings of the 2022 IEEE 95th Vehicular Technology Conference; 2022 Jun 19-22; Helsinki, Finland. New York City: IEEE; 2022. p. 1-6.

[87]

W.L. Chan, M.L. Moravec, R.G. Baraniuk, D.M. Mittleman. Terahertz imaging with compressed sensing and phase retrieval. Opt Lett, 33 (9) (2008), pp. 974-976.

[88]

Cao R, He H, Yu X, Song S, Huang K, Zhang J, et al. Joint channel estimation and cooperative localization for near-field ultra-massive MIMO. 2023. arXiv:231213683.

[89]

S. Liu, X. Yu, Z. Gao, J. Xu, D.W.K. Ng, S. Cui. Sensing-enhanced channel estimation for near-field XL-MIMO systems. IEEE J Sel Areas Commun, 43 (3) (2024), pp. 628-643.

[90]

Cao R, Yu W, He H, Yu X, Song S, Zhang J. Newtonized near-field channel estimation for ultra-massive MIMO systems. In: Proceedings of the 2024 IEEE Wireless Communications and Networking Conference; 2024 Apr 21-24; Dubai, United Arab Emirates. New York City: IEEE; 2024. p. 1-6.

[91]

Lee T, Park J, Kim H, Andrews JG, Generating high dimensional user-specific wireless channels using diffusion models. 2024. arXiv:240903924.

[92]

Chi G, Yang Z, Wu C, Xu J, Gao Y, Liu Y, et al. RF-diffusion:radio signal generation via time-frequency diffusion. In:Proceedings of the 30th Annual International Conference on Mobile Computing and Networking; 2024 Nov 18-22; Washington, DC, USA. New York City: Association for Computing Machinery (ACM); 2024. p. 77-92.

[93]

M. Arvinte, J.I. Tamir. MIMO channel estimation using score-based generative models. IEEE Trans Wirel Commun, 22 (6) (2023), pp. 3698-3713.

[94]

W. Yu, H. He, X. Yu, S. Song, J. Zhang, R. Murch, et al. Bayes-optimal unsupervised learning for channel estimation in near-field holographic MIMO. IEEE J Sel Top Signal Process, 18 (4) (2024), pp. 714-729.

[95]

Olutayo T, Champagne B. Score-based generative modeling for MIMO detection without knowledge of noise statistics. In: Proceedings of the 2023 IEEE 34th Annual International Symposium on Personal, Indoor and Mobile Radio Communications; 2023 Sep 5-8; Toronto, ON, Canada. New York City: IEEE; 2023. p. 1-7.

[96]

K. He, L. He, L. Fan, Y. Deng, G.K. Karagiannidis, A. Nallanathan. Learning-based signal detection for MIMO systems with unknown noise statistics. IEEE Trans Commun, 69 (5) (2021), pp. 3025-3038.

[97]

E. Balevi, J.G. Andrews. Unfolded hybrid beamforming with GAN compressed ultra-low feedback overhead. IEEE Trans Wirel Commun, 20 (12) (2021), pp. 8381-8392.

[98]

Jayashankar T, Lee GCF, Lancho A, Weiss A, Polyanskiy Y, Wornell G. Score-based source separation with applications to digital communication signals. In:Proceedings of the 37th International Conference on Advances in Neural Information Processing Systems; 2023 Dec 10-16; New Orleans, LA, USA. Red Hook: Curran Associates Inc.; 2023. p. 5092-125.

[99]

Y. Ma, Y. Shen, X. Yu, J. Zhang, S.H. Song, K.B. Letaief. Learn to communicate with neural calibration: scalability and generalization. IEEE Trans Wirel Commun, 21 (11) (2022), pp. 9947-9961.

[100]

N.T. Nguyen, M. Ma, O. Lavi, N. Shlezinger, Y.C. Eldar, S.A. Lee. Deep unfolding hybrid beamforming designs for THz massive MIMO systems. IEEE Trans Signal Process, 71 (2023), pp. 3788-3804.

[101]

H. He, R. Wang, W. Jin, S. Jin, C.K. Wen, G.Y. Li. Beamspace channel estimation for wideband millimeter-wave MIMO: a model-driven unsupervised learning approach. IEEE Trans Wirel Commun, 22 (3) (2022), pp. 1808-1822.

[102]

H. He, X. Yu, J. Zhang, S. Song, K.B. Letaief. Message passing meets graph neural networks: a new paradigm for massive MIMO systems. IEEE Trans Wirel Commun, 23 (5) (2024), pp. 4709-4723.

[103]

G. Liu, Z. Hu, L. Wang, H. Zhang, J. Xue, M. Matthaiou. A hypernetwork based framework for non-stationary channel prediction. IEEE Trans Veh Technol, 73 (6) (2024), pp. 8338-8351.

[104]

Y. Ding, B.D. Rao. Dictionary learning-based sparse channel representation and estimation for FDD massive MIMO systems. IEEE Trans Wirel Commun, 17 (8) (2018), pp. 5437-5451.

[105]

Wen C, Tong J, Hu Y, Lin Z, Zhang J. WRF-GS:wireless radiation field reconstruction with 3D Gaussian splatting. In:Proceedings of IEEE Conference on Computer Communications; 2025 May 19-22; London, UK. New York City: IEEE; 2025. p. 1-10.

[106]

Meng X, Kabashima Y. Quantized compressed sensing with score-based generative models. In:Proceedings of the Eleventh International Conference on Learning Representations; 2023 May 1-5; Kigali, Rwanda. New York City: IEEE; 2023.

[107]

Elata N, Michaeli T, Elad M. Adaptive compressed sensing with diffusion-based posterior sampling. In: Proceedings of the European Conference on Computer Vision; 2024 Sep 29-Oct 4; Milan, Italy. Berlin:Springer; 2025. p. 290-308.

[108]

N. Zilberstein, C. Dick, R. Doost-Mohammady, A. Sabharwal, S. Segarra. Annealed Langevin dynamics for massive MIMO detection. IEEE Trans Wirel Commun, 22 (6) (2022), pp. 3762-3776.

[109]

Wang Z, Zhou Y, Shi Y, Letaief KB. Federated fine-tuning for pre-trained foundation models over wireless networks. 2024. arXiv:240702924.

[110]

B. Liu, X. Liu, S. Gao, X. Cheng, L. Yang. LLM4CP: adapting large language models for channel prediction. J Commun Inf Netw, 9 (2) (2024), pp. 113-125.

[111]

J. Shao, J. Tong, Q. Wu, W. Guo, Z. Li, Z. Lin, et al. WirelessLLM: empowering large language models towards wireless intelligence. J Commun Inf Netw, 9 (2) (2024), pp. 99-112.

[112]

B. Romera-Paredes, M. Barekatain, A. Novikov, M. Balog, M.P. Kumar, E. Dupont, et al. Mathematical discoveries from program search with large language models. Nature, 625 (7995) (2024), pp. 468-475.

[113]

Y. Shen, J. Shao, X. Zhang, Z. Lin, H. Pan, D. Li, et al. Large language models empowered autonomous edge AI for connected intelligence. IEEE Commun Mag, 62 (10) (2024), pp. 140-146.

[114]

Zou H, Zhao Q, Tian Y, Bariah L, Bader F, Lestable T, et al. TelecomGPT: a framework to build telecom-specfic large language models. 2024. arXiv:2407.09424.

[115]

K. Hornik, M. Stinchcombe, H. White. Multilayer feedforward networks are universal approximators. Neural Netw, 2 (5) (1989), pp. 359-366.

[116]

H. Ye, G.Y. Li, B.H. Juang. Power of deep learning for channel estimation and signal detection in OFDM systems. IEEE Wirel Commun Lett, 7 (1) (2018), pp. 114-117.

[117]

H. Sun, X. Chen, Q. Shi, M. Hong, X. Fu, N.D. Sidiropoulos. Learning to optimize: training deep neural networks for interference management. IEEE Trans Signal Process, 66 (20) (2018), pp. 5438-5453.

[118]

H. He, C.K. Wen, S. Jin, G.Y. Li. Deep learning-based channel estimation for beamspace mmWave massive MIMO systems. IEEE Wirel Commun Lett, 7 (5) (2018), pp. 852-855.

[119]

J. Zhang, C.K. Wen, L. Liang, S. Jin. Universal model-driven deep learning for MIMO transceiver design. IEEE Commun Mag, 62 (4) (2023), pp. 74-80.

[120]

Ma Y, Yu W, Yu X, Zhang J, Song S, Letaief KB. Lightweight and flexible deep equilibrium learning for CSI feedback in FDD massive MIMO. In: Proceedings of the 2024 IEEE International Conference on Machine Learning for Communication and Networking; 2024 May 5-8; Stockholm, Sweden. New York City: IEEE; 2024. p. 299-304.

[121]

H. Bauschke, P. Combettes. Convex analysis and monotone operator theory in Hilbert spaces. Springer, Berlin (2011).

[122]

Fung SW, Heaton H, Li Q, McKenzie D, Osher S, Yin W. JFB:Jacobian-free backpropagation for implicit networks. In:Proceedings of the 36th AAAI Conference on Artificial Intelligence; 2022 Feb 22-Mar 1; online. Washington, DC: Association for the Advancement of Artificial Intelligence (AAAI); 2022. p. 6648-56.

[123]

J. Gao, X. Chen, G.Y. Li. Deep unfolding based channel estimation for wideband terahertz near-field massive MIMO systems. Front Inf Technol Electron Eng, 25 (8) (2024), pp. 1162-1172.

[124]

C.A. Metzler, A. Maleki, R.G. Baraniuk. From denoising to compressed sensing. IEEE Trans Inf Theory, 62 (9) (2016), pp. 5117-5144.

[125]

Q. Shi, M. Razaviyayn, Z.Q. Luo, C. He. An iteratively weighted MMSE approach to distributed sum-utility maximization for a MIMO interfering broadcast channel. IEEE Trans Signal Process, 59 (9) (2011), pp. 4331-4340.

[126]

Zhao W, Han C, Song HJ, Björnson E, DNN based two-stage compensation algorithm for THz hybrid beamforming with imperfect hardware. 2024. arXiv:2411.14699.

[127]

He H, Wen CK, Jin S. Generalized expectation consistent signal recovery for nonlinear measurements. In: Proceedings of the 2017 IEEE International Symposium on Information Theory; 2017 Jun 25-30; Aachen, Germany. New York City: IEEE; 2017. p. 2333-7.

[128]

H. Huang, W. Xia, J. Xiong, J. Yang, G. Zheng, X. Zhu. Unsupervised learning-based fast beamforming design for downlink MIMO. IEEE Access, 7 (2018), pp. 7599-7605.

[129]

C.M. Stein. Estimation of the mean of a multivariate normal distribution. Ann Stat, 9 (6) (1981), pp. 1135-1151.

[130]

Y.C. Eldar. Generalized SURE for exponential families: applications to regularization. IEEE Trans Signal Process, 57 (2) (2009), pp. 471-481.

[131]

Yu W, He H, Yu X, Song S, Zhang J, Letaief KB. Blind performance prediction for deep learning based ultra-massive MIMO channel estimation. In:Proceedings of IEEE International Conference on Communications; 2023 May 28-Jun 1; Rome, Italy. New York City: IEEE; 2023. p. 2613-8.

[132]

Tian H, Lian L. GSURE-based unsupervised deep equilibrium model learning for large-scale channel estimation. In:Proceedings of IEEE Global Communications Conference; 2024 Dec 8-12; Cape Town, South Africa. New York City: IEEE; 2024. p. 1-6.

[133]

M. Raphan, E.P. Simoncelli. Least squares estimation without priors or supervision. Neural Comput, 23 (2) (2011), pp. 374-420.

[134]

Y. Shen, Y. Shi, J. Zhang, K.B. Letaief. Graph neural networks for scalable radio resource management: architecture design and theoretical analysis. IEEE J Sel Areas Commun, 39 (1) (2021), pp. 101-115.

[135]

Y. Shen, J. Zhang, S.H. Song, K.B. Letaief. Graph neural networks for wireless communications: from theory to practice. IEEE Trans Wirel Commun, 22 (5) (2023), pp. 3554-3569.

[136]

He H, Kosasih A, Yu X, Zhang J, Song H, Hardjawana W. GNN-enhanced approximate message passing for massive/ultra-massive MIMO detection. In: Proceedings of the 2023 IEEE Wireless Communications and Networking Conference; 2023 Mar 26-29; Glasgow, UK. New York City: IEEE; 2023. p. 1-6.

[137]

S. He, S. Xiong, Y. Ou, J. Zhang, J. Wang, Y. Huang, et al. An overview on the application of graph neural networks in wireless networks. IEEE Open J Commun Soc, 2 (2021), pp. 2547-2565.

[138]

Ha D, Dai A, Le QV. Hypernetworks. 2016. arXiv:160909106.

[139]

Jin W, He H, Wen CK, Jin S, Li GY. Adaptive channel estimation based on model-driven deep learning for wideband mmWave systems. In: Proceedings of the 2021 IEEE Global Communications Conference; 2021 Dec 7-11; Madrid, Spain. New York City: IEEE; 2021. p. 1-6.

[140]

Xie S, He H, Li H, Song S, Zhang J, Zhang YJA, et al. Deep learning-based adaptive joint source-channel coding using hypernetworks. In: Proceedings of the 2024 IEEE International Mediterranean Conference on Communications and Networking; 2024 Jul 8-11; Madrid, Spain. New York City: IEEE; 2024. p. 191-6.

[141]

M. Cui, L. Dai. Channel estimation for extremely large-scale MIMO: far-field or near-field?. IEEE Trans Commun, 70 (4) (2022), pp. 2663-2677.

[142]

Zhao X, An Z, Pan Q, Yang L. NeRF2:neural radio-frequency radiance fields. In: Proceedings of the 29th Annual International Conference on Mobile Computing and Networking; 2023 Oct 2-6; Madrid, Spain. New York City: Association for Computing Machinery (ACM); 2023. p. 1-15.

[143]

H. Zhang, N. Shlezinger, F. Guidi, D. Dardari, M.F. Imani, Y.C. Eldar. Beam focusing for near-field multiuser MIMO communications. IEEE Trans Wirel Commun, 21 (9) (2022), pp. 7476-7490.

[144]

Z. Wan, Z. Gao, F. Gao, M. Di Renzo, M.S. Alouini. Terahertz massive MIMO with holographic reconfigurable intelligent surfaces. IEEE Trans Commun, 69 (7) (2021), pp. 4732-4750.

[145]

A.M. Elbir, K.V. Mishra, S. Chatzinotas. Terahertz-band joint ultra-massive MIMO radar-communications: model-based and model-free hybrid beamforming. IEEE J Sel Top Signal Process, 15 (6) (2021), pp. 1468-1483.

[146]

J. Ma, L. Ping. Orthogonal AMP. IEEE Access, 5 (2017), pp. 2020-2033.

[147]

J. Cespedes, P.M. Olmos, M. Sanchez-Fernandez, F. Perez-Cruz. Expectation propagation detection for high-order high-dimensional MIMO systems. IEEE Trans Commun, 62 (8) (2014), pp. 2840-2849.

[148]

S. Wu, L. Kuang, Z. Ni, J. Lu, D. Huang, Q. Guo. Low-complexity iterative detection for large-scale multiuser MIMO-OFDM systems using approximate message passing. IEEE J Sel Top Signal Process, 8 (5) (2014), pp. 902-915.

[149]

X. Xu, L. Zheng. Neural feature learning in function space. J Mach Learn Res, 25 (142) (2024), pp. 1-76.

[150]

Xu X, Zheng L. Multiuser detection with neural feature learning. In:Proceedings of IEEE Military Communications Conference; 2024 Oct 28-Nov 1; Washington, DC, USA. New York City: IEEE; 2024. p. 715-20.

[151]

Xu X, Zheng L, Agrawal I. Neural feature learning for engineering problems. In: Proceedings of the 59th Annual Allerton Conference on Communication, Control, and Computing; 2023 Sep 26-29; Monticello, IL, USA. New York City: IEEE; 2023. p. 1-8.

[152]

Song Y, Sohl-Dickstein J, Kingma DP, Kumar A, Ermon S, Poole B. Score-based generative modeling through stochastic differential equations. In:Proceedings of the International Conference on Learning Representations; 2021 May 3-7; online. Trier: dblp Computer Science Bibliography; 2021. p. 37799-812.

[153]

Ho J, Jain A, Abbeel P. Denoising diffusion probabilistic models. In:Proceedings of the 34th International Conference on Neural Information Processing Systems; 2020 Dec 6-12; Vancouver, BC, Canada. Red Hook: Curran Associates Inc.; 2020. p. 6840-51.

[154]

Alikhani S, Charan G, Alkhateeb A. Large wireless model (LWM): a foundation model for wireless channels. 2024. arXiv:241108872.

[155]

Song Y, Ermon S. Generative modeling by estimating gradients of the data distribution. In:Proceedings of the 33rd International Conference on Neural Information Processing Systems; 2019 Dec 8-14; Vancouver, BC, Canada. Red Hook: Curran Associates Inc.; 2019. p. 11918-30.

[156]

P. Vincent. A connection between score matching and denoising autoencoders. Neural Comput, 23 (7) (2011), pp. 1661-1674.

[157]

Cai C, Yuan X, Zhang YJA. Score-based turbo message passing for plug-and-play compressive image recovery. 2025. arXiv:2503.22140.

[158]

Rombach R, Blattmann A, Lorenz D, Esser P, Ommer B. High-resolution image synthesis with latent diffusion models. In:Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition; 2023 Jun 17-24; Vancouver, BC, Canada. New York City: IEEE; 2022. p. 10684-95.

[159]

Yu W, He H, Yu X, Song S, Zhang J, Murch RD. Learning Bayes-optimal channel estimation for holographic MIMO in unknown EM environments. In:Proceedings of IEEE International Conference on Communications; 2024 Jun 9-13; Denver, CO, USA. New York City: IEEE; 2024. p. 3592-7.

[160]

Aali A, Arvinte M, Kumar S, Tamir JI. Solving inverse problems with score-based generative priors learned from noisy data. In: Proceedings of the 2023 57th Asilomar Conference on Signals, Systems, and Computers; 2023 Oct 29-Nov 1; Grove, CA, USA. New York City: IEEE; 2023. p. 837-43.

[161]

H. Du, R. Zhang, Y. Liu, J. Wang, Y. Lin, Z. Li, et al. Enhancing deep reinforcement learning: a tutorial on generative diffusion models in network optimization. IEEE Commun Surv Tutor, 26 (4) (2024), pp. 2611-2646.

[162]

Kim M, Fritschek R, Schaefer RF. Learning end-to-end channel coding with diffusion models. In:Proceedings of the 26th International ITG Workshop on Smart Antennas and 13th Conference on Systems, Communications, and Coding; 2023 Feb 27; Braunschweig, Germany. New York City: IEEE; 2023. p. 1-6.

[163]

Letafati M, Ali S, Latva-aho M. Diffusion models for wireless communications. 2023. arXiv:2310.07312.

[164]

Meng X, Kabashima Y. Diffusion model based posterior sampling for noisy linear inverse problems. 2024. arXiv:2211.12343.

[165]

Zhou X, Liang L, Zhang J, Jiang P, Li Y, Jin S. Generative diffusion models for high dimensional channel estimation. 2024. arXiv:240810501.

[166]

J. Guo, C.K. Wen, S. Jin, G.Y. Li. Overview of deep learning-based CSI feedback in massive MIMO systems. IEEE Trans Commun, 70 (12) (2022), pp. 8017-8045.

[167]

Zilberstein N, Swami A, Segarra S. Joint channel estimation and data detection in massive MIMO systems based on diffusion models. In:Proceedings of IEEE International Conference on Acoustics, Speech and Signal Processing; 2024 Apr 14-19; Seoul, Republic of Korea. New York City: IEEE; 2024. p. 13291-5.

[168]

Ho J, Salimans T. Classifier-free diffusion guidance. 2022. arXiv:220712598.

[169]

Dhariwal P, Nichol A. Diffusion models beat GANs on image synthesis. In:Proceedings of the 35th International Conference on Neural Information Processing Systems; 2021 Dec 6-14; online. Red Hook: Curran Associates Inc.; 2021. p 8780-94.

[170]

H. Xie, Z. Qin, G.Y. Li, B.H. Juang. Deep learning enabled semantic communication systems. IEEE Trans Signal Process, 69 (2021), pp. 2663-2675.

[171]

Li H, Yu W, He H, Shao J, Song S, Zhang J. Task-oriented communication with out-of-distribution detection:an information bottleneck framework. In:Proceedings of IEEE Global Communications Conference; 2023 Dec 4-8; Kuala Lumpur, Malaysia. New York City: IEEE; 2023. p. 3136-41.

[172]

J. Shao, Y. Mao, J. Zhang. Learning task-oriented communication for edge inference: an information bottleneck approach. IEEE J Sel Areas Commun, 40 (1) (2022), pp. 197-211.

[173]

Houlsby N, Giurgiu A, Jastrzebski S, Morrone B, De Laroussilhe Q, Gesmundo A, et al. Parameter-efficient transfer learning for NLP. In:Proceedings of the 36th International Conference on Machine Learning; 2019 Jun 10-15; Long Beach, CA, USA. New York City: ML Research Press; 2019. p. 2790-9.

[174]

Bu Z, Wang YX, Zha S, Karypis G. Differentially private bias-term fine-tuning of foundation models. In:Proceedings of the Forty-First International Conference on Machine Learning; 2024 Jul 21-27; Vienna, Austria. New York City: JMLR. org; 2024. p. 4730-51.

[175]

Hu EJ, Shen Y, Waills P, Allen-Zhu Z, Li Y, Wang S, et al. LoRA: low-rank adaptation of large language models. 2021. arXiv:210609685.

[176]

H. Sun, H. Tian, W. Ni, J. Zheng, D. Niyato, P. Zhang. Federated low-rank adaptation for large models fine-tuning over wireless networks. IEEE Trans Wirel Commun, 24 (1) (2025), pp. 659-675.

[177]

Kang T, Wang Z, He H, Zhang J, Song S, Letaief KB. Federated low-rank adaptation with differential privacy over wireless networks. 2024. arXiv:241107806.

[178]

Zhang K, He H, Song S, Zhang J, Letaief KB. Communication-efficient distributed on-device LLM inference over wireless networks. 2025. arXiv:2503.14882.

[179]

Zou Q, Yang H.A concise tutorial on approximate message passing. 2022. arXiv:220107487.

[180]

B. Efron. Tweedie’s formula and selection bias. J Am Stat Assoc, 106 (496) (2011), pp. 1602-1614.

[181]

K. Wang, Z. Gao, S. Chen, B. Ning, G. Chen, Y. Su. Knowledge and data dual-driven channel estimation and feedback for ultra-massive MIMO systems under hybrid field beam squint effect. IEEE Trans Wirel Commun, 23 (9) (2024), pp. 11240-11259.

[182]

Open AI; Achiam J, Adler S, Agarwai S, Ahmad L, Akkaya I, et al. GPT-4 technical report. 2024. arXiv:2303.08774.

[183]

DeepSeek-AI; Guo D, Yang D, Zhang H, Song J, Zhang R, et al. DeepSeek-R1: incentivizing reasoning capability in LLMs via reinforcement learning. 2025. arXiv:2501.12948.

[184]

Zhao WX, Zhou K, Li J, Tang T, Wang X, Hou Y, et al. A survey of large language models. 2025. arXiv:2303.18223.

[185]

Liu B, Gao S, Liu X, Cheng X, Yang L.WiFo: wireless foundation model for channel prediction. 2025. arXiv:2412. 08908.

[186]

Yu L, Shi L, Zhang J, Wang J, Zhang Z, Zhang Y, et al. ChannelGPT: a large model to generate digital twin channel for 6G environment intelligence. 2024. arXiv:2410.13379.

[187]

Zheng T, Dai L.Large language model enabled multi-task physical layer network. 2025. arXiv:2412. 20772.

[188]

Liu X, Gao S, Liu B, Cheng X, Yang L. LLM4WM: adapting LLM for wireless multi-tasking. 2025. arXiv:2501.12983.

[189]

Yang T, Zhang P, Zheng M, Shi Y, Jing L, Huang J, et al. WirelessGPT: a generative pre-trained multi-task learning framework for wireless communication. 2025. arXiv:2502.06877.

[190]

Zhang E, Goto R, Sagan N, Mutter J, Phillips N, Alizadeh A, et al. LLM-LASSO: a robust framework for domain-informed feature selection and regularization. 2025. arXiv:2502.10648.

[191]

Lee W, Park J.LLM-empowered resource allocation in wireless communications systems. 2024. arXiv:2408. 02944.

[192]

Zhou H, Hu C, Yuan D, Yuan Y, Wu D, Chen X, et al. Large language models for wireless networks: an overview from the prompt engineering perspective. 2024. arXiv:2411.04136.

[193]

Nazar AM, Celik A, Selim MY, Abdallah A, Qiao D, Eltawil AM.ENWAR: a RAG-empowered multi-modal LLM framework for wireless environment perception. 2024. arXiv:2410. 18104.

[194]

Weindel F, Heckel R.LLM-guided search for deletion-correcting codes. 2025. arXiv:2504. 00613.

[195]

Tong J, Shao J, Wu Q, Guo W, Li Z, Lin Z, et al. WirelessAgent: large language model agents for intelligent wireless networks. 2024. arXiv:2409.07964.

[196]

Shao J, Li X. AI Flow at the network edge. IEEE Netw 2026; 40(1):330-6.

[197]

Nikbakht R, Benzaghta M, Geraci G. TSpec-LLM: an open-source dataset for LLM understanding of 3GPP specifications. 2024. arXiv:2406.01768.

[198]

Bariah L, Zou H, Zhao Q, Mouhouche B, Bader F, Debbah M. Understanding telecom language through large language models. In:Proceedings of IEEE Global Communications Conference; 2023 Dec 8-12; Kuala Lumpur, Malaysia. New York City: IEEE; 2023. p. 6542-7.

[199]

Bornea AL, Ayed F, De Domenico A, Piovesan N, Maatouk A.Telco-RAG: navigating the challenges of retrieval-augmented language models for telecommunications. 2024. arXiv:2404. 15939.

[200]

G.M. Yilma, J.A. Ayala-Romero, A. Garcia-Saavedra, X. Costa-Perez. TelecomRAG: taming telecom standards with retrieval augmented generation and LLMs. ACM SIGCOMM Comput Commun Rev, 54 (3) (2024), pp. 18-23.

[201]

C.M. Bishop, N.M. Nasrabadi. Pattern recognition and machine learning. Springer, Berlin (2006).

AI Summary AI Mindmap
PDF (2397KB)

378

Accesses

0

Citation

Detail

Sections
Recommended

AI思维导图

/