We propose enhancements to score-based generative modeling techniques for low-latency pilot-based channel estimation in a point-to-point single-carrier multiple-input multiple-output (MIMO) wireless system. Building on recent advances in score-based models, we investigate a specific noise schedule design and sampling acceleration by step-skipping to reduce the number of denoising steps during inference. We additionally propose a single-step signal-to-noise ratio informed denoiser as an extreme case of the step-skipping approach. Our methods achieve significant latency reductions without performance degradation, as demonstrated on a synthetic channel dataset representing an urban macrocell MIMO communications scenario.
Score-based generative modeling, diffusion model, low-latency channel estimation, MIMO system.
Accurate acquisition is a key factor when scaling up wireless communication systems [1]. However, conventional estimators based on maximum likelihood, , and Bayesian techniques are reaching their limits. Therefore, and in particular based methods have been proposed due to their great potential in learning the underlying prior distribution of a wireless channel [2]. This distribution covers the site-specific channel characteristics of a ’s surrounding. In accordingly designed algorithms, exploiting this prior knowledge can significantly improve the channel estimation quality. Hence, various -based channel estimation algorithms using, e.g, [3], [4], and [5] have recently been proposed.
The latest advancements derive channel estimators based on [6] and the closely related [7], showcasing great potential. Especially, their prevailing generative abilities compared to, e.g., , together with a stable training process and design flexibility, make them interesting candidates for wireless algorithm development. The major drawback of and is their high latency during inference due to the iterative sampling procedure. To deploy algorithms that utilize / in real-time wireless communications systems, we thus have to overcome this limitation.
The authors of [8] introduced a -based framework, incorporating the pilot information in a guidance term during posterior sampling. However, their implementation exhibits high computational complexity. The channel estimator proposed in [9] features a deterministic adaptation of the [6] and a -dependent denoising procedure. Also, by exploiting structural properties of the system model, the authors can utilize a lightweight implementation, reducing computational complexity. Building on this work, [10] extends the channel estimator to a broader set of problems, specifically under-determined and quantized channel estimation. Additionally, they propose a framework solely using noisy pilots for training, altogether avoiding the necessity of ground-truth channels. In [11], the - and -based methods are compared to a channel estimator based on a [12], which exhibits better performance and also reduces the number of denoising steps. Another recently published work [13], incorporating many ideas of [9], improves the estimator’s efficiency even more by using the DPM-Solver framework [14]. The authors of [15] propose to use a Brownian-Bridge in an integrated sensing and communications system to estimate the sensing channel.
Apart from channel estimation, and have also been proposed for semantic communications [16]–[18], data augmentation [19], [20], joint channel estimation and data detection [21] and end-to-end learning [22].
The main focus of this work is to further reduce the complexity of a -based channel estimation algorithm. We examine improving the iterative denoising procedure of the rather than any improvements in the architecture itself. Our contributions can be summarized as follows:
We adapt the channel estimation algorithm proposed in [9] using formulations for a with a Gaussian kernel.
We propose two different techniques to accelerate the denoising procedure, i.e., a specific noise schedule design and a step-skipping approach.
We derive a single-step -informed channel estimator, which is the extreme case of the step-skipping idea.
We validate the effectiveness of our proposed enhancements with numerical simulations conducted on synthetic channel data.
The rest of this work is structured as follows. We introduce the channel estimation problem and the system model in Section 2. Section 3 shortly discusses the framework and explains its usage in the channel estimation algorithm. Enhancements of the -based channel estimator are proposed in Section 4 and supported with simulation results presented in Section 5. Section 6 concludes this work.
Bold-faced lower-case and upper-case letters denote vectors and matrices, respectively. The matrix \({\boldsymbol{F}}_N\) describes an \(N\)-dimensional matrix and \(\mathbf{I}_M\) is the \(M\)-dimensional identity matrix. The vectorization (stacking of columns) of a matrix \({\boldsymbol{A}} \in \mathbb{C}^{M \times N}\) is denoted as \(\mathop{\mathrm{vec}}({{\boldsymbol{A}}}) \in \mathbb{C}^{MN}\), and \({\boldsymbol{A}} \otimes {\boldsymbol{B}} \in \mathbb{C}^{MP \times NQ}\) is the Kronecker product of \({\boldsymbol{A}}\) and \({\boldsymbol{B}} \in \mathbb{C}^{P \times Q}\). A complex-valued multivariate Gaussian distribution with mean \({\boldsymbol{\mu}}\) and covariance matrix \({\boldsymbol{C}}\) is described by \(\mathcal{N}_{\mathbb{C}}({\boldsymbol{\mu}}, {\boldsymbol{C}})\). The gradient of a function \(f\) w.r.t. a variable \({\boldsymbol{x}}\) is written as \(\nabla_{{\boldsymbol{x}}} f\).
We consider a point-to-point single-carrier system with an \({N_{\mathrm{tx}}}\)-antenna transmitter and an \({N_{\mathrm{rx}}}\)-antenna receiver. The pilot-based transmission scheme is described by \[\label{eq:system95model95matrix} {\boldsymbol{Y}}= {\boldsymbol{H}}{\boldsymbol{P}} + {\boldsymbol{N}},\tag{1}\] where \({\boldsymbol{Y}}\in \mathbb{C}^{{N_{\mathrm{rx}}}\times {N_{\mathrm{p}}}}\) comprises the received signals, \({\boldsymbol{H}}\in \mathbb{C}^{{N_{\mathrm{rx}}}\times {N_{\mathrm{tx}}}}\) describes the unknown channel matrix, \({\boldsymbol{P}} \in \mathbb{C}^{{N_{\mathrm{tx}}}\times {N_{\mathrm{p}}}}\) is the unitary pilot matrix satisfying \({\boldsymbol{P}} {\boldsymbol{P}}^{\mathrm{H}}= \mathbf{I}_{N_{\mathrm{tx}}}\) and \({\boldsymbol{N}}\in \mathbb{C}^{{N_{\mathrm{rx}}}\times {N_{\mathrm{p}}}}\) describes Gaussian noise with columns \({\boldsymbol{n}}_i \sim \mathcal{N}_{\mathbb{C}}({\boldsymbol{0}}, \eta^2 \mathbf{I})\). The objective is to recover the unknown channel matrix \({\boldsymbol{H}}\) from the observed signal \({\boldsymbol{Y}}\). We assume \({N_{\mathrm{p}}}= {N_{\mathrm{tx}}}\), resulting in a fully-determined channel estimation scenario. Rewriting 1 as \[\label{eq:system95model95vector} {\boldsymbol{y}}= ({\boldsymbol{P}}^{\mathrm{T}}\otimes \mathbf{I}_{{N_{\mathrm{rx}}}}) {\boldsymbol{h}}+ {\boldsymbol{n}}= {\boldsymbol{A}} {\boldsymbol{h}}+ {\boldsymbol{n}},\tag{2}\] with \({\boldsymbol{y}}= \mathop{\mathrm{vec}}({\boldsymbol{Y}})\), \({\boldsymbol{h}}= \mathop{\mathrm{vec}}({\boldsymbol{H}})\), \({\boldsymbol{n}}= \mathop{\mathrm{vec}}({\boldsymbol{N}})\) and \({\boldsymbol{A}}=({\boldsymbol{P}}^{\mathrm{T}}\otimes \mathbf{I}_{{N_{\mathrm{rx}}}})\) enables us to work with vectors instead of matrices. Note that \({\boldsymbol{A}}\) is also unitary since it is derived by the Kronecker product of two unitary matrices.
The channels are assumed to be distributed according to an unknown prior \(p({\boldsymbol{h}})\). Assuming \(\mathop{\mathrm{\mathbb{E}}}_{p}[{\boldsymbol{h}}] = {\boldsymbol{0}}\) and \(\mathop{\mathrm{\mathbb{E}}}_{p}[\| {\boldsymbol{h}}\|^2] = {N_{\mathrm{tx}}}{N_{\mathrm{rx}}}\) leads to the definition \(\mathrm{SNR} ({\boldsymbol{y}}) = 1/\eta^2\).
The considered in this work is based on the framework introduced in [7], which is briefly reviewed in this section. We use the forward \[\label{eq:forward95sde} \operatorname{d} \!{\boldsymbol{h}}(t) = \sqrt{\frac{\operatorname{d} \![\sigma^2(t)]}{\operatorname{d} \!t}} \operatorname{d} \!{\boldsymbol{w}} (t)\tag{3}\] for \(t \in [0, T]\) with \({\boldsymbol{h}}(0) \mathrel{\vcenter{:}}= {\boldsymbol{h}}\sim p({\boldsymbol{h}})\) and a standard Wiener process \({\boldsymbol{w}}(t)\) to continuously diffuse a ground-truth channel sample into Gaussian noise. The noise schedule \(\sigma(t)\) controls the diffusion process and must satisfy the condition \(\lim_{t\to0} \sigma(t) =0\). Alternatively, a noisy sample can directly be produced by sampling from the Gaussian kernel \(q_\sigma({\boldsymbol{h}}(t) | {\boldsymbol{h}}(0)) = \mathcal{N}_{\mathbb{C}}({\boldsymbol{h}}(t); {\boldsymbol{h}}(0), \sigma^2(t) \mathbf{I})\), such that \[\label{eq:forward95kernel} {\boldsymbol{h}}(t) = {\boldsymbol{h}}(0) + \sigma(t) {\boldsymbol{\epsilon}}(t)\tag{4}\] using the reparameterization trick with \({\boldsymbol{\epsilon}}(t) \sim \mathcal{N}_{\mathbb{C}}({\boldsymbol{0}}, \mathbf{I})\). This leads to the ’s definition \(\mathrm{SNR}^{\mathrm{SBM}}(t) = 1 / \sigma^2(t)\).
The reverse of 3 moving from \(t=T\) down to \(t=0\) is derived by \[\label{eq:reverse95sde} \operatorname{d} \!{\boldsymbol{h}}(t) = - \frac{\operatorname{d} \![\sigma^2(t)]}{\operatorname{d} \!t} \nabla_{{\boldsymbol{h}}(t)}\!\log p({\boldsymbol{h}}(t)) +\sqrt{\frac{\operatorname{d} \![\sigma^2(t)]}{\operatorname{d} \!t}} \operatorname{d} \!\bar{{\boldsymbol{w}}} (t)\tag{5}\] introducing the score function \(\nabla_{{\boldsymbol{h}}(t)}\!\log p({\boldsymbol{h}}(t))\) and the reverse-time Wiener process \(\bar{{\boldsymbol{w}}} (t)\). Discretizing 5 according to the Euler-Maruyama method [23] yields \[\label{eq:reverse95sde95discrete952} {\boldsymbol{h}}_{k-1} = {\boldsymbol{h}}_k + (\sigma_k^2 - \sigma_{k-1}^2) \nabla_{{\boldsymbol{h}}_k}\!\log p({\boldsymbol{h}}_k) + \sqrt{\sigma_k^2 - \sigma_{k-1}^2}{\boldsymbol{z}}_k.\tag{6}\] From the discretization follows that \({\boldsymbol{h}}(t) \rightarrow \{{\boldsymbol{h}}_k\}_{k=1}^K\) and \(\sigma(t) \rightarrow \{\sigma_k\}_{k=1}^K\), where \(K\) is the total number of discretization points. However, the score function in 6 is generally unknown and is thus approximated with a \({\boldsymbol{s}}_{{\boldsymbol{\theta}}}\) parametrized by \({{\boldsymbol{\theta}}}\) at each discretization step using [24]. The corresponding training procedure is outlined in Algorithm 1. Upon successful training, the replaces the score function in 6 to get \[\label{eq:reverse95sde95discrete} {\boldsymbol{h}}_{k-1} = {\boldsymbol{h}}_k + (\sigma_k^2 - \sigma_{k-1}^2) {\boldsymbol{s}}_{{\boldsymbol{\theta}}}({\boldsymbol{h}}_k, k) + \sqrt{\sigma_k^2 - \sigma_{k-1}^2}{\boldsymbol{z}}_k.\tag{7}\] The resulting iterative sampling procedure can also be interpreted as a Markov chain with Gaussian transition distributions \(p_{{\boldsymbol{\theta}}}({\boldsymbol{h}}_{k-1} | {\boldsymbol{h}}_{k}) = \mathcal{N}_{\mathbb{C}}({\boldsymbol{h}}_{k-1}; {\boldsymbol{\mu}}_{{{\boldsymbol{\theta}}}} ({\boldsymbol{h}}_k, k), {\boldsymbol{C}}_k)\) with means \({\boldsymbol{\mu}}_{{{\boldsymbol{\theta}}}} ({\boldsymbol{h}}_k, k) = {\boldsymbol{h}}_k + (\sigma_k^2 - \sigma_{k-1}^2) {\boldsymbol{s}}_{{\boldsymbol{\theta}}}({\boldsymbol{h}}_k, k)\) and covariance matrices \({\boldsymbol{C}}_k = (\sigma_k^2 - \sigma_{k-1}^2) \mathbf{I}\).
We build on the -based channel estimation algorithm proposed in [9], which combines an -informed estimator with a lightweight . We also train our in the beamspace domain by applying a Kronecker to the spatial channels, thereby exploiting structural properties of wireless signals. Also, we adopt their implementation and the -dependent estimation procedure. However, the authors of [9] employ a architecture instead of Gaussian kernels. Therefore, some derivations have to be slightly adapted for this work.
Algorithm 2 concisely presents the adapted -based channel estimation algorithm. After the initial decorrelation step of the pilot observation \({\boldsymbol{y}}\), we directly conduct the beamspace transformation without a preceding normalization step. This can be omitted due to the equivalence between the decorrelated pilot observation and the ’s Gaussian kernel. The appropriate initial step is found by identifying the variance \(\sigma_k^2\) closest to the received observation’s variance \(\eta^2\) in line 3. The algorithm then enters the denoising procedure, where it iteratively applies 7 but without the stochastic term \(\sqrt{\sigma_k^2 - \sigma_{k-1}^2}{\boldsymbol{z}}_k\) in line 6. Including this term is reasonable for generative tasks but not for point-wise estimation [25]. Upon completion, the channel estimate in the beamspace domain is returned and transformed back into the spatial domain to finish the algorithm.
The main bottleneck in Algorithm 2 is the number of required denoising steps \({\hat{k}}\) as each step \(k\) requires a costly of \({\boldsymbol{s}}_{{\boldsymbol{\theta}}}({\boldsymbol{h}}_k, k)\). A naive approach to reducing \({\hat{k}}\) is to reduce the total number of discretization points \(K\). However, a drawback of this idea is that the worsened resolution of the discretized is affecting the estimator’s performance [9]. The following two techniques introduce new approaches to reducing denoising steps.
We consider the commonly used exponential noise schedule \(\sigma(t)\), which is discretized by the power-exponentiated geometric sequence \[\label{eq:noise95schedule} \sigma_k = \sigma_\mathrm{min}\left(\frac{\sigma_\mathrm{max}}{\sigma_\mathrm{min}}\right)^{(\frac{k-1}{K-1})^\gamma}\tag{8}\] with the parameter \(\gamma\) controlling the schedule’s steepness. The minimum/maximum variances are derived from the maximum/minimum the should cover via the relation \[\label{eq:sbm95snr95relation} \sigma_{\text{min/max}} = \sqrt{\frac{1}{\mathrm{SNR}^{\mathrm{SBM}}_{\text{max/min}}}}.\tag{9}\]
From Algorithm 2, line 3, we can infer that the noise schedule directly influences \({\hat{k}}\) and thus affects the algorithm’s time complexity. For \(\gamma < 1\), the sequence in 8 grows faster and therefore fewer denoising steps \({\hat{k}}\) compared to a standard geometric sequence with \(\gamma = 1\) are required for the same \(\eta^2\). However, one has to take care that the ’s resolution does not get too coarse for high .
The method to speed up the algorithm even more introduced in this section is based on the general idea of skipping certain latent variables of a during sampling as proposed in, e.g., [12], [14]. In our -based framework, the adapted deterministic denoising procedure incorporating this approach is inferred from 7 and given by \[\label{eq:accelerated95reverse} \hat{{\boldsymbol{h}}}_{k-\Delta} = {\boldsymbol{h}}_k + (\sigma_k^2 - \sigma_{k-\Delta}^2) {\boldsymbol{s}}_{{\boldsymbol{\theta}}}({\boldsymbol{h}}_k, k),\tag{10}\] where we introduced the step variable \(\Delta \in \mathbb{N}\). The case \(\Delta=1\) exactly resembles the standard method discussed in Section 3.2. For \(\Delta>1\), we skip certain points of the original discretization and, as a consequence, enlarge the step-size given by \(\sigma_k^2 - \sigma_{k-\Delta}^2\) since \(\{\sigma_k\}_{k=1}^K\) is a monotonically increasing sequence. This effect can also be interpreted as a coarser discretization of the reverse during the denoising procedure. It is important to mention that we still maintain the fine resolution of the for identifying the initial step \({\hat{k}}\).
In Algorithm 2, line 6 has to be replaced with 10 to realize the accelerated denoising. Also, we have to ensure that the loop is indexed by \(k = {{\hat{k}}, {\hat{k}}-\Delta, {\hat{k}}-2\Delta, \ldots, {\hat{k}}\mod \Delta}\). In the last denoising step, we set \(\sigma_{k-\Delta}=0\) to completely remove the remaining noise in the final iteration. The number of required denoising steps and thereby the algorithm’s complexity decreases to \({\tilde{k}}_{\Delta} = \lceil \frac{{\hat{k}}}{\Delta}\rceil\).
Single-Step SNR-informed Denoiser. In the extreme case of \(\Delta = \Delta_{\mathrm{max}} = K\), the number of denoising steps reduces to a single step \[\label{eq:single95step} \hat{{\boldsymbol{h}}}_0 = {\boldsymbol{h}}_k + \sigma_k^2 {\boldsymbol{s}}_{{\boldsymbol{\theta}}}({\boldsymbol{h}}_k, k)\tag{11}\] which completely removes the denoising loop as depicted in Algorithm 3. The changes compared to Algorithm 2 are highlighted in blue. This algorithm rather resembles a -informed -based denoiser than a framework using a .
To better understand how the single-step denoiser works, recall that in , the is trained to estimate the score function of the diffused channel data conditioned on a ground-truth sample \({\boldsymbol{h}}_0\). This can also be reformulated as \[{\boldsymbol{s}}_{{\boldsymbol{\theta}}}({\boldsymbol{h}}_k, k) \approx \nabla_{{\boldsymbol{h}}_k} q_{\sigma}({\boldsymbol{h}}_k | {\boldsymbol{h}}_0) = - \frac{1}{\sigma_k} {\boldsymbol{\epsilon}}_k,\] where \({\boldsymbol{\epsilon}}_k\) is the noise realization used to generate \({\boldsymbol{h}}_k\). Therefore, during training, instead of directly estimating the score, we train a \({\boldsymbol{\epsilon}}_{{\boldsymbol{\theta}}}({\boldsymbol{h}}_k, k)\) to approximate \({\boldsymbol{\epsilon}}_k\) and set the learned score to \({\boldsymbol{s}}_{{\boldsymbol{\theta}}}({\boldsymbol{h}}_k, k) = - \frac{1}{\sigma_k} {\boldsymbol{\epsilon}}_{{\boldsymbol{\theta}}}({\boldsymbol{h}}_k, k)\). Incorporating these reformulations into 11 yields \[\hat{{\boldsymbol{h}}}_0 = {\boldsymbol{h}}_k - \sigma_k {\boldsymbol{\epsilon}}_{{\boldsymbol{\theta}}}({\boldsymbol{h}}_k, k).\] This can be interpreted as the approximated inverse of the Gaussian forward kernel expressed in 4 . Thus, adding the scaled estimated score in 11 can be interpreted as a direct denoising step — effectively identifying and removing the added noise in a single step.
We work with version 2.6 of the channel simulator [26] and consider a scenario with a strong path and additional multi-path components. The is placed at a height of \(\SI{25}{m}\) and equipped with a consisting of \({N_{\mathrm{rx}}}= 64\) antennas. It covers a sector of \(120^\circ\) in which are placed at random with distances ranging between \(\SI{35}{m}\) and \(\SI{500}{m}\). Each itself consists of a equipped with \({N_{\mathrm{tx}}}= 16\) antennas. We consider half-wavelength antenna spacing for all and the respective channels are simulated at a frequency of \(\SI{6}{GHz}\) for a single snapshot. Although we choose a large number of antennas at the in this work, the characteristics identified in the simulation results also hold for fewer antennas, and especially also in the case of a system with \({N_{\mathrm{tx}}}=1\). A post-processing step removes the effective path gain for further processing according to [26]. Knowledge of the effective path gain is a viable assumption as it can be estimated with small efforts in a real communications system. In total, this generates \(1.2 \cdot 10^5\) channel realizations \({\boldsymbol{H}}_m \sim p({\boldsymbol{H}})\) for different locations in the scenario, which are split up in sets of \(M_\text{train} = 10^5\), \(M_\text{val} = 10^4\) and \(M_\text{test} = 10^4\) training, validation and test samples, respectively.
The is designed upon the noise schedule in 8 with a total of \(K=100\) discretization steps. The maximum and minimum covered by the are set to \(\mathrm{SNR}^{\mathrm{SBM}}_{\text{max}} = \SI{40}{dB}\) and \(\mathrm{SNR}^{\mathrm{SBM}}_{\text{min}} = \SI{-22}{dB}\), resulting in minimum and maximum noise variances of \(\sigma_\mathrm{min} = 0.01\) and \(\sigma_\mathrm{max} \approx 12.6\) according to 9 . If not stated otherwise, \(\gamma=1.0\) and \(\Delta=1\) hold.
As already stated in Section 3.2, we reuse the implementation of [9] without performing any specific hyperparameter tuning. The architecture combines a lightweight 2D- with a positional Transformer sinusoidal position embedding of step \(k\). This enables parameter sharing of the across all discretization steps. The is trained with the AdamW optimizer [27] using a batch size of \(128\), an adaptive learning rate, and a stopping criterion depending on the validation loss evolution over training epochs. For each simulation, we train \(5\) randomly initialized and pick the one exhibiting the lowest validation loss after training for further use in channel estimation inference. We refer to our publicly available GitHub repository for further implementation details.2
The proposed -based channel estimator is compared to the following baseline algorithms.
The estimate denoted as "LS" solves the optimization problem \(\hat{{\boldsymbol{h}}}_\mathrm{LS} = \mathop{\mathrm{\arg\!\min}}_{{\boldsymbol{h}}} \|{\boldsymbol{y}}- {\boldsymbol{A}} {\boldsymbol{h}}\|^2\) and can be derived as \(\hat{{\boldsymbol{h}}}_\mathrm{LS} = {\boldsymbol{A}}^{\mathrm{H}}{\boldsymbol{y}}\) as \({\boldsymbol{A}}\) is assumed to be unitary. We also evaluate a based estimate denoted as "SCov-LMMSE" with a global sample covariance matrix \({\boldsymbol{C}} = \frac{1}{M_\text{train}} \sum_{m=1}^{M_\text{train}} {\boldsymbol{h}}_m {\boldsymbol{h}}_m^{\mathrm{H}}\) as \(\hat{{\boldsymbol{h}}}_\textrm{Scov-LMMSE} = {\boldsymbol{C}} ({\boldsymbol{C}} + \eta^2 \mathbf{I})^{-1}{\boldsymbol{y}}\).
As prior-aided baselines, we choose two different -based estimators for comparison. The first version, denoted as "GMM" and proposed in [3], fits a to the ground-truth data such that it approximates the underlying distribution well. The learnt means and covariances of the Gaussian components are then exploited in an adapted estimator. The second version, denoted as "GMM-kron" and proposed in [28], follows the same reasoning but trains two separate on the receiver and transmitter side, respectively. Afterwards, covariance matrices of both are combined with the Kronecker product to build the final covariance matrices used in the adapted .
We opt for the metric to assess the channel estimator’s performance. Therefore, for each noise level, a noisy observation of each channel sample \({\boldsymbol{h}}_m\) in the test set is created according to 1 . Then, a channel estimate \(\hat{{\boldsymbol{h}}}_m\) is computed with one of the considered estimators. The for each level is evaluated as \[\mathrm{NMSE} = \frac{1}{{N_{\mathrm{rx}}}{N_{\mathrm{tx}}}M_\text{test}} \sum_{m=1}^{M_\text{test}}\|{\boldsymbol{h}}_m - \hat{{\boldsymbol{h}}}_m \|_2^2.\] It is important to note that our -based channel estimator does not require retraining for different but operates on the whole range.
The effect of choosing different noise schedules on the number of required denoising steps is depicted in Fig. 4 for different choices of the parameter \(\gamma\). Because the noise schedule is an exponential function and we plot the on a logarithmic scale, a linear decrease of \({\hat{k}}\) can be seen for the regular geometric sequence with \(\gamma=1\) in black. As proposed in Section 4.1 and validated in Fig. 4, decreasing \(\gamma\) significantly decreases \({\hat{k}}\) for every \(\mathrm{SNR}({\boldsymbol{y}})\). We also include a simulation for \(\gamma>1\) to investigate the effect on the performance when increasing the number of denoising steps.
Fig. [fig:nmse95noise95schedule] depicts the for the different choices of \(\gamma\). There is almost no degradation in performance when decreasing \(\gamma\), only in the high- regime. The deterioration stems from the discretization being too coarse in the low-noise region for small values of \(\gamma\). Also, as the graph for \(\gamma=1.6\) is on par with the others, there is no benefit in increasing the number of denoising steps by adapting the noise schedule. This indicates that the algorithm can denoise a pilot observation well as long as we maintain a good enough discretization.
Even more impressive is the reduction of denoising steps with the accelerated denoising approach, cf. Section 4.2, as depicted in Fig. 5. The black curve again resembles the baseline implementation, where every discretization step is conducted in the denoising procedure. With increasing \(\Delta\), we can accelerate the denoising procedure by roughly a factor of \(\Delta\) until only a single step is required for \(\Delta=\Delta_\mathrm{max}\) independent of the .
Surprisingly, increasing the step variable \(\Delta\) does not worsen the . Fig. [fig:nmse95accelerated] showcases that it even improves the channel estimation performance by a small but significant margin, not stemming from computational instabilities, over the whole range. Especially, in the mid- range, the improvement is clearly visible. This result seems counterintuitive at first sight; however, we make the following hypothesis describing the effect: We refer to [9], where the authors show the evolution of the performance during the denoising procedure for different levels. However, one artifact not discussed in the paper is a rise in the at the final steps of the estimation procedure after the has reached its minimum. Although we still do not fully understand this effect, we claim that increasing \(\Delta\) mitigates this effect, which explains the counterintuitive improvement in Fig. [fig:nmse95accelerated].
Fig. 6 depicts the performance of different channel estimators measured in terms of the . The LS and SCov-LMMSE estimators, as two conventional methods, are barely able to reconstruct the channel well, thereby emphasizing the need for more advanced estimation algorithms. The -based approaches and especially the Kronecker version thereof significantly improve the performance in the low-to-mid -range while converging to the classical estimators’ curves for higher .
The estimator denoted as "SBM" in Fig. 6 implements the -based channel estimator from Algorithm 2 without step-skipping (\(\Delta=1\)) and the regular exponential noise schedule (\(\gamma=1.0\)). It therefore requires the regular number of denoising steps given by \(\hat{k}\) and thus is computationally complex. However, it is capable of outperforming the other baselines over the whole -range by roughly 5 dB. The second -based method, denoted as "SBM single-step", implements the single-step -informed denoiser (\(\Delta = \Delta_\mathrm{max}\)) from Algorithm 3. As already discussed in the previous section and shown in Figs. 5 and [fig:nmse95accelerated], this technique is capable of further improving the while drastically reducing the computational complexity.
This paper introduced two methods to accelerate the denoising procedure of a -based channel estimator. By carefully adapting the forward ’s noise schedule, we significantly reduced the number of denoising steps with only minor performance losses. An even faster acceleration was achieved by skipping certain steps in the denoising procedure, which in the extreme case results in the single-step denoiser. This approach even improves the estimation performance due to the mitigation of an artifact in a previously proposed -based channel estimator.
Further research should clarify the theoretical optimality guarantees of the proposed denoiser and identify its limitations. Also, investigating the performance for different wireless communications scenarios is essential for the algorithm’s generalization capabilities. Regarding this point, an interesting research question not considered in this work is, e.g., what effect a rotation of the has on the channel estimation performance. As a final suggestion, the proposed estimator should be adapted to operate in a ed MIMO system for actual deployment in real-world communications scenarios.
This work was funded by Huawei Technologies Sweden AB, Stockholm.↩︎