Hybrid Neural Representations for Spherical Data


Abstract

In this paper, we study hybrid neural representations for spherical data, a domain of increasing relevance in scientific research. In particular, our work focuses on weather and climate data as well as comic microwave background (CMB) data. Although previous studies have delved into coordinate-based neural representations for spherical signals, they often fail to capture the intricate details of highly nonlinear signals. To address this limitation, we introduce a novel approach named Hybrid Neural Representations for Spherical data (HNeR-S). Our main idea is to use spherical feature-grids to obtain positional features which are combined with a multilayer perception to predict the target signal. We consider feature-grids with equirectangular and hierarchical equal area isolatitude pixelization structures that align with weather data and CMB data, respectively. We extensively verify the effectiveness of our HNeR-S for regression, super-resolution, temporal interpolation, and compression tasks.

1 Introduction↩︎

Coordinate-based neural representations (CNRs) form a family of techniques to parameterize the target signal using a neural network that maps input coordinates to signal values [1]. This results in a continuous function that seamlessly interpolates discretized coordinates of the target signal. As a case in point, when applied to image representation, CNRs trained to fit RGB values on 2D-pixel coordinates can also predict the continuum of coordinates between pixels. This capability has been proven to be useful in representing, interpolating, and increasing the signal resolutions of various modalities, including images [2], signed distance functions [3], and radiance fields [4] as special cases.

However, CNRs have been less investigated for spherical signals despite the important scientific applications, e.g., global weather and climate modeling [5] or the analysis of cosmic microwave background (CMB) radiation [6]. This is noteworthy given the necessity for interpolating and increasing the resolution of spherical signals in these applications. For example, spatial or temporal interpolation is required for building a fine-grained weather forecasting system [7], [8]. Increasing the resolution of CMB data may discover new insights about the early universe [9][11].

To this end, recent works [12][16] have investigated CNRs based on coordinate-based multilayer perceptrons (MLPs). To featurize the spherical coordinates, they use the sinusoids from the Cartesian coordinate system [15], [16], spectral embeddings of the discretized manifold [13], [14], or the spherical harmonics with fixed frequency components [12]. However, these methods suffer from the inherent limitation of coordinate-based MLPs: lack of expressive power to approximate highly nonlinear signals to their finest details.

For Euclidean data, recent works [17][21] have resolved this issue by using hybrid neural representations that combine feature-grid structures with the MLP. Their key idea is to decompose the input space into grid structures and assign learnable parameters as positional features for each point in the grid. Then the positional features can largely fluctuate accordingly with the unconstrained point-wise parameters. However, extending the existing hybrid neural representation to spherical data is non-trivial since they rely on grid structures and interpolation algorithms specific to Euclidean domains, e.g., sparse voxel grids and Euclidean distance-based interpolation, that are highly redundant or ill-defined for spherical data.

Contribution. In this work, we develop a novel hybrid neural representation for spherical data (HNeR-S) that combines spherical feature-grids with MLP. Our main idea is to use a hierarchy of coarse- to fine-grained spherical feature-grids to generate positional features of a spherical coordinate at multiple frequencies. We focus on scientific data: weather and climate data as well as CMB data. We provide an overview of HNeR-S in 1.

To compute the positional feature for an input point, we assign parameters to the grid point in the spherical feature-grid and apply a spherical interpolation algorithm to output the feature for the input position. Especially, we choose to align the feature-grid structure with the inherent data structure. This leads to designing our method with (1) an equirectangular feature-grid for the weather and climate data and (2) a hierarchical equal area iso-latitude pixelization (HEALPix) feature-grid for the CMB data.

We thoroughly evaluate our algorithm on weather and climate [5] and CMB [11] data for super-resolution, regression, temporal interpolation, and compression tasks. Furthermore, we implement and compare with five CNR baselines [4], [12], [22][24] applicable to spherical data.

To summarize, our contributions are as follows:

  • We propose a novel hybrid neural representation for spherical data (HNeR-S), which tailors the existing hybrid neural representations for spherical data using spherical feature-grids and interpolation algorithms.

  • We empirically evaluate our algorithms on climate and weather data as well as CMB data. The results demonstrate our superiority over five INR baselines on super-resolution, regression, temporal interpolation, and compression tasks.

2 Related Work↩︎

Coordinate-based neural representations (CNRs). CNRs aim to learn a mapping from coordinates to the signal values via neural networks. The choice of positional encodings and activation functions are important for CNRs to capture high-frequency information of the underlying signal [25]. To this end, sinusoidal positional encoding [24] and activation function [22] have been proposed for Euclidean data, e.g., images and videos. [23] proposed the Gabor wavelet activations to improve the performance and the robustness of sinusoidal encodings for the image domain. [24] analyzed the benefits of positional encoding to learn high-frequency functions via neural tangent kernel theory [26].

CNRs for non-Euclidean data. Recently, researchers have also considered generalizing CNRs to non-Euclidean data [12][15], [27]. A simple approach is to lift the two-dimensional sphere into three-dimensional Euclidean space so that the Cartesian coordinates of the sphere can be used as inputs to the existing three-dimensional CNRs [15], [16]. To incorporate the geometry of the data, researchers considered using the eigenfunctions of the Laplace-Beltrami operator for general manifolds [13], [14]. For spherical data, this corresponds to using spherical harmonics to featurize the spherical coordinates [12].

Learning weather and climate data. Recently, weather and climate data have gained popularity as an application of machine learning [28], which is attributed to its significant impact on climate prediction, mitigation, and adaptation. Employing neural networks for weather and climate data tasks has shown impressive results across a variety of tasks, e.g., super-resolution [29], [30], temporal modeling [31], [32], and compression [16], [33]. Neural network architectures like spherical convolutional neural networks [34] and spherical Fourier neural operators [35] have been tailored for this domain.

Learning cosmic microwave background (CMB) data. CMB data encapsulates the thermal radiation left over from the Big Bang, the origin of the universe. Machine learning has been applied to predict the posterior distribution of the cosmological parameters [36] as well as signal recovery and dust cleaning [37][39]. A potential application of machine learning for CMB data is to increase resolution of the data, which has mainly been progressed from the physical development of new sensors [9][11].

3 Method↩︎

In this section, we introduce our hybrid neural representations for spherical data (HNeR-S). We first elaborate on the general grid-agnostic framework (3.1) and provide detailed descriptions for the equirectangular and HEALPix feature-grids ([subsec:equi] [subsec:heal]). We provide a brief overview of each grid in 2.

Figure 2: Overview of equirectangular and HEALPix grid. Note that the uniformity of feature-grid refers to the consistency in the area covered by each unit cell within the grid.

3.1 General Framework↩︎

We formulate our problem as learning a mapping from spherical coordinates to the target signal. To this end, our HNeR-S employs a hierarchy of spherical feature grids and a bilinear spherical interpolation algorithm to output the positional feature of the input coordinate. Then HNeR-S applies a multi-layered perception (MLP) on the positional feature to predict the target signal. We provide an overview of HNeR-S in 1.

Problem definition. To be specific, we consider learning some ground-truth spherical signal \(f(\mathbf{x}_{\text{in}}) \in \mathbb{R}^{D_{\text{out}}}\), where the input coordinate \(\mathbf{x}_{\text{in}}=(\psi, \phi)\) is a pair of latitude \(\psi \in [-\frac{1}{2}\pi, \frac{1}{2}\pi]\) and longitude \(\phi \in [0, 2\pi)\). To approximate this signal, we train a neural network \(f_{\theta}\) on a dataset of tuples \((\mathbf{x}_\text{in}, \mathbf{y})\) with ground truth label \(\mathbf{y} = f(\mathbf{x}_\text{in})\).

Overall architecture. Our HNeR-S makes a prediction \(\hat{\mathbf{y}}=f_{\theta}(\mathbf{x}_{\text{in}})\) from an input point \(\mathbf{x}_{\text{in}}\) as follows: \[f_{\theta}(\mathbf{x}) = \operatorname{MLP}(\mathbf{Z}(\mathbf{x}_{\text{in}})),\] where \(\operatorname{MLP}(\cdot)\) denotes the MLP and \(\mathbf{Z}(\mathbf{x}_{\text{in}}) \in \mathbb{R}^{D\times L}\) denotes the concatenation of \(D\)-dimensional positional features constructed from \(L\) levels of feature-grids: \[\mathbf{Z}(\mathbf{x}_{\text{in}}) = \operatorname{Concat}(\mathbf{z}^{(1)}(\mathbf{x}_{\text{in}}),\ldots, \mathbf{z}^{(L)}(\mathbf{x}_{\text{in}})),\] where \(\operatorname{Concat}(\cdot)\) denotes concatenation of elements and the positional feature \(\mathbf{z}^{(\ell)}(\mathbf{x}_{\text{in}}) \in \mathbb{R}^{D}\) is constructed from the \(\ell\)-th level feature-grid for \(\ell=1,\ldots, L\). We assign a higher level \(\ell\) for a finer grid.

Positional feature. The \(\ell\)-th positional feature \(\mathbf{z}^{(\ell)}(\mathbf{x}_{\text{in}})\) is constructed by interpolating the positional features assigned to grid points \(\mathcal{V}^{(\ell)}\) in the \(\ell\)-th feature-grid. We let \(\mathbf{x}_{i}\) denote the coordinate of a grid point \(i \in \mathcal{V}^{(\ell)}\). Different grid types, e.g., equirectangular and HEALPix grids, correspond to a different organization of position \(\mathbf{x}_{i}^{(\ell)}\) for each grid point \(i\). We also let \(\mathbf{z}_{i}^{(\ell)} \in \mathbb{R}^{D}\) denote the associated positional feature, which is also a parameter to train for our framework.

Then given an input point with position \(\mathbf{x}_{\text{in}}\), the HNeR-S constructs the \(\ell\)-th positional feature \(\mathbf{z}^{(\ell)}(\mathbf{x}_{\text{in}})\) by interpolation of neighboring points \(\mathcal{N}^{(\ell)}(\mathbf{x}_{\text{in}}) \subseteq \mathcal{V}^{(\ell)}\) in the feature-grid: \[\mathbf{z}^{(\ell)}(\mathbf{x}_{\text{in}}) = \operatorname{Interpolate}(\mathbf{x}_{\text{in}}, \{(\mathbf{x}_{i}^{(\ell)}, \mathbf{z}_{i}^{(\ell)}):i \in \mathcal{N}^{(\ell)}(\mathbf{x}_{\text{in}})\}).\] The neighborhood structure of the equirectangular and HEALPix grid allows our HNeR-S to use bilinear spherical interpolation based on the latitude and the longitude.

Neighborhood structure. We consider neighborhood functions that output \(\mathcal{N}^{(\ell)}(\mathbf{x}_{\text{in}})\subseteq \mathcal{V}^{(\ell)}\) as a set of four points, denoted by \(\mathbf{x}_{i,j} = (\psi_{i,j}, \phi_{i,j})\) associated with positional feature \(\mathbf{z}_{i,j}^{(\ell)}\) for \(i, j \in \{1,2\}\). Importantly, we consider grids that allow the construction of the neighborhood as pairs of isolatitude points and let \(\psi_{1} = \psi_{1,1}=\psi_{1,2}, \psi_{2} = \psi_{2,1}=\psi_{2,2}\) without loss of generality.

Furthermore, the neighbors bound the input point by a latitude interval \(\Psi = [\psi_{1,1}, \psi_{1,2}]\) and two longitude intervals \(\Phi_{1}, \Phi_{2}\). When a neighbor of the input point does not overlap with the prime meridian (\(\phi=0\)), the longitude intervals are defined similarly, e.g., \(\Phi_{1} = [\phi_{1,1}, \phi_{1,2}]\). Otherwise, the longitude intervals are set by \(\Phi_{1} = [\phi_{1,1}, 2\pi] \cup [0, \phi_{1,2}]\) and \(\Phi_{2}\) is set similarly.

Bilinear spherical interpolation. Given the neighbor as pairs of isolatitude points, our framework employs bilinear interpolation using the longitude and the latitude to compute the associated feature. To be specific, given an input point \(\mathbf{x}_{\text{in}}=(\psi, \phi)\), we compute \(\mathbf{z}^{(\ell)}(\psi, \phi)\) via two steps: (1) computing \(\mathbf{z}^{(\ell)}(\psi_{1}, \phi)\) and \(\mathbf{z}^{(\ell)}(\psi_{2}, \phi)\) via isolatitude interpolation and (2) computing \(\mathbf{z}^{(\ell)}(\psi, \phi)\) via interpolation along the longitude between features computed in (1).

Precisely, the interpolation along the isolatitude neighbors \(\mathbf{x}_{1, 1} = (\psi_{1}, \phi_{1,1}), \mathbf{x}_{1, 2}=(\psi_{1}, \phi_{1,2})\) is defined as follows: \[\begin{align} \mathbf{z}^{(\ell)}(\psi_{1}, \phi) = \lambda \mathbf{z}_{1,1}^{(\ell)} + (1-\lambda) \mathbf{z}_{1,2}^{(\ell)}, \end{align}\] where \(\mathbf{z}^{(\ell)}_{1,1} = \mathbf{z}^{(\ell)}(\psi_{1}, \phi_{1,1})\) and \(\mathbf{z}^{(\ell)}_{1,2} = \mathbf{z}^{(\ell)}(\psi_{1}, \phi_{1,2})\) are learnable parameters. For the interpolation weight \(\lambda\), the algorithm sets \(\lambda=\frac{d(\phi, \phi_{1,1})}{d(\phi_{1,1}, \phi_{1,2})}\) where \(d(\cdot, \cdot)\) is the difference in the longitude for isolatitude points, e.g., \(d(\phi_{1,1}, \phi_{1,2}) = \phi_{1,2} - \phi_{1,1}\) if the interval \(\Phi_{1}\) does not overlap with the prime meridian. One can also compute \(\mathbf{z}^{(\ell)}(\psi_{2}, \phi)\) similarly using \(\mathbf{z}^{(\ell)}_{2,1}\) and \(\mathbf{z}^{(\ell)}_{2,2}\).

Finally, the algorithm computes the final positional feature by interpolation along the latitude: \[\mathbf{z}^{(\ell)}(\psi, \phi) = \mu \mathbf{z}^{(\ell)}(\psi_{1}, \phi) + (1-\mu) \mathbf{z}^{(\ell)}(\psi_{2}, \phi),\] where the interpolation weight is set to \(\mu = \frac{\psi - \psi_{1}}{\psi_{2} - \psi_{1}}\).

3.2 Equirectangular Feature-Grid↩︎

Figure 3: Neighborhood structure of the equirectangular grid. The yellow point indicates the input point \mathbf{x}_{\text{in}} and green points indicate the neighborhood grid points \mathcal{N}^{(\ell)}(\mathbf{x}_{\text{in}}).
Figure 4: Pole singularity and periodicity of equirectangular grid. Different spherical coordinates can indicate the same point on a sphere, i.e., points at the North pole (\psi=-\frac{\pi}{2}), South pole (\psi=\frac{\pi}{2}), and the prime meridian (\phi=0 and \phi=2\pi). Our HNeR-S avoided assigning different parameters for such spherical coordinates. Points with the same marker and color share the associated parameters.

We use HNeR-S based on the equirectangular feature-grid for weather and climate data since the data is also distributed along an equirectangular grid. It is based on a projection that maps meridians (isolongitude lines) to vertical straight lines of constant spacing and circles of latitude (isolatitude lines) to horizontal straight lines of constant spacing. We index each point in the \(\ell\)-th grid by a tuple of integers \((n, m)\) with the position \(\mathbf{x}_{n, m}^{(\ell)}\) described as follows: \[\begin{align} \mathbf{x}_{n,m}^{(\ell)} = (\psi_{n}^{(\ell)}, \phi_{m}^{(\ell)}) = \left(\pi\left(\frac{n}{N_{\text{lat}}^{(\ell)}} - \frac{1}{2}\right), 2\pi\frac{m}{N_{\text{lon}}^{(\ell)}}\right). \end{align}\] The non-negative integers \(n \leq N_{\text{lat}}^{(\ell)}\) and \(m \leq N_{\text{lon}}^{(\ell)}\) index the point-wise latitude and longitude, respectively. The integers \(N_{\text{lat}}^{(\ell)}, N_{\text{lon}}^{(\ell)}\) decide the latitude-wise and longitude-wise resolutions of the feature-grid. We choose them by: \[\begin{align} N_\text{lat}^{(\ell)}=\lfloor \gamma^{\ell-1}\cdot N_\text{lat}^{(1)}\rfloor, \quad N_\text{lon}^{(\ell)}=\lfloor \gamma^{\ell-1}\cdot N_\text{lon}^{(1)}\rfloor, \end{align}\] where \(\gamma > 1\) is the scaling factor, \(\lfloor\cdot\rfloor\) is the floor function, and \(N_\text{lat}^{(1)}, N_\text{lon}^{(1)}\) are resolutions of the coarsest grid (\(\ell=1\)).

Neighborhood structure. Given an input point \(\mathbf{x} = (\psi, \phi)\) as an input of the CNR, the neighborhood forms a unit cell of the feature-grid as follows: \[\begin{align} \mathcal{N}^{(\ell)}(\mathbf{x}) &= \{(n, m), (n, m+1), (n+1, m), (n+1, m+1)\}. \end{align}\] Our framework determines the index \((n, m)\) to ensure the position \(\mathbf{x}\) is inside the rectangle: \[(n, m) = \left(\left\lfloor \left(\frac{\psi}{\pi}+\frac{1}{2}\right)N^{(\ell)}_{\text{lat}} \right\rfloor, \left\lfloor \frac{\phi}{2\pi}N^{(\ell)}_{\text{lon}} \right\rfloor\right).\] We describe the neighborhood structure in 3.

Non-uniformity. Since the equirectangular projection does not preserve area, the resulting equirectangular grid is irregular and the cell-wise areas are non-uniformly distributed. To incorporate this during training and evaluation, one needs to reweight each data point by the corresponding cell area [16]. The non-uniformity of the equirectangular grid is illustrated in 2.

Pole singularity. It is important to note that the equirectangular grid assigns multiple points on a single pole, i.e., \(\mathbf{x}_{0,m} = \mathbf{x}_{0,m'}\) for any \(m, m' < N_{\text{lon}}\). To incorporate this fact, we merge the associated parameters into a single parameter for both training and evaluation. We also remark on how the interpolation of our algorithm is well-defined even for the input points near the poles. We illustrate the pole singularity in 4.

Periodicity. The equirectangular projection of maps inherently introduces discontinuity, where the grid points sharing identical coordinates find themselves at the opposite ends of the map, on the left and right sides of the prime meridian, as described in 4. To address this issue, we merge the parameters corresponding to grid points at the same latitude from both ends of the map into a unified parameter.

3.3 HEALPix Grid↩︎

Figure 5: Neighborhood structure of the HEALPix grid. The yellow point indicates the input point \mathbf{x} and black points indicate the neighborhood grid points \mathcal{N}^{(\ell)}(\mathbf{x}_\text{in}).

We consider feature-grids with hierarchical equal area isolatitude pixelation (HEALPix) structure for the CMB data [40], since the data points are distributed according to the HEALPix grid structure. Notably, the data points are located in the center of the HEALPix grid cells. The grid structure uniformly partitions the sphere and the data points are uniformly distributed over the sphere.

In particular, the HEALPix center points are located on \(N_{\text{line}}\) lines of constant latitude, and each line (or azimuth) is uniformly divided by the points. The lines with latitude \(\psi\) satisfying \(|\cos \psi| \leq \frac{2}{3}\) are divided into the same number of pixels \(N_{\text{eq}}\). The remaining lines are located within the polar cap (\(|\cos \psi| > \frac{2}{3}\)) and contain a varying number of pixels, increasing from line to line with increasing distance from the poles by one pixel within each quadrant. We refer to [40] for a detailed description.

The resolution of the grid is expressed by the parameter \(N_{\text{side}}\), which denotes the number of divisions along the side of a base-resolution pixel that is needed to reach a desired high-resolution partition. One can observe that \(N_{\text{line}}=4N_{\text{side}}-1\) and \(N_{\text{eq}}=4N_{\text{side}}\). This results in \(N_\text{pix}^{(\ell)} = 12 (N_\text{side}^{(\ell)})^2\) pixels and the corresponding center points. The resolution increases by the power of two for each level, i.e., we set \(N_{\text{side}}^{(\ell)}=2^{\ell-1}\) for the \(\ell\)-th grid.

Neighborhood structure. We choose the neighborhood \(\mathcal{N}^{(\ell)}(\mathbf{x}_\text{in})\) in the HEALPix grid as the center of pixels that are adjacent to the pixel that includes the point \(\mathbf{x}_\text{in}\). Furthermore, while a pixel in the HEALPix grid is adjacent to at most eight pixels, we choose the neighborhood as the center of the closest pixels on the two lines above and below the location. Such neighborhood allows the application of bilinear interpolation as described in 3.1. We illustrate the HEALPix neighborhood structure in 5.

4 Experiments↩︎

In this section, we assess the performance of our two hierarchical hybrid neural representations designed for spherical data: the equirectangular grid and the HEALPix grid. To this end, we first validate our framework for climate data regarding four tasks: regression, super-resolution, temporal interpolation, and compression (4.1). We also explore two tasks on CMB data with the HEALPix grid regarding super-resolution and regression (4.2). We will release our code upon acceptance.

Table 1: Details on the datasets used for weather and climate (ERA5) and CMB (PR1) data.
ERA5 PR1
Grid Equirectangular HEALPix
Target Geopotential, Temperature CMB Temperature
Spatial Res. \(0.25^\circ\), \(0.50^\circ\), \(1.00^\circ\) \(5\) arcmin
Temporal Res. Daily, Weekly -

Datasets. Mainly, we utilized climate and CMB data for our experiment. Weather and climate data is gathered from ECMWF reanalysis 5th generation [5] archive where data can be directly downloaded by using Climate Data Storage API [41]. To be specific, we gathered geopotential and temperature data in the year 2000 for super-resolution, regression, and temporal interpolation. For the compression task, we gathered geopotential data in the year 2016 following [16]. The detailed description of datasets is in 1.

Next, we use the CMB temperature data from Planck Public Data Release 1 (PR1) Mission Science Maps data at NASA/IPAC Infrared Science Archive (IRSA)1. We use the version where missing values are determined by spectral matching independent component analysis [42].

Baselines. On one hand, for super-resolution, regression, and temporal interpolation tasks, we compare our methods to five different encoding schemes that are applicable for learning spherical signals: logarithmically spaced sinusoidal positional encoding [4], sinusoidal representation networks [22], wavelet implicit neural representation [23], spherical harmonics implicit neural representation [12], and Fourier feature networks [24].

On the other hand, for the compression task, we compare our methods to the state-of-the-art compression scheme without neural network [43] and the recently proposed coordinate-based neural network for compression [16].

Experimental setting. We conduct all experiments using a single RTX 3090 GPU. For the regression, super-resolution, and temporal interpolation experiments, we employ a consistent architecture, consisting of a 4-layer multilayer perceptron (MLP) with 256 units in each hidden layer. Detailed settings for the compression experiment are provided in 4.1.

4.1 Weather and Climate Data↩︎

Table 2: Results of regression on weather and climate data. The evaluation metric is weighted PSNR and the best metric is highlighted in bold.
Geopotential Temperature
ReLU+P.E. 60.20 55.63
SIREN 55.10 55.20
SHINR 54.98 45.75
WIRE 65.93 59.81
FFN 61.15 57.14
HNeR-S (ours) 66.17 63.45

Figure 6: image.

Figure 7: Prediction (top) and error maps (bottom) for \times2 super-resolution in climate and weather temperature data.

In this section, we evaluate our model on weather and climate data for \(\times2\) and \(\times4\) super-resolution, regression, temporal interpolation, and compression. The climate data is often represented using an equirectangular grid (i.e., a regular latitude-longitude grid), where data distribution is concentrated around the poles. To address this unique challenge, we employ our hierarchical equirectangular feature-grid. In addition, every task on an equirectangular grid is evaluated by latitude-weighted metric, i.e., weighted RMSE, MAE, and PSNR [16]. This assigns smaller weights to the poles, compensating for the irregularity of data sampled from the equirectangular grid.

Figure 8: Test PSNR curves. These plots show the test PSNR curve for (\times 4) super-resolution (left) and regression (middle) for the temperature of weather and climate data and regression for CMB data (right). Our model converges to the best PSNR in the fastest manner.

Regression. We first evaluate our framework for the regression task where only a portion of the grid is used for training and the rest is used for evaluation. We report the results in 2. One can observe how our method consistently demonstrates superiority in PSNR performance and faster convergence over the baselines. We further visualize our results in 8. The figure illustrates our method’s accuracy in regression tasks, showcasing minimal errors across diverse geographical regions.

In 8, one can observe how our algorithm demonstrates significant improvement over the baselines. The improvement is more pronounced for predicting the temperature, where our HNeR-S achieves a 63.45 PSNR score whereas the second-best baseline (WIRE) achieves 59.81.

Super-resolution. Next, we evaluate our method for the super-resolution task. To this end, we train the models on low-resolution grids and evaluate them on high-resolution grids. Precisely, we train the models on the resolution of 0.50\(^\circ\) and \(1.00^{\circ}\) evaluate them on the resolution of 0.25\(^\circ\) for the \(\times 2\) and \(\times 4\) setting, respectively.

We report the results in 6. Here, one can observe how our algorithm demonstrates significant improvements over the baselines. We also visualize the results in 7, which demonstrate our method’s ability to reconstruct high-resolution details accurately, improving clarity and definition in climate data visualizations.

Table 3: Performance for temporal interpolation on weather and climate data. The best number is highlighted in bold.
Geopotential Temperature
2-3 (lr)4-5 Daily Weekly Daily Weekly
ReLU+P.E. 28.43 23.92 27.12 24.35
SIREN 28.80 23.16 27.26 23.34
SHINR 28.46 22.75 27.27 22.90
WIRE 28.66 22.99 27.07 23.18
FFN 28.53 23.54 26.65 23.73
HNeR-S (ours) 28.79 23.97 27.55 24.50

Temporal interpolation. Finally, we evaluate the ability of our framework to interpolate between snapshots of the weather at different timesteps. This task employs two temporal resolutions: daily and weekly, with a total of 30 data snapshots sampled. Since this task requires the CNRs to condition on the timestep of the snapshot, we modify our algorithm and the baselines by incorporating the timestep \(t\) as a concatenation of the positional feature, which serves as input for the MLP.

The outcomes, detailed in 3 and 8, highlight our method’s exceptional PSNR performance. The visualizations in 7 reveal our model’s capability to accurately interpolate temporal changes.

Table 4: Results of compression on weather and climate data. The best results of spatial compression models are highlighted in bold. \(^{\dagger}\)FFN-T and SZ3 results are from [16].
\(0.031 \leq\) BPP \(\leq 0.080\) \(0.099 \leq\) BPP \(\leq 0.119\)
2-4 (lr)5-7 BPP WRMSE WMAE BPP WRMSE WMAE
\((\downarrow)\) \((\downarrow)\) \((\downarrow)\) \((\downarrow)\) \((\downarrow)\) \((\downarrow)\)
Spatiotemporal Compression
SZ3\(^{\dagger}\) 0.080 698.5 574.9 0.115 493.9 408.8
FFN-T\(^{\dagger}\) 0.031 0.111 73.6 53.0
Spatial Compression
SIREN 0.063 669.5 462.1 0.119 616.0 404.8
WIRE 0.069 2617.2 1433.3 0.099 688.7 478.6
HNeR-S (ours) 0.057 144.8 99.4 0.114 124.0 50.3
Table 5: No caption
Reference RELU+P.E. WIRE FFN Ours
[0pt][0pt]

image

image

image

image

image

[0pt][0pt]image

image

image

image

image

Compression. As an additional experiment we apply HNeR-S for compression, driven by the demand to store daily-growing high-resolution climate and weather datasets [16]. To this end, we combine our method with the existing HNeR-based compression scheme [44] to apply image-wise compression, i.e., we train one neural network for the compression of each image. As a non-machine learning baseline, we consider a modular composable SZ3 [43] framework. We also compare with the Fourier feature network with temporal encoding [16] proposed for compression of climate and weather data. We note that FFN-T is a spatiotemporal compressor, i.e., it takes both spatial and temporal coordinate that trains a series of snapshots, while ours is a spatial compressor, i.e., it trains a network for each image. Additionally, we consider two implicit neural network baselines including WIRE [23] and SIREN [12] for fair comparison on spatial compression.

We report the results in 4. We compare the trade-off between bits per pixel (BPP) and the quality of the compressed data measured in weighted RMSE and weighted MAE.2 Here, one can observe that our method shows superior results to spatial compression baselines. Our method achieves comparable results with FFN-T, which is a spatial-temporal compression algorithm that is not directly comparable to ours. In particular, FFN-T takes advantage of sharing parameters across different time steps, while our scheme can easily incorporate the daily growth of the dataset.3

4.2 Cosmic Microwave Background Data↩︎

Here, we evaluate our model for the CMB data, which is stored as a HEALPix grid where the distribution of the data is uniform across the grids. For every CMB experiment, we used a resolution of \(N_{\text{side}}=2048\), which is the highest resolution that the Planck satellite provides. One should note that we used the highest resolution of the CMB data (i.e., \(N_\text{side}\) = 2048) where the data is often downgraded or only utilizes part of the map due to its extensive data size [45], [46]. Meanwhile, our method utilizes the full resolution and highlights the scalability and ability to learn intricate signals of the CMB data.

Super-resolution. Similar to the regression task, we can apply super-resolution for CMB data. However, the experiment setting is slightly adjusted for CMB data since the resolution of HEALPix data is defined by \(N_{\text{side}}\). Here, the \(\times 2\) setting involves training on \(N_{\text{side}}=1024\) to predict \(N_{\text{side}}=2048\), and the \(\times 4\) setting trains the model on \(N_{\text{side}}=512\) to predict \(N_{\text{side}}=2048\).

Note that training on high-resolution \(N_{\text{side}}=2048\) data is challenging due to its large size, a hurdle our CNR-based approach overcomes, unlike prior works that often use downgraded data such as \(N_{\text{side}}=64\) [47]. As can be seen in the 9 our approach outperforms the other baselines. Despite the highly intricate and noisy pattern of the CMB data, our model significantly outperforms the baselines. This implies that our model can learn complex real-world data as well.

Figure 9: image.

We provide experimental results in 9 and 5. We observe that our method consistently shows superior PSNR results. Moreover, in 5, our model converges into desirable PSNR in the early stages of the training.

Regression. Similar to the regression in climate data, we predict the unknown signal from the constructed hybrid neural representation. We demonstrate the results in 9. One can observe that our method highly outperforms compared to the baselines. As depicted in 5, both the prediction and error map indicate that our method can predict the target signal within a small error. Moreover, according to 8, our model converges to the promising accuracy by showing a large performance gap compared to the other baselines.

5 Conclusion↩︎

In this paper, we presented two new hybrid neural representations tailored for spherical data, utilizing hierarchical spherical feature-grids to generate positional encodings through spherical interpolation algorithms. Our extensive experimental evaluations across multiple datasets and tasks have verified the effectiveness of our method. An interesting avenue for future work will be the application of adaptive grid [19], [48] which will lead to a more compact feature grid. Also, there remains an exploration for interpolation methods tailored for spherical grids, i.e., geodesic-aware interpolation.

Broader Impact↩︎

Our framework can advance machine learning for applications like climatology and cosmology. Machine learning for weather and climate forecasting can benefit our society via improved resilience to extreme weather events, agriculture productivity, and water resource management. Furthermore, machine learning for cosmology can accelerate scientific discovery, in particular our understanding of dark matter and dark energy.

Acknowledgements↩︎

We would like to express our sincere gratitude to Seonghyun Park, Hyosoon Jang, Seongsu Kim, Juwon Hwang, Kiyoung Seong, and Dongyeop Woo for their valuable comments and offering essential experimental tools for our research. Also, we are grateful to Langwen Huang for providing the detailed results of their work, which served as a foundational baseline method for our research. Additionally, our thanks extend to Jihoon Tack and Doyup Lee for their constructive feedback that greatly contributed to the improvement of this paper.

6 Architecture Details↩︎

Equirectangular feature-grid. For the equirectangular feature-grid we adjust the level of the hierarchical map and the base resolution of the feature-grid, by considering the resolution of the training data. For example, comparing the task of \(\times\)​2 and \(\times\)​4 super-resolution, we assign the higher level and base resolution for \(\times\)​2 task. Mainly, we used upscaling factor \(\gamma=1.5\). We initialized the parameter for the feature-grid with uniform initialization \(\mathcal{U}(-1\mathrm{e}{-4},1\mathrm{e}{-4})\).

HEALPix feature-grid. Similarly to the equirectangular feature-grid, we adjust the feature grid level by the resolution of the training data. The choice of the level is quite intuitive. In the case for super-resolution and regression task, mostly the optimal choice was to choose the level as \(\ell_{\text{grid}}=\ell_{\text{data}}-1\). This choice was beneficial since it enables overall parameters to be trained while capturing the fine details. The parameter of the feature-grid is initialized by normal initialization \(\mathcal{N}(0,1\mathrm{e}{-4})\).

7 Baseline Details↩︎

ReLU+P.E. [4] is a positional encoding defined as: \[\gamma(x)=\left(\sin \left(2^0 \pi x\right), \cos \left(2^0 \pi x\right), \cdots, \sin \left(2^{L-1} \pi x\right), \cos \left(2^{L-1} \pi x\right)\right)\] which separately encodes each dimension of the input point \(\mathbf{x}\in\mathbb{R}^n\) with sinusoidals in a axis-aligned and logarithmically spaced manner. In this study, we adjust the level of the frequency \(L\) with \(\lbrace 2, 5, 7\rbrace\).

SIREN [22] is a sinusoidal activation \(\sin(\omega_0 W\mathbf{x}+b)\) where the frequency of a sinusoidal function is composed by linear weight and scaled by \(\omega_0\). Here, \(\omega_0\) is set as a hyperparameter and we adjust its range within \(\lbrace 10, 20, 30 \rbrace\).

FFN [24] is a random fourier feature positional encoding \(\gamma\left(\mathbf{x})=((\cos(2\pi\mathbf{B}\mathbf{x}),\sin(2\pi\mathbf{B}\mathbf{x})\right)^T\) where \(\mathbf{x}\in\mathbb{R}^n\) and \(\mathbf{B}\in \mathbb{R}^{m\times n}\) is a sampled frequency from \(\mathcal{N}(0, \sigma^2)\). In our study, we adjust \(\sigma\) in the range of \(\lbrace 2, 5, 10\rbrace\).

WIRE [23] is a Gabor wavelet based activation \(\exp({j\omega_0 \mathbf{x}})\exp({-|s_0 \mathbf{x}|^2})\), which assigns the scale (\(\omega_0\)) and width (\(s_0\)) as hyper parameters. In our study, we adjust the both scale and width parameters in the range of \(\lbrace 1, 10\rbrace\).

SHINR [12] is a spherical harmonics based positional encoding \(\gamma(\mathbf{x})=\bigoplus_{l=0}^L\bigoplus_{m=-l}^l Y_m^l(\theta,\psi)\), which takes spherical coordinate \((\theta,\psi)\) and \(L\) is a level of the encoding. In our experiment, we adjust the level \(L\) in the range of \(\lbrace 2, 3, 4\rbrace\).

8 Experimental Details↩︎

8.1 Hyperparameters.↩︎

In this section, we present the spectrum of hyperparameters applied to our approach. Without loss of generality, the hyperparameters selected for weather and climate datasets correspond to those used for the equirectangular feature grid, while those chosen for cosmic microwave background data pertain to the HEALPix feature grid. The common settings across the tasks are 4 hidden layers, 256 hidden dimensions, ReLU [49] activation and AdamW optimizer [50] for the MLP.

Table 6: Hyperparameter table for super-resolution task for geopotential, temperature and CMB data.
INR training Geo. SR (\(\times\) 2) Geo. SR (\(\times\) 4) Temp. SR (\(\times\)​2) Temp. (\(\times\)​4) CMB (\(\times\)​2) CMB (\(\times\)​4)
Level (\(L\)) \(\lbrace 8, 6\rbrace\) \(\lbrace 8, 6\rbrace\) \(\lbrace 8, 6\rbrace\) \(\lbrace 8, 6\rbrace\) \(\lbrace 11, 10\rbrace\) \(\lbrace 10, 9\rbrace\)
Parameter dim. (\(d\)) \(\lbrace 2, 3\rbrace\) \(\lbrace 2, 3\rbrace\) \(\lbrace 2, 3\rbrace\) \(\lbrace 2, 3\rbrace\) \(\lbrace 2, 3\rbrace\) \(\lbrace 2, 3\rbrace\)
Scaling factor (\(\gamma\)) \(1.5\) \(1.5\) \(1.5\) \(1.5\) \(-\) \(-\)
Base resolution \(16\) \(16\) \(16\) \(16\) \(-\) \(-\)
Table 7: Hyperparameter table for regression and temporal task for geopotential, temperature and CMB data. 1 day stands for daily and 1 week stands for weekly interpolation task.
INR training Geo. (Reg.) Temp. (Reg.) CMB (Reg.) Geo. (1 day) Temp. (1 day) Geo. (1 week) Temp. (1 week)
Level (\(L\)) \(\lbrace 8, 6\rbrace\) \(\lbrace 8, 6\rbrace\) \(\lbrace 11, 12\rbrace\) \(\lbrace 7, 8\rbrace\) \(\lbrace 6, 8\rbrace\) \(\lbrace 8\rbrace\) \(\lbrace 6, 8\rbrace\)
Parameter dim. (\(d\)) \(\lbrace 2, 3\rbrace\) \(\lbrace 2, 3\rbrace\) \(\lbrace 2, 3\rbrace\) \(\lbrace 6, 8\rbrace\) \(\lbrace 2, 3\rbrace\) \(\lbrace 8,9\rbrace\) \(\lbrace 2, 3\rbrace\)
Scaling factor (\(\gamma\)) \(1.5\) \(1.5\) \(-\) \(1.5\) \(1.5\) \(\lbrace 1.1, 1.2\rbrace\) \(1.5\)
Base resolution \(16\) \(16\) \(-\) \(16\) \(4\) \(4\) \(16\)

8.2 Training Details.↩︎

We train our model with weighted RMSE [16] and used AdamW optimizer [50] with learning rate of 1e-5. For the compression task, we trained our model with . We utilize the C3 [51] framework for compression task which also utilizes the learnable positional encoding to encode the data. Specifically, we applied our equirectangular grid feature instead of the original learnable positional grid which is targeted for a single or a series of 2D image data.

8.3 Task Details.↩︎

Super-resolution. The study applies two settings for enhancing the resolution of CMB, weather, and climate data: \(\times2\) and \(\times4\). In the \(\times2\) setting, the model is trained using \(361 \times 720\) training points, each representing a 0.50\(^\circ\) degree. For the \(\times4\) setting, training involves \(181\times 360\) points, with each point representing a 1.00\(^\circ\) degree. The model’s performance is assessed on its ability to predict data at a finer resolution of 0.25\(^\circ\), across a total of \(721\times 1440\) points. In the context of CMB super-resolution, the \(\times2\) setting involves training on data downsampled to \(N_\text{side}=1024\), yielding approximately 10M training points at a resolution of about \(N\mu\)min. For the \(\times4\) setting, the training data is further downsampled to \(N_\text{side}=512\), with around 30K points. The prediction task targets data with \(N_\text{side}=2048\), encompassing about 50M points at the full resolution provided by the Planck satellite.

Regression. The regression task involves dividing data points into training and testing sets in an 8:2 ratio. Specifically, for weather and climate data, the split is conducted on geopotential and temperature datasets that total \(721\times 1440\) points, resulting in approximately 83K training points and 20K test points. For CMB data, the split includes 50M points, allocating 40M for training and 10M for testing.

Temporal interpolation. This task employs two temporal resolutions: daily and weekly, with a total of 30 data snapshots sampled. Here, snapshots indicate the data that represents the captured moment. The spacing between snapshots corresponds to the chosen temporal resolution. Training data is constructed by selecting the first snapshot, skipping the next, and then selecting the third snapshot. Conversely, test data is compiled by skipping the first snapshot, selecting the second, and continuing this pattern. This method ensures a consistent and desired time step between training and test datasets.

Compression. The two settings of the compression task are divided by the bits per pixel (BPP). The BPP is computed by dividing the model size by the number of pixels. On one hand, for the spatiotemporal compression task, the number of pixels is the multiplication of the number of time indices, pressure levels, and the number of latitude and longitude levels. In our dataset, they are 366, 11, 361, and 720, respectively, following [16]. On the other hand, for the spatial compression task, the number of pixels is the multiplication of the number of latitude and longitude levels.

References↩︎

[1]
Xie, Y., Takikawa, T., Saito, S., Litany, O., Yan, S., Khan, N., Tombari, F., Tompkin, J., Sitzmann, V., and Sridhar, S. Neural fields in visual computing and beyond. arXiv preprint 2111.11426, 2021.
[2]
Chen, Y., Liu, S., and Wang, X. Learning continuous image representation with local implicit image function. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 8628–8638, 2021.
[3]
Park, J. J., Florence, P., Straub, J., Newcombe, R., and Lovegrove, S. Deepsdf: Learning continuous signed distance functions for shape representation. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 165–174, 2019.
[4]
Mildenhall, B., Srinivasan, P. P., Tancik, M., Barron, J. T., Ramamoorthi, R., and Ng, R. Nerf: Representing scenes as neural radiance fields for view synthesis. Communications of the ACM, 65 (1): 99–106, 2021.
[5]
Hersbach, H., Bell, B., Berrisford, P., Biavati, G., Horányi, A., Muñoz Sabater, J., Nicolas, J., Peubey, C., Radu, R., Rozum, I., et al. Era5 hourly data on single levels from 1979 to present. Copernicus climate change service (c3s) climate data store (cds), 10 (10.24381), 2018.
[6]
de Bernardis, P., Ade, P. A., Bock, J. J., Bond, J., Borrill, J., Boscaleri, A., Coble, K., Crill, B., De Gasperis, G., Farese, P., et al. A flat universe from high-resolution maps of the cosmic microwave background radiation. Nature, 404 (6781): 955–959, 2000.
[7]
Lam, R., Sanchez-Gonzalez, A., Willson, M., Wirnsberger, P., Fortunato, M., Alet, F., Ravuri, S., Ewalds, T., Eaton-Rosen, Z., Hu, W., et al. Graphcast: Learning skillful medium-range global weather forecasting. arXiv preprint arXiv:2212.12794, 2022.
[8]
Bi, K., Xie, L., Zhang, H., Chen, X., Gu, X., and Tian, Q. Pangu-weather: A 3d high-resolution model for fast and accurate global weather forecast. arXiv preprint arXiv:2211.02556, 2022.
[9]
Bennett, C. L., Banday, A. J., Górski, K. M., Hinshaw, G., Jackson, P., Keegstra, P., Kogut, A., Smoot, G. F., Wilkinson, D. T., and Wright, E. L. Four-year cobe* dmr cosmic microwave background observations: maps and basic results. The Astrophysical Journal, 464 (1): L1, 1996.
[10]
Spergel, D. N., Verde, L., Peiris, H. V., Komatsu, E., Nolta, M., Bennett, C. L., Halpern, M., Hinshaw, G., Jarosik, N., Kogut, A., et al. First-year wilkinson microwave anisotropy probe (wmap)* observations: determination of cosmological parameters. The Astrophysical Journal Supplement Series, 148 (1): 175, 2003.
[11]
Ade, P. A., Aghanim, N., Alves, M., Armitage-Caplan, C., Arnaud, M., Ashdown, M., Atrio-Barandela, F., Aumont, J., Aussel, H., Baccigalupi, C., et al. Planck 2013 results. i. overview of products and scientific results. Astronomy & Astrophysics, 571: A1, 2014.
[12]
Esteves, C., Lu, T., Suhail, M., fan Chen, Y., and Makadia, A. Generalized fourier features for coordinate-based learning of functions on manifolds, 2022. URL https://openreview.net/forum?id=g6UqpVislvH.
[13]
Grattarola, D. and Vandergheynst, P. Generalised implicit neural representations. Advances in Neural Information Processing Systems, 35: 30446–30458, 2022.
[14]
Koestler, L., Grittner, D., Moeller, M., Cremers, D., and Lähner, Z. Intrinsic neural fields: Learning functions on manifolds. In European Conference on Computer Vision, pp. 622–639. Springer, 2022.
[15]
Schwarz, J. R., Tack, J., Teh, Y. W., Lee, J., and Shin, J. Modality-agnostic variational compression of implicit neural representations. arXiv preprint arXiv:2301.09479, 2023.
[16]
Huang, L. and Hoefler, T. Compressing multidimensional weather and climate data into neural networks. In The Eleventh International Conference on Learning Representations, 2023. URL https://openreview.net/forum?id=Y5SEe3dfniJ.
[17]
Chabra, R., Lenssen, J. E., Ilg, E., Schmidt, T., Straub, J., Lovegrove, S., and Newcombe, R. Deep local shapes: Learning local sdf priors for detailed 3d reconstruction. In Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XXIX 16, pp. 608–625. Springer, 2020.
[18]
Müller, T., Evans, A., Schied, C., and Keller, A. Instant neural graphics primitives with a multiresolution hash encoding. ACM Transactions on Graphics (ToG), 41 (4): 1–15, 2022.
[19]
Martel, J. N., Lindell, D. B., Lin, C. Z., Chan, E. R., Monteiro, M., and Wetzstein, G. Acorn: Adaptive coordinate networks for neural scene representation. arXiv preprint arXiv:2105.02788, 2021.
[20]
Liu, L., Gu, J., Zaw Lin, K., Chua, T.-S., and Theobalt, C. Neural sparse voxel fields. Advances in Neural Information Processing Systems, 33: 15651–15663, 2020.
[21]
Peng, S., Zhang, Y., Xu, Y., Wang, Q., Shuai, Q., Bao, H., and Zhou, X. Neural body: Implicit neural representations with structured latent codes for novel view synthesis of dynamic humans. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 9054–9063, 2021.
[22]
Sitzmann, V., Martel, J., Bergman, A., Lindell, D., and Wetzstein, G. Implicit neural representations with periodic activation functions. Advances in neural information processing systems, 33: 7462–7473, 2020.
[23]
Saragadam, V., LeJeune, D., Tan, J., Balakrishnan, G., Veeraraghavan, A., and Baraniuk, R. G. Wire: Wavelet implicit neural representations. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 18507–18516, 2023.
[24]
Tancik, M., Srinivasan, P., Mildenhall, B., Fridovich-Keil, S., Raghavan, N., Singhal, U., Ramamoorthi, R., Barron, J., and Ng, R. Fourier features let networks learn high frequency functions in low dimensional domains. Advances in Neural Information Processing Systems, 33: 7537–7547, 2020.
[25]
Rahaman, N., Baratin, A., Arpit, D., Draxler, F., Lin, M., Hamprecht, F., Bengio, Y., and Courville, A. On the spectral bias of neural networks. In International Conference on Machine Learning, pp. 5301–5310. PMLR, 2019.
[26]
Jacot, A., Gabriel, F., and Hongler, C. Neural tangent kernel: Convergence and generalization in neural networks. Advances in neural information processing systems, 31, 2018.
[27]
Huang, L. and Hoefler, T. Compressing multidimensional weather and climate data into neural networks. arXiv preprint arXiv:2210.12538, 2022.
[28]
Yuval, J. and O’Gorman, P. A. Stable machine-learning parameterization of subgrid processes for climate modeling at a range of resolutions. Nature communications, 11 (1): 3295, 2020.
[29]
Wang, J., Liu, Z., Foster, I., Chang, W., Kettimuthu, R., and Kotamarthi, V. R. Fast and accurate learned multiresolution dynamical downscaling for precipitation. Geoscientific Model Development, 14 (10): 6355–6372, 2021.
[30]
Yang, H., Yan, J., Liu, Y., and Song, Z. Statistical downscaling of numerical weather prediction based on convolutional neural networks. Global Energy Interconnection, 5 (2): 217–225, 2022.
[31]
Wang, X., Yu, K., Wu, S., Gu, J., Liu, Y., Dong, C., Qiao, Y., and Change Loy, C. Esrgan: Enhanced super-resolution generative adversarial networks. In Proceedings of the European conference on computer vision (ECCV) workshops, pp. 0–0, 2018.
[32]
Stengel, K., Glaws, A., Hettinger, D., and King, R. N. Adversarial super-resolution of climatological wind and solar data. Proceedings of the National Academy of Sciences, 117 (29): 16805–16815, 2020.
[33]
Dupont, E., Loya, H., Alizadeh, M., Goliński, A., Teh, Y. W., and Doucet, A. Coin++: Neural compression across modalities. arXiv preprint arXiv:2201.12904, 2022.
[34]
Cohen, T. S., Geiger, M., Köhler, J., and Welling, M. Spherical CNNs. In International Conference on Learning Representations, 2018. URL https://openreview.net/forum?id=Hkbd5xZRb.
[35]
Bonev, B., Kurth, T., Hundt, C., Pathak, J., Baust, M., Kashinath, K., and Anandkumar, A. Spherical fourier neural operators: Learning stable dynamics on the sphere. In International Conference on Learning Representations, 2023. URL https://openreview.net/forum?id=TwsJ9IOZDx.
[36]
Hortúa, H. J., Volpi, R., Marinelli, D., and Malagò, L. Parameter estimation for the cosmic microwave background with bayesian neural networks. Physical Review D, 102 (10): 103509, 2020.
[37]
Caldeira, J., Wu, W. K., Nord, B., Avestruz, C., Trivedi, S., and Story, K. T. Deepcmb: Lensing reconstruction of the cosmic microwave background with deep neural networks. Astronomy and Computing, 28: 100307, 2019.
[38]
Wang, G.-J., Shi, H.-L., Yan, Y.-P., Xia, J.-Q., Zhao, Y.-Y., Li, S.-Y., and Li, J.-F. Recovering the cmb signal with machine learning. The Astrophysical Journal Supplement Series, 260 (1): 13, 2022.
[39]
Casas, J., Bonavera, L., González-Nuevo, J., Baccigalupi, C., Cueli, M., Crespo, D., Goitia, E., Santos, J., Sánchez, M., and de Cos, F. Cenn: A fully convolutional neural network for cmb recovery in realistic microwave sky simulations. arXiv preprint arXiv:2205.05623, 2022.
[40]
Gorski, K. M., Hivon, E., Banday, A. J., Wandelt, B. D., Hansen, F. K., Reinecke, M., and Bartelmann, M. Healpix: A framework for high-resolution discretization and fast analysis of data distributed on the sphere. The Astrophysical Journal, 622 (2): 759, 2005.
[41]
Buontempo, C., Hutjes, R., Beavis, P., Berckmans, J., Cagnazzo, C., Vamborg, F., Thépaut, J.-N., Bergeron, C., Almond, S., Amici, A., et al. Fostering the development of climate services through copernicus climate change service (c3s) for agriculture applications. Weather and Climate Extremes, 27: 100226, 2020.
[42]
Delabrouille, J., Cardoso, J.-F., and Patanchon, G. Multidetector multicomponent spectral matching and applications for cosmic microwave background data analysis. Monthly Notices of the Royal Astronomical Society, 346 (4): 1089–1102, 2003.
[43]
Liang, X., Zhao, K., Di, S., Li, S., Underwood, R., Gok, A. M., Tian, J., Deng, J., Calhoun, J. C., Tao, D., et al. Sz3: A modular framework for composing prediction-based error-bounded lossy compressors. IEEE Transactions on Big Data, 9 (2): 485–498, 2022.
[44]
Ladune, T., Philippe, P., Henry, F., Clare, G., and Leguay, T. Cool-chic: Coordinate-based low complexity hierarchical image codec. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 13515–13522, 2023.
[45]
Krachmalnicoff, N. and Tomasi, M. Convolutional neural networks on the healpix sphere: a pixel-based algorithm and its application to cmb data analysis. Astronomy & Astrophysics, 628: A129, 2019.
[46]
Montefalcone, G., Abitbol, M. H., Kodwani, D., and Grumitt, R. Inpainting cmb maps using partial convolutional neural networks. Journal of Cosmology and Astroparticle Physics, 2021 (03): 055, 2021.
[47]
Defferrard, M., Milani, M., Gusset, F., and Perraudin, N. Deepsphere: a graph-based spherical cnn. arXiv preprint arXiv:2012.15000, 2020.
[48]
Martinez-Castellanos, I., Singer, L. P., Burns, E., Tak, D., Joens, A., Racusin, J. L., and Perkins, J. S. Multiresolution healpix maps for multiwavelength and multimessenger astronomy. The Astronomical Journal, 163 (6): 259, 2022.
[49]
Agarap, A. F. Deep learning using rectified linear units (relu). arXiv preprint arXiv:1803.08375, 2018.
[50]
Loshchilov, I. and Hutter, F. Decoupled weight decay regularization. arXiv preprint arXiv:1711.05101, 2017.
[51]
Kim, H., Bauer, M., Theis, L., Schwarz, J. R., and Dupont, E. C3: High-performance and low-complexity neural compression from a single image or video. arXiv preprint arXiv:2312.02753, 2023.

  1. https://irsa.ipac.caltech.edu/data/Planck/release_1/all-sky-maps/↩︎

  2. Weights are assigned based on varying cell areas as in training.↩︎

  3. Our scheme can train a small network only for the new data, while FFN-T requires re-training on the updated dataset.↩︎