Beyond holography:
the entropic quantum gravity foundations of anisotropic diffusion


Abstract

Recently, thanks to the development of artificial intelligence (AI) there is increasing scientific attention in establishing the connections between theoretical physics and AI. Traditionally, these connections have been focusing mostly on the relation between string theory and image processing and involve important theoretical paradigms such as holography. Recently G. Bianconi has formulated the Gravity from Entropy (GfE) approach to quantum gravity in which gravity is derived from the geometric quantum relative entropy (GQRE) between two metrics associated with the Lorentzian spacetime. Here it is demonstrated that the famous Perona-Malik algorithm for image processing is the gradient flow of the GfE action in its simple warm-up scenario. Specifically, this algorithm is the outcome of the minimization of the GQRE between two Euclidean metrics: the one of the support of the image and the one induced by the image. As the Perona-Malik algorithm is known to preserve sharp contours, this implies that the GfE action, does not in general lead to uniform images upon iteration of the gradient flow dynamics as it would be intuitively expected from entropic actions maximising classical entropies. Rather, the outcome of the minimization of the GQRE is compatible with the preservation of complex structures. These results provide the geometrical and information theory foundations for the Perona-Malik algorithm and might contribute to establish deeper connections between GfE, machine learning and brain research.

Recently there is an increasing recognition of the common mathematical foundations of theoretical physics and artificial intelligence (AI) algorithms [1][8]. Specifically there is a growing consensus on the fundamental role of topology and geometry to inform the most recent developments of AI and network theory. This scientific interest is currently leading to the fast developments of very vibrant research fields such as topological and geometrical machine learning [9], [10] and to topological higher-order network dynamics [7], [11], [12] that might lead to a more comprehensive understanding of brain dynamics [13][17]. At the interface between problems in brain research an in AI, a key research question is whether theoretical physics can inspire geometric diffusion models [18], [19] that can be crucial to develop the foundation of the next generation of unsupervised or self-supervised learning algorithms [20]. In addition to geometry and topology, information theory is also recognized as fundamental for both AI [21], [22] and brain research [23]. In particular, the relative entropy, also known as Kullback-Leibler entropy, is acquiring a fundamental role in AI and has given rise to very successful theoretical concepts and algorithms as the information bottleneck principle [24] and diffusion models [25], [26]. In this work we define the geometric quantum relative entropy (GQRE) in Euclidean spaces and we show that the GQRE provides solid information theory foundations to geometric anisotropic diffusion algorithms. By doing so, we establish the connection between these anisotropic diffusion algorithms and the recently proposed Gravity from entropy (GfE) approach to quantum gravity [27].

The relation between artificial intelligence, network theory and theoretical physics has a long history. Theoretical physics inspiration from computer vision has led to important conceptual frameworks such as the formulation of the holographic principle [28][30]. The holographic principle was originally motivated [28] to obtain the area law for the entropy of black-holes. More in general, this principle states that our three-dimensional universe might be encoded in a two dimensional surface as an hologram and leading to a very vibrant and active research direction in theoretical physics [29], [31].

On the other side, important proposals to reconcile image processing algorithms and in general AI algorithm with theoretical physics are at the forefront of physics-inspired AI [18], [19]. The root of this field can be traced back to the work [32] by Sochen, Kimmel and Malladi. This work is the first work that has framed image processing into a differential geometry theory involving the metric induced by the image. Moreover this work has proposed the adoption of the Polyakov action of string theory in order to formulate modified anisotropic diffusion models for image processing.

The cross-fertilization of ideas between theoretical physics and specifically quantum information and network science is also very fertile [33][39]. In this context, quantum entropy has found applications in the characterization of complex network structure, thanks to the use of the Von Neumann entropy associated with the graph Laplacian. This so-called Von Neumann entropy of networks has been originally proposed in Ref. [33] by Passerini and Severini and since then has found wide applications in the theory of simple and multilayer networks [34][36], [38] and provides the underlying theoretical framework for the recently proposed renormalization group approach in network theory [37]. However, the quantum relative entropy that is so central in quantum information theory  [40] has not yet been recognized to have wide applications in either AI or network science.

Recently, in Ref. [27], the author has combined quantum mechanics with gravity in a comprehensive statistical mechanics and quantum information theory framework known as Gravity from Entropy (GfE). Specifically she has formulated a geometrical definition of quantum relative entropy called GQRE in Lorentzian spacetimes. In this framework the metrics associated with spacetime are treated as quantum operators and the GfE action is determined by a Lagrangian given by the GQRE between the metric of the manifold and the metric induced by the matter field and curvature. Thus, the action of GfE is a quantum information theory action that fully captures the interplay between geometry and the matter fields defined on the manifold. Interestingly, the GQRE can also be calculated for the Schwarzschild black hole giving rise to an area law for large Schwarzschild radius [41] without obeying the holographic principle.

Here, we demonstrate that the action of GfE, is key for geometric anisotropic diffusion models. In particular we reveal that in one of its simplest incarnations, the GfE action, associated to a GQRE Lagrangian defined on Euclidean spaces, provides the quantum information theory foundations for the famous Perona-Malik algorithm  [42] for image processing. The Perona-Malik algorithm is a fundamental reference model in the field of image processing [43][45] that denoises an image by implementing a diffusion process. This diffusion is however anisotropic and takes place in presence of a metric that is chosen to have an ad hoc functional form dependent on the contrast of the image. Previous attempts to relate the Perona-Malik algorithm to string theories and in particular the Polyakov action, [32] have not been able to justify the empirical choice of the functional form of the metric assumed in the Perona-Malik algorithm. Here we provide solid quantum information theory foundations for the Perona-Malik algorithm by defining the GQRE in an Euclidean setting and showing that the Perona-Malik algorithm is given by the gradient flow of the GfE action proposed in Ref. [27]. The definition of the GQRE is here shown to be rooted in the theory of von Neumann algebras  [46], [47] and to constitute a significant advance with respect to the Araki entropy [48], [49]. The GQRE provides an information theory quantity that is built from two metrics associated to the image: the true metric of the 2D flat Euclidean space that offers the support to the image, and the metric induced by the image defined on it. Anisotropic diffusion emerges from the tension between these two metrics to reduce their differences quantified through the GQRE Lagrangian of GfE. As the Perona-Malik is known to preserve sharp contours of the image, the GfE foundation of the Perona-Malik action also points out an important property of the GfE action. Indeed, the minimization of the GfE action is compatible with the preservation of structure and complexity of the image. This is a notable feature of the GfE action as it reveals that, contrary to the expectations in scenarios when the classical entropy is maximized, the minimization of the GfE action, based on the GQRE Lagrangian, is compatible with non-homogeneous and complex outcomes.

Figure 1: Schematic representation of the metrics associated to a black-and-white image and the GQRE Lagrangian of the GfE action which provides the foundations of the Perona-Malik algorithm. The flat 2D Euclidean manifold \Omega (in green) that offers the support of the image, has points {\boldsymbol{r}} of coordinates {\boldsymbol{r}}=(x_1,x_2) and an Euclidean metric g_{\mu\nu}=\eta_{\mu\nu}. The infinitesimal distance ds between points in \Omega is defined by this metric and obeys ds^2=g_{\mu\nu}dx^{\mu}dx^{\nu}. The image is associated to the surface \mathcal{K} (in orange) embedded in a flat 3D Euclidean metric,thus every point of \mathcal{K} has coordinates ({\boldsymbol{r}}, \phi({\boldsymbol{r}})). The infinitesimal distance d\hat{s} between two points in this surface obeys d\hat{s}^2=G_{\mu\nu}dx^{\mu}dx^{\nu} where G_{\mu\nu}, given by Eq.(3 ), is the symmetric rank 2 tensor that defines the metric induced by the surface \mathcal{K} on the 2D flat support of the image \Omega. The GfE action, associated to the Lagrangian \mathcal{L}, is given by the GQRE between the metrics {\boldsymbol{G}} and g. In this work we show that the Perona-Malik algorithm is the gradient flow of the GfE action. Therefore the Perona-Malik algorithm emerges from the principle of minimization of the GQRE between the metric induced by the image {\boldsymbol{G}}_{\mu\nu} and the metric g_{\mu\nu} of the 2D image support.

The Perona-Malik algorithm. We consider a \(2D\) flat Euclidean manifold \(\Omega\) of coordinates \({\boldsymbol{r}}=(x_1,x_2)\in \Omega\) with metric \(g_{\mu\nu}=\eta_{\mu\nu}\) with \(\eta_{\mu\nu}=1\) if \(\mu=\nu\) and \(\eta_{\mu\nu}=0\) otherwise. The infinitesimal distance \(ds\) between points in \(2D\) is defined by this metric and obeys \(ds^2=g_{\mu\nu}dx^{\mu}dx^{\nu}\). Note that the metric \(g_{\mu\nu}\) and its inverse \(g^{\mu\nu}\) will be central to transform vectors in one-forms and vice versa by lowering or raising the indices, i.e. \[\begin{align} g_{\mu\nu}V^{\nu}=V_{\mu},\quad g^{\mu\nu}V_{\nu}=V^{\mu}. \end{align}\]

On top of this manifold we define a function \(\phi({\boldsymbol{r}})\in \mathbb{R}\) indicating the intensity of the colour of the single colour, black-white image in our simple setting. Given an initial noisy image determined by the function \(\psi({\boldsymbol{r}})\) the Perona-Malik algorithm [42], [44], [45] proposes to reconstruct the true image by performing a Laplace-Beltrami diffusion with metric. Specifically, the reconstructed image \(\phi({\boldsymbol{r}})\) is found by integrating the system \[\begin{align} \frac{d\phi({\boldsymbol{r}},t)}{dt}&=&\nabla_{\mu}\rho(|\nabla\phi|^2)\nabla^{\mu}\phi({\boldsymbol{r}},t)\nonumber\\ \phi({\boldsymbol{r}},0)&=&\psi({\boldsymbol{r}}). \label{PM} \end{align}\tag{1}\] where the metric \(\rho(|\nabla\phi|^2)\) is taken to be \[\begin{align} \rho(|\nabla\phi|^2)=\frac{1}{1+\alpha |\nabla\phi|^2}, \label{rho} \end{align}\tag{2}\] and where \(\alpha\in \mathbb{R}^+\) is a parameter of the model. The power and beauty of this model is that the metric evolves together with the diffusion model and the reconstruction of the image. It is clearly highly desirable to derive the specific functional form of the metric \(\rho(|\nabla\phi|^2)\) in a principled way. However, in the original work of Perona and Malik [42] there are no fundamental information theory principle driving this choice which remains an ad hoc choice in order to achieve anisotropic diffusion and a good performance of the algorithm. Additionally, also the string theory approach to anisotropic diffusions proposed in Ref. [32] does not provide this explanation.

As we will see in this article, this particular choice of the metric \(\rho(|\nabla\phi|^2)\) is exactly what is predicted by the GfE action.

Induced metric. The set of points \(({\boldsymbol{r}},\phi({\boldsymbol{r}}))\) with \({\boldsymbol{r}}\in \Omega\) defines a \(2D\) surface \(\mathcal{K}\) immersed in \(3D\) (see Figure 1). At any given point of \(\mathcal{K}\), the tangent vectors are given by \({\boldsymbol{e}}_1=(1,0,\nabla_{x^1}\phi)\) and \({\boldsymbol{e}}_2=(0,1,\nabla_{x^2}\phi)\). Let us assume that the 3D embedding space of \(\mathcal{K}\) has flat Euclidean metric with diagonal elements \((1,1,\alpha)\), where \(\alpha\) is a positive real constant. The infinitesimal distance \(d\hat{s}\) between points \({\boldsymbol{X}}\) and \({\boldsymbol{X}}+\delta{\boldsymbol{X}}\) with \(\delta {\boldsymbol{X}}={\boldsymbol{e}}_1 dx^1+{\boldsymbol{e}}_2 dx^2\) in \(\mathcal{K}\) obeys \(d\hat{s}^2=G_{\mu\nu}dx^{\mu}dx^{\nu}\) where \({G}_{\mu\nu}\) are the elements of the real and symmetric rank \(2\) tensor that defines the induced metric on \(\Omega\). Specifically, we have that the induced metric \({\boldsymbol{G}}\) on the 2D manifold \(\Omega\) that provides the support of the image, is given by \[\begin{align} { G}_{\mu\nu}=g_{\mu\nu}+\alpha \nabla_{\mu}\phi\nabla_{\nu}\phi. \label{induced} \end{align}\tag{3}\]

Thus the considered manifold \(\Omega\) is associated with two metrics; the metric \(g_{\mu\nu}\) and the metric \(G_{\mu\nu}\). Note that in the following we will use \(\hat{\boldsymbol{G}}_{\mu\nu}\) when referring to either one of these two metrics.

According to the GfE approach here we will treat both \(g\) and \({\boldsymbol{G}}\) as quantum operators and we will consider the action \(\mathcal{S}\) given by the GQRE between these two metrics. We will show that the Perona-Malik algorithm can be obtained as the gradient flow of the GfE action.
Eigenvalues of the metrics associated to the manifold. The GQRE will be defined in terms of the eigenvalues of the true metric \(g\) and the induced metric \({\boldsymbol{G}}\). In order to define the eigenvalues and eigenvectors of these two metrics in a rotationally invariant way, we define \(\lambda\) as an eigenvalue of \(\hat{\boldsymbol{G}}_{\mu\nu}\) if it solves the eigenvalue problem \[\begin{align} \hat{G}_{\mu\nu}[V^{(\lambda)}]^{\nu}=\lambda V^{(\lambda)}_{\mu}. \label{eig1} \end{align}\tag{4}\] Thus this eigenvalue problem is the usual eigenvalue problem for the matrix \(\hat{\boldsymbol{G}}g^{-1}\) as the above equation reduces to \[\begin{align} \hat{\boldsymbol{G}}_{\mu\nu}g^{\nu\rho}V_{\rho}=\lambda V_{\mu}. \end{align}\] It follows that all the eigenvalues \(\lambda^{\prime}\) of \(g_{\mu\nu}\) are equal to one \(\lambda^{\prime}_n=1\), for \(n\in \{1,2\}\) independently of the choice of the metric \(g_{\mu\nu}\) Instead, the eigenvalues \(\lambda\) and the associated (non-normalized) eigenvectors \(V_{\mu}\) of the induced metrics \({\boldsymbol{G}}_{\mu\nu}\), are given by \[\begin{align} \lambda_1=(1+\alpha|\nabla \phi|^2),\quad V_{\mu}=\nabla_{\mu}\phi\nonumber \\ \lambda_2=1,\quad \left(\eta_{\mu\nu}-\frac{\nabla_{\mu}\phi\nabla_{\nu}\phi}{|\nabla\phi|^2}\right)V^{\nu}. \label{lambda2} \end{align}\tag{5}\] An action that is only dependent on these eigenvalues will be clearly rotational invariant.
The GfE action. The GfE action [27] is associated with a Lagrangian given by the GQRE between two Lorentzian metrics: the metric of the manifold and the metric induced by the matter fields, and curvature. In the context of the Perona-Malik algorithm we consider the warm-up scenario of GfE where the two metrics associated to the GQRE are Euclidean of the metric induced by the matter-field and curvature is played by \({\boldsymbol{G}}_{\mu\nu}\) which is defined in Eq.(3 ) and schematically described in Figure 1. The action \(\mathcal{S}\) of GfE is given by \[\begin{align} \mathcal{S}=\frac{1}{2}\int_{\Omega} \sqrt{|-g|}\mathcal{L} d{\boldsymbol{r}}, \label{action} \end{align}\tag{6}\] where the Lagrangian \(\mathcal{L}\) is given by the GQRE between \(g\) and \({\boldsymbol{G}}\). Anticipating the results of the next paragraph, the GQRE Lagrangian \(\mathcal{L}\) in Euclidean space that encodes for the Perona-Malik algorithm is given by \[\begin{align} \mathcal{L}=-Tr\ln {\boldsymbol{G}}g^{-1}. \label{Lag00} \end{align}\tag{7}\] This definition of the GQRE implies that the GQRE that encodes for the Perona-Malik algorithm, can be also expressed in terms of the eigenvalues \(\lambda_n\) of \({\boldsymbol{G}}\) and the eigenvalues \(\lambda^{\prime}_n\) of \(g\) in a familiar form for the relative entropy. Indeed we have \[\begin{align} \mathcal{L}=-\sum_{n=1}^2\ln \lambda_n=\sum_{n=1}^2\lambda^{\prime}_n(\ln \lambda^{\prime}_n-\ln \lambda_n), \end{align}\] where we have used \(\lambda_n^{\prime}=1\) for \(n\in \{1,2\}\). Using the explicit expression of \(\lambda_n\) given by Eq.(5 ), we obtain \[\begin{align} \mathcal{L}=-\ln(1+\alpha|\nabla\phi|^2), \label{Lag} \end{align}\tag{8}\] which constitute the Euclidean version of the warm-up scenario of GfE proposed in Ref.[27].

First principles derivation of the GQRE in Euclidean space Eq. (7 ) and its connection with Araki entropy. In this paragraph we formulate the GQRE in Euclidean space. We discuss the quantum information theory foundations of the GQRE Lagragian \(\mathcal{L}\) (Eq.(7 )) and we relate it to the Araki entropy [48], [49]. Consistently with the GfE approach [27], here we consider a Dirac-Kähler [50], [51] interpretation of the gradient in which \(\nabla_{\mu}\phi\) describes the component of a one-form. Therefore, in order to provide the theoretical foundations for our definition of the GQRE, we will consider Hilbert spaces formed by the direct sum of a zero-form and a one-form. The generic vector \(\ket{\Psi}\) of the considered Hilbert space \(\mathcal{H}\) is given by \[\begin{align} \ket{\Psi}={{\phi}}\oplus {\omega}_{\mu} dx^{\mu},\label{phi1} \end{align}\tag{9}\] Thus both the image (encoded by \(\phi\oplus 0_{\mu}dx^{\mu}\)) and the gradient of the image encoded by \(0\oplus\nabla_{\mu}\phi dx^{\mu}\) can be interpreted as elements of this Hilbert space. The scalar product between \(\ket{\Phi}\) and another generic vector \(\ket{\Phi}\) given by \(\ket{\Phi}=\hat{\phi}\oplus \hat{\omega}_{\mu} dx^{\mu}\) is defined as \[\begin{align} \left\langle{\langle \Psi,\Phi\rangle}\right\rangle=\int \sqrt{-|g|} \Big(\bar{\phi}\hat{\phi}+\bar{\omega}_{\mu}\hat{\omega}^{\mu}\Big) d{\boldsymbol{r}}, \end{align}\] where \(\hat{\omega}^{\mu}=g^{\mu\rho}\hat{\omega}_{\rho}\) . Thus the metric tensor \(\tilde{g}^{-1}\) associated to this scalar product is given by \[\begin{align} \tilde{g}^{-1}&=&1\oplus g^{\mu\nu}dx_{\mu}\otimes dx_{\nu}. \end{align}\]

All vectors \(\ket{\Phi}\) in the Hilbert space \(\mathcal{H}\) must satisfy \[\begin{align} \left\langle{\bra{\Phi}\Phi\rangle}\right\rangle<\infty. \label{hilb} \end{align}\tag{10}\]

Starting from the induced metric \({\boldsymbol{G}}\) we can construct the topological induced metric \(\tilde{\boldsymbol{G}}\) given by \[\begin{align} \tilde{\boldsymbol{G}}&=&1\oplus {G}_{\mu\nu}dx^{\mu}\otimes dx^{\nu}. \end{align}\] This topological induced metric can be interpreted as a quantum operator \(\tilde{\boldsymbol{G}}:\mathcal{H}\to \mathcal{H}\) where \(\tilde{\boldsymbol{G}}\cdot \ket{\Phi}\in \mathcal{H}\), \[\begin{align} \tilde{\boldsymbol{G}}\cdot \ket{\Phi}&=&\phi\oplus {G}_{\mu\nu}\omega^{\nu}dx^{\mu}. \end{align}\]

The metric \(\tilde{g}\) of the manifold \(\Omega\) can be used to define a dual Hilbert space \(\mathcal{H}^{\star}\). To this end we define the dual of \(\ket{\Psi}\) as \(\ket{\Phi^{\star}}\) and the dual of \(\ket{\Phi}\) as \(\ket{\Phi^{\star}}\) given by \[\begin{align} \ket{\Psi^{\star}}={{\phi}}\oplus {\omega}^{\mu} dx_{\mu},\quad \ket{\Phi^{\star}}=\hat{\phi}\oplus \hat{\omega}^{\mu} dx_{\mu}, \end{align}\] where \(\omega^{\mu}=g^{\mu\rho}\omega_{\rho},\hat{\omega}^{\mu}=g^{\mu\rho}\hat{\omega}_{\rho}\).

The scalar product \(\left\langle{\left\langle{ \Psi^{\star},\Phi^{\star}}\right\rangle}\right\rangle_{\star}\) is mediated by \(\tilde{g}\) given by \[\begin{align} \tilde{g}=1\oplus g_{\mu\nu}dx^{\mu}\otimes dx^{\nu} \end{align}\] and satisfies \[\begin{align} \left\langle{\langle{\Psi},\Phi\rangle}\right\rangle=\left\langle{\langle \Psi^{\star},\Phi^{\star}\rangle}\right\rangle_{\star}. \label{Hdual} \end{align}\tag{11}\]

The dual operator \(\tilde{\boldsymbol{G}}^{\star}:\mathcal{H}^{\star}\to\mathcal{H}^{\star}\) of \(\tilde{\boldsymbol{G}}\) is given by \[\begin{align} \tilde{\boldsymbol{G}}^{\star}&=&1\oplus {[G^{\star}]}^{\mu\nu}dx_{\mu}\otimes dx_{\nu}. \end{align}\] which must satisfy \[\begin{align} \left\langle{\left\langle{{\Psi},\tilde{\boldsymbol{G}}\cdot\Phi}\right\rangle}\right\rangle=\left\langle{\left\langle{ \tilde{\boldsymbol{G}}^{\star}\cdot \Psi^{\star}, \Phi^{\star}}\right\rangle}\right\rangle_{\star}. \label{HdualG} \end{align}\tag{12}\] for any arbitrary choices of \(\ket{\Psi}\) and \(\ket{\Phi}\). Here the action of \(\tilde{\boldsymbol{G}}^{\star}\) on the generic dual vector \(\bra{\Phi^{\star}}\) obeys \[\begin{align} \tilde{\boldsymbol{G}}^{\star}\cdot \ket{\Phi^{\star}}&=&\phi\oplus {[G_{(1)}^{\star}]}^{\mu\nu}\omega_{\nu}dx_{\mu}. \end{align}\] Thus, using the symmetry of \(G_{\mu\nu}\) we conclude that \(\tilde{\boldsymbol{G}}^{\star}\) is related to \(\tilde{\boldsymbol{G}}\) by \[\begin{align} \tilde{\boldsymbol{G}}^{\star}=\tilde{g}^{-1}\tilde{\boldsymbol{G}}\tilde{g}^{-1}. \end{align}\] indicating that \[\begin{align} {[G^{\star}]}^{\mu\nu}&=&g^{\mu\rho}{G}_{\rho\sigma}g^{\nu\sigma}. \end{align}\]

The topological metrics \(\tilde{\boldsymbol{G}}\) and \(\tilde{\boldsymbol{G}}^{\star}\) that we consider in this work are respectively elements of the algebras \(\textswab{U}\) and \(\textswab{U}^*\) that generalizes the \(C^*\) algebra [47]. The norm associated to the topological metric \(\tilde{\boldsymbol{G}}\in \textswab{U}\) is equal to the norm associated to its dual, i.e.. \(\|\tilde{\boldsymbol{G}}\|=\|\tilde{\boldsymbol{G}}^{\star}\|\) with \[\begin{align} \|\tilde{\boldsymbol{G}}\|=\|\tilde{\boldsymbol{G}}^{\star}\|=\int \sqrt{|-g|}Tr_F \Big(\tilde{\boldsymbol{G}}\tilde{\boldsymbol{G}}^{\star}\Big) d{\boldsymbol{r}}, \label{norm} \end{align}\tag{13}\] where \(Tr_F \Big(\tilde{\boldsymbol{G}}\tilde{\boldsymbol{G}}^{\star}\Big)\) is given by \[\begin{align} Tr_F \Big(\tilde{\boldsymbol{G}}\tilde{\boldsymbol{G}}^{\star}\Big)&=&1+{G}_{\mu\nu}{G}^{\nu\mu}. \end{align}\]

For a topological metric \(\tilde{\boldsymbol{G}}\) interpreted as a quantum operator in \(\textswab{U}\) we define the square root of the modular operator \(\boldsymbol{\Delta}^{1/2}_{\tilde{\boldsymbol{G}},g}:\mathcal{H}\to \mathcal{H}\) as \[\begin{align} \boldsymbol{\Delta}^{1/2}_{\tilde{\boldsymbol{G}},g}=\sqrt{\tilde{\boldsymbol{G}}\tilde{\boldsymbol{G}}^{\star}}=\tilde{\boldsymbol{G}}\tilde{g}^{-1}, \end{align}\] where the last identity is derived under the assumption that \(\tilde{\boldsymbol{G}}\) is positively definite, i.e. it has only positive eigenvalues. By this notation we indicate the square -root modular operator \(\boldsymbol{\Delta}_{\tilde{\boldsymbol{G}},g}^{1/2}\) acting on the topological field \(\ket{\Phi}\) as \[\begin{align} \boldsymbol{\Delta}^{1/2}_{\tilde{\boldsymbol{G}},g}\ket{\Phi}&=&\phi\oplus {[G_{(1)}]}_{\mu\rho}g^{\rho\nu}\omega_{\nu}dx^{\mu}. \end{align}\] Using the flattened trace (\(Tr_F\)) formalism introduced in Ref. [27] we obtain that the Lagrangian \(\mathcal{L}\) indicating the GQRE can be expressed as \[\begin{align} \mathcal{L}=-Tr_F\ln \boldsymbol{\Delta}^{1/2}_{\tilde{\boldsymbol{G}},g}=-Tr\ln {\boldsymbol{G}}{g}^{-1}, \end{align}\] where the trace \(Tr\) of the last expression indicates the usual trace of a matrix. Therefore the action \(\mathcal{S}\) defined in Eq.(6 ) and associated with the GQRE Lagrangian given by Eq.(7 ), extends the definition of Araki [48], [49] quantum relative entropy and provides geometrical definition of this entropy by treating metrics as quantum operators.

The Perona-Malik algorithm as the gradient flow of the GQRE The GfE action, associated with the Lagrangian given by Eq.(8 ), expressing the GQRE between the metric \(g_{\mu\nu}=\eta_{\mu\nu}\) of the 2D support of the image and the induced metric \({ G}_{\mu\nu}\) defined in (3 ), is given by \[\begin{align} \mathcal{S}=-\frac{1}{2}\int_\Omega d{\boldsymbol{r}} \ln(1+\alpha|\nabla \phi|^2). \end{align}\] Thus, the Perona-Malik algorithm can be easily shown to be the gradient flow of this action, which is given by \[\begin{align} \frac{d\phi({\boldsymbol{r}},t)}{dt}&=&-\frac{\delta \mathcal{S}}{\delta \phi({\boldsymbol{x}},t)}. \end{align}\] Indeed in this way we get the dynamical equations \[\begin{align} \frac{d\phi({\boldsymbol{r}},t)}{dt}&=&\alpha\nabla_{\mu}\rho(|\nabla\phi|^2)\nabla^{\mu}\phi({\boldsymbol{r}},t), \end{align}\] where \(\rho(| \nabla\phi|^2)\) is given by Eq.(2 ), i.e. \[\begin{align} \rho(|\nabla\phi|^2)=\frac{1}{1+\alpha|\nabla\phi|^2} \end{align}\] with initial condition \(\phi({\boldsymbol{r}},0)=\psi({\boldsymbol{r}}).\) Therefore, upon a rescaling of the time \(t\to t/\alpha\) we recover the Perona-Malik algorithm defined in Eq.(1 ). As anticipated above, the GfE action, given by the GQRE between the metric \(g\) and the metric induced by the image \({\boldsymbol{G}}\) provides the information theory principle that justifies the choice of \(\rho(|\nabla\phi|^2)\).

Conclusions. This work shows that the GfE approach that gives rise to modified gravity, provides the entropic quantum gravity foundations of the Perona-Malik algorithm for anisotropic diffusion. In particular, we have revealed that the Perona-Malik algorithm is the gradient flow of the GfE action whose Lagrangian is given by the GQRE between the flat \(2D\) metric \(g_{\mu\nu}=\eta_{\mu\nu}\) and the metric \(G_{\mu\nu}\) induced by the image. This implies that the Perona-Malik algorithm can be interpreted as the outcome of the tension between the metric of the flat \(2D\) support of the image and the metric induced by the image on the \(2D\) plane trying to minimize their GQRE. Interestingly, the minimization of the GfE action is compatible with heterogeneous and complex images with sharp contours. This result is in sharp contrast with the expectations arising from the classical maximum entropy principle and might be reflected also in the properties of the solutions of the GfE modified gravity equations. Therefore this result might indicate relevant consequences of adopting the GQRE as the Lagrangian for both quantum gravity and image processing algorithms.

From the point of view of artificial intelligence, our findings establish solid quantum information-theoretic grounds for anisotropic diffusion and the Perona-Malik algorithm, justifying the ad hoc choice for the functional expression of the metric adopted by Perona and Malik. From the point of view of entropic quantum gravity, these findings provide a rather immediate application of the GfE approach to machine learning opening new perspectives on the development of the next generation of AI diffusion algorithms. Moreover, in this application, the \(2D\) metric associated with the image is flat and Euclidean and remains unchanged during the learning of the image, while, going beyond the warm-up scenario, the GfE approach envisages that this metric is associated to the curvature, and can evolve in time. Therefore, future research could explore the full potential of GfE action to develop the next generation of AI algorithms.

Based on this considerations, our expectations are that the GfE approach could provide new perspectives on physics inspired artificial intelligence, unsupervised learning [20] and brain research. On one side, these results might lead to the formulation of a new generation of geometric diffusion algorithms. On the other side, they might inspire research at the interface between topological and geometrical learning and brain research [52][55]. which could potentially capture, among the other things, the main mechanisms beyond brain illusions such as the Kanizsa triangle [56]. Thus the GfE approach might turn to be a fertile framework for proposing the next generation of diffusion models fully based on the information encoded in the geometrical description of data.

In summary, this work demonstrates the common foundations of GfE and anisotropic diffusion for image processing and indicates that the full GfE action could potentially offer valuable insights for developing future generations of AI algorithms.

Acknowledgements↩︎

The author acknowledges interesting scientific discussions with Giovanna Citti and Alessandro Sarti and thanks them for pointing out Ref. [32]. This work was partially supported by a grant from the Simons Foundation. The author would like to thank the Isaac Newton Institute for Mathematical Sciences, Cambridge, for support and hospitality during the programme Hypergraphs: Theory and Applications, where work on this paper was undertaken. This work was supported by EPSRC grant EP/V521929/1.

References↩︎

[1]
Herbert Levine and Yuhai Tu. Machine learning meets physics: A two-way street. Proceedings of the National Academy of Sciences, 121(27):e2403580121, 2024.
[2]
Giuseppe Carleo, Ignacio Cirac, Kyle Cranmer, Laurent Daudet, Maria Schuld, Naftali Tishby, Leslie Vogt-Maranto, and Lenka Zdeborová. Machine learning and the physical sciences. Reviews of Modern Physics, 91(4):045002, 2019.
[3]
Jonathan Carifio, James Halverson, Dmitri Krioukov, and Brent D Nelson. Machine learning in the string landscape. Journal of High Energy Physics, 2017(9):1–36, 2017.
[4]
Yang-Hui He. Deep-learning the landscape. In Machine Learning: In Pure Mathematics and Theoretical Physics, pages 183–221. World Scientific, 2023.
[5]
Juan Carrasquilla and Roger G Melko. Machine learning phases of matter. Nature Physics, 13(5):431–434, 2017.
[6]
Tiago Mendes-Santos, Markus Schmitt, Adriano Angelone, Alex Rodriguez, Pascal Scholl, Hannah J Williams, Daniel Barredo, Thierry Lahaye, Antoine Browaeys, Markus Heyl, et al. Wave-function network description and kolmogorov complexity of quantum many-body systems. Physical Review X, 14(2):021029, 2024.
[7]
Ana P Millán, Hanlin Sun, Lorenzo Giambagli, Riccardo Muolo, Timoteo Carletti, Joaquı́n J Torres, Filippo Radicchi, Jürgen Kurths, and Ginestra Bianconi. Topology shapes dynamics of higher-order networks. Nature Physics, pages 1–9, 2025.
[8]
Runyue Wang, Yu Tian, Pietro Liò, and Ginestra Bianconi. Dirac-equation signal processing: Physics boosts topological machine learning. arXiv preprint arXiv:2412.05132, 2024.
[9]
Theodore Papamarkou, Tolga Birdal, Michael Bronstein, Gunnar Carlsson, Justin Curry, Yue Gao, Mustafa Hajij, Roland Kwitt, Pietro Lio, Paolo Di Lorenzo, et al. Position: Topological deep learning is the new frontier for relational learning. arXiv preprint arXiv:2402.08871, 2024.
[10]
Michael M Bronstein, Joan Bruna, Yann LeCun, Arthur Szlam, and Pierre Vandergheynst. Geometric deep learning: going beyond euclidean data. IEEE Signal Processing Magazine, 34(4):18–42, 2017.
[11]
Ginestra Bianconi. Higher-order networks: An introduction to simplicial complexes. Cambridge University Press, 2021.
[12]
Federico Battiston, Enrico Amico, Alain Barrat, Ginestra Bianconi, Guilherme Ferraz de Arruda, Benedetta Franceschiello, Iacopo Iacopini, Sonia Kéfi, Vito Latora, Yamir Moreno, et al. The physics of higher-order interactions in complex systems. Nature Physics, 17(10):1093–1098, 2021.
[13]
Giovanni Petri, Paul Expert, Federico Turkheimer, Robin Carhart-Harris, David Nutt, Peter J Hellyer, and Francesco Vaccarino. Homological scaffolds of brain functional networks. Journal of The Royal Society Interface, 11(101):20140873, 2014.
[14]
Michael W Reimann, Max Nolte, Martina Scolamiero, Katharine Turner, Rodrigo Perin, Giuseppe Chindemi, Paweł Dłotko, Ran Levi, Kathryn Hess, and Henry Markram. Cliques of neurons bound into cavities provide a missing link between structure and function. Frontiers in computational neuroscience, 11:266051, 2017.
[15]
Andrea Santoro, Federico Battiston, Giovanni Petri, and Enrico Amico. Higher-order organization of multivariate time series. Nature Physics, 19(2):221–229, 2023.
[16]
Joshua Faskowitz, Richard F Betzel, and Olaf Sporns. Edges in brain networks: Contributions to models of structure and function. Network Neuroscience, 6(1):1–28, 2022.
[17]
Dániel L Barabási, Ginestra Bianconi, Ed Bullmore, Mark Burgess, SueYeon Chung, Tina Eliassi-Rad, Dileep George, István A Kovács, Hernán Makse, Thomas E Nichols, et al. Neuroscience needs network science. Journal of Neuroscience, 43(34):5989–5995, 2023.
[18]
Francesco Di Giovanni, James Rowbottom, Benjamin P Chamberlain, Thomas Markovich, and Michael M Bronstein. Understanding convolution on graphs via energies. arXiv preprint arXiv:2206.10991, 2022.
[19]
Michael M Bronstein, Joan Bruna, Taco Cohen, and Petar Veličković. Geometric deep learning: Grids, groups, graphs, geodesics, and gauges. arXiv preprint arXiv:2104.13478, 2021.
[20]
Lars Schmarje, Monty Santarossa, Simon-Martin Schröder, and Reinhard Koch. A survey on semi-, self-and unsupervised learning for image classification. IEEE Access, 9:82146–82168, 2021.
[21]
Cristopher Moore and Stephan Mertens. The nature of computation. Oxford University Press, 2011.
[22]
Marc Mezard and Andrea Montanari. Information, physics, and computation. Oxford University Press, 2009.
[23]
Karl Friston. The free-energy principle: a unified brain theory? Nature reviews neuroscience, 11(2):127–138, 2010.
[24]
Naftali Tishby, Fernando C Pereira, and William Bialek. The information bottleneck method. arXiv preprint physics/0004057, 2000.
[25]
Ling Yang, Zhilong Zhang, Yang Song, Shenda Hong, Runsheng Xu, Yue Zhao, Wentao Zhang, Bin Cui, and Ming-Hsuan Yang. Diffusion models: A comprehensive survey of methods and applications. ACM Computing Surveys, 56(4):1–39, 2023.
[26]
Ben Chamberlain, James Rowbottom, Maria I Gorinova, Michael Bronstein, Stefan Webb, and Emanuele Rossi. Grand: Graph neural diffusion. In International conference on machine learning, pages 1407–1418. PMLR, 2021.
[27]
Ginestra Bianconi. Gravity from entropy. Physical Review D, 111(6):066001, 2025.
[28]
Gerard’t Hooft. The holographic principle. In Basics and Highlights in Fundamental Physics, pages 72–100. World Scientific, 2001.
[29]
Leonard Susskind. The world as a hologram. Journal of Mathematical Physics, 36(11):6377–6396, 1995.
[30]
Juan Maldacena. The illusion of gravity. Scientific American, 293(5):56–63, 2005.
[31]
Juan Maldacena. The large-n limit of superconformal field theories and supergravity. International journal of theoretical physics, 38(4):1113–1133, 1999.
[32]
Ron Kimmel, Nir Sochen, and Ravi Malladi. From high energy physics to low level vision. In Scale-Space Theory in Computer Vision: First International Conference, The Netherlands, July 2–4, 1997 Proceedings 1, pages 236–247. Springer, 1997.
[33]
Filippo Passerini and Simone Severini. The von neumann entropy of networks. arXiv preprint arXiv:0812.2597, 2008.
[34]
Kartik Anand and Ginestra Bianconi. Entropy measures for networks: Toward an information theory of complex topologies. Physical Review E—Statistical, Nonlinear, and Soft Matter Physics, 80(4):045102, 2009.
[35]
Kartik Anand, Ginestra Bianconi, and Simone Severini. Shannon and von Neumann entropy of random networks with heterogeneous expected degree. Physical Review E, 83(3):036109, 2011.
[36]
Manlio De Domenico and Jacob Biamonte. Spectral entropies as information-theoretic tools for complex network comparison. Physical Review X, 6(4):041062, 2016.
[37]
Pablo Villegas, Tommaso Gili, Guido Caldarelli, and Andrea Gabrielli. Laplacian renormalization group for heterogeneous networks. Nature Physics, 19(3):445–450, 2023.
[38]
Manlio De Domenico, Vincenzo Nicosia, Alexandre Arenas, and Vito Latora. Structural reducibility of multilayer networks. Nature Communications, 6(1):1–9, 2015.
[39]
Flaviano Morone, Ian Leifer, and Hernán A Makse. Fibration symmetries uncover the building blocks of biological networks. Proceedings of the National Academy of Sciences, 117(15):8306–8314, 2020.
[40]
Vlatko Vedral. The role of relative entropy in quantum information theory. Reviews of Modern Physics, 74(1):197, 2002.
[41]
Ginestra Bianconi. The quantum relative entropy of the schwarzschild black hole and the area law. Entropy, 27(3):266, 2025.
[42]
Pietro Perona and Jitendra Malik. Scale-space and edge detection using anisotropic diffusion. IEEE Transactions on pattern analysis and machine intelligence, 12(7):629–639, 1990.
[43]
Satyanad Kichenassamy. The Perona–Malik paradox. SIAM Journal on Applied Mathematics, 57(5):1328–1342, 1997.
[44]
Patrick Guidotti. Some anisotropic diffusions. Ulmer Seminare, 14:215–221, 2009.
[45]
Antoni Buades, Bartomeu Coll, and Jean-Michel Morel. A review of image denoising algorithms, with a new one. Multiscale modeling & simulation, 4(2):490–530, 2005.
[46]
Edward Witten. edal for Exceptional Achievement in Research: Invited article on entanglement properties of quantum field theory. Reviews of Modern Physics, 90(4):045003, 2018.
[47]
Huzihiro Araki. Mathematical theory of quantum fields. Oxford University Press, USA, 1999.
[48]
Huzihiro Araki. Relative entropy of states of von Neumann algebras. Publications of the Research Institute for Mathematical Sciences, 11(3):809–833, 1975.
[49]
Masanori Ohya and Dénes Petz. Quantum entropy and its use. Springer Science & Business Media, 2004.
[50]
SI Kruglov. irac–Kähler equation. International Journal of Theoretical Physics, 41(4):653–687, 2002.
[51]
Peter Becher and Hans Joos. The Dirac-Kähler equation and fermions on the lattice. Zeitschrift für Physik C Particles and Fields, 15:343–365, 1982.
[52]
Fernando AN Santos, Ernesto P Raposo, Maurı́cio D Coutinho-Filho, Mauro Copelli, Cornelis J Stam, and Linda Douw. Topological phase transitions in functional brain networks. Physical Review E, 100(3):032414, 2019.
[53]
Yuansheng Zhou, Brian H Smith, and Tatyana O Sharpee. Hyperbolic geometry of the olfactory space. Science advances, 4(8):eaaq1458, 2018.
[54]
Giovanna Citti and Alessandro Sarti. A gauge field model of modal completion. Journal of Mathematical Imaging and Vision, 52:267–284, 2015.
[55]
Giovanna Citti and Alessandro Sarti. Neuromathematics of vision, volume 32. Springer, 2014.
[56]
Gaetano Kanizsa. Subjective contours. Scientific American, 234(4):48–53, 1976.