We consider the unique continuation (data assimilation) problem for the Helmholtz equation and study its numerical approximation based on physics-informed neural networks (PINNs). Exploiting the conditional stability of the problem, we first give a bound on the generalization error of PINNs. We then present numerical experiments in 2d for different frequencies and for geometric configurations with different stability bounds for the continuation problem. The results show that vanilla PINNs provide good approximations even for noisy data in configurations with robust stability (both low and moderate frequencies), but may struggle otherwise. This indicates that more sophisticated techniques are needed to obtain PINNs that are frequency-robust for inverse problems subject to the Helmholtz equation.
Authors
Mihai Nechita
Tiberiu Popoviciu Institute of Numerical Analysis, Romanian Academy, Cluj-Napoca, Romania
Department of Mathematics, Babeṣ-Bolyai University, Cluj-Napoca, Romania
P. Escapil-Inchauspé and G. A. Ruz, Hyper-parameter tuning of physics-informed neural networks: Application to Helmholtz problems, preprint arXiv:2205.06704, 2022
T. G Grossmann, U. J. Komorowska, J. Latz, and C.-B. Schönlieb, Can physics-informed neural networks beat the finite element method?, preprint arXiv:2302.04107, 2023
M. Raissi, P. Perdikaris, and G. E. Karniadakis, Physics-informed neural networks: A deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations, Journal of Computational physics, 378 (2019), pp. 686-707,https://doi.org/10.1016/j.jcp.2018.10.045DOI:https://doi.org/10.1016/j.jcp.2018.10.045
2023_1_Nechita
SOLVING ILL-POSED HELMHOLTZ PROBLEMS WITH PHYSICS-INFORMED NEURAL NETWORKS*
MIHAI NECHITA ^(†‡){ }^{\dagger \ddagger}
Abstract
We consider the unique continuation (data assimilation) problem for the Helmholtz equation and study its numerical approximation based on physicsinformed neural networks (PINNs). Exploiting the conditional stability of the problem, we first give a bound on the generalization error of PINNs. We then present numerical experiments in 2d for different frequencies and for geometric configurations with different stability bounds for the continuation problem. The results show that vanilla PINNs provide good approximations even for noisy data in configurations with robust stability (both low and moderate frequencies), but may struggle otherwise. This indicates that more sophisticated techniques are needed to obtain PINNs that are frequency-robust for inverse problems subject to the Helmholtz equation.
Physics-informed neural networks [16] have recently emerged as an effective way of approximating ill-posed/inverse problems for partial differential equations (PDEs), which are challenging to solve numerically due their inherent instability [11]. One of their main advantages is the versatility with which they can solve both forward and inverse problems. As reported in [9] for well-posed boundary value problems, PINNs are not able, however, to outperform finite elements methods for low dimensional forward problems ( 2 d and 3 d ).
In this paper we explore PINNs for an important class of inverse problems, namely the unique continuation (data assimilation) problem in which measurements are available in a subset of the domain and no boundary conditions are given. The goal is to find a solution that extends the data subject to the PDE
considered. Numerical methods for such problems typically consider regularization at the continuous level (Tikhonov or quasi-reversibility). More recently, other kinds of methods have been proposed that make use of regularization at the discrete level in the framework of stabilized finite element methods (FEM). In this latter approach, conditional stability estimates can be used to prove error bounds and convergence in terms of the degree of ill-posedness and the approximation order, see e.g. [3, 5, 15]. Conditional stability estimates have also been employed in [14] to obtain bounds for the generalization error of PINNs solving unique continuation problems for Poisson, Stokes, heat and wave equations.
We are interested in the ill-posed unique continuation problem for the Helmholtz equation, whose stability properties (in terms of the frequency) depend on the geometry of the data set relative to the geometry of the target set where the solution is reconstructed [6]. We consider PINNs for approximating this problem and we first give a bound in Section 3 on the generalization error that takes into account the frequency dependence. We then explore numerically in Section 4 how different frequencies and geometric configuration impact the effectiveness of PINNs. We see that PINNs give good approximations when the frequency is low and the target set is inside the convex hull of the data set (robust stability in terms of frequency as discussed in Section 2.1). This includes results where data perturbations are present. When the target set is outside of the convex hull of the data set (stability highly sensitive to frequency) or when the frequency is large we observe poor results with vanilla PINNs. This indicates that more sophisticated techniques are needed to obtain PINNs that are robust in frequency for inverse problems subject to the Helmholtz equation.
2. UNIQUE CONTINUATION FOR THE HELMHOLTZ EQUATION
Let Omega subR^(2)\Omega \subset \mathbb{R}^{2} be a bounded domain (open and connected), and let omega sub Omega\omega \subset \Omega be an open and connected subset. We consider the ill-posed unique continuation problem for the Helmholtz equation: find u inH^(1)(Omega)u \in H^{1}(\Omega) such that
{[(1)-Delta u-k^(2)u=f" in "Omega","],[u=g" in "omega","]:}\left\{\begin{align*}
-\Delta u-k^{2} u=f & \text { in } \Omega, \tag{1}\\
u=g & \text { in } \omega,
\end{align*}\right.
with wave number k > 0k>0, source term f inL^(2)(Omega)f \in L^{2}(\Omega) and datum g inH^(1)(omega)g \in H^{1}(\omega). In other words, partial measurements are given in a subset omega sub Omega\omega \subset \Omega of the domain for the solution of the Helmholtz equation, while no boundary conditions are prescribed. If problem (1) has a solution, then its uniqueness is guaranteed by the unique continuation principle for elliptic operators. We will assume that the function gg is the restriction on omega\omega of a solution to the Helmholtz equation with source term ff.
Problem (1) is ill-posed in the sense of Hadamard: there is no uniform stability with respect to the data ff and gg, see e.g. [1] or [11]. A small
perturbation in data can lead to a large change in the solution. Nonetheless, assuming an additional a priori bound, the solution can be bounded by the data in the following way.
2.1. Conditional stability. We denote by B sub OmegaB \subset \Omega a target set containing omega\omega such that B\\ bar(omega)sub OmegaB \backslash \bar{\omega} \subset \Omega, i.e. B\\omegaB \backslash \omega does not touch the boundary of Omega\Omega. A standard result for quantitative unique continuation of elliptic operators, see e.g. [1], states that there exist constants C_(st)(k) > 0C_{s t}(k)>0 and alpha in(0,1)\alpha \in(0,1) such that the following Hölder stability estimate holds
for any u inH^(1)(Omega)u \in H^{1}(\Omega) satisfying (1), where the stability constant C_(st)(k)C_{s t}(k) depends on the frequency kk. The exponent alpha in(0,1)\alpha \in(0,1) encodes the degree of ill-posedness for the continuation problem: as alpha < 1\alpha<1 decreases the Hölder stability deteriorates, while alpha=1\alpha=1 would give a well-posed problem with Lipschitz stability. Both C_(st)C_{s t} and alpha\alpha depend on the geometric configuration in a nontrivial way, and we give details below regarding C_(st)C_{s t}.
Fig. 1. Frequency dependence for the stability constant. Data set omega\omega (dark grey) and target set BB (light grey). Omega\Omega is the whole square.
An important aspect for the stability of this ill-posed Helmholtz problem is the dependence of the stability constant C_(st)(k)C_{s t}(k) on the frequency kk. For example, when there is a straight line that intersects BB but not bar(omega)\bar{\omega}, as in Fig. 1a, it was proven in [6, Example 4] that for any N inN,C_(st)(k) <= k^(N)N \in \mathbb{N}, C_{s t}(k) \leq k^{N} cannot hold uniformly in kk; this means that the stability constant in (2) grows superpolynomially in the frequency. Also, for three-ball inequalities (where omega,B,Omega\omega, B, \Omega are concentric balls) it was recently shown in [2] that in the maximum norm C_(st)(k)C_{s t}(k) grows exponentially in kk and this dependence is optimal.
Bounds that have a different behavior with respect to the frequency can be obtained under a convexity condition of the target domain BB relative to the data set omega\omega, essentially that BB is included in the convex hull of omega\omega, as for example in Fig. 1b. Such a condition was first considered in [10], where it was
shown that the stability of the solution in the L^(2)L^{2}-norm can actually improve in a certain sense as the frequency increases. In this vein, similar results that make use of a convexity condition were proven in [6] for a particular geometric setting prototypical for continuation inside the convex hull of omega\omega. It was shown in [6, Corollary 2] that there exist constants C > 0C>0 and alpha in(0,1)\alpha \in(0,1) such that
for any u inH^(1)(Omega)u \in H^{1}(\Omega) satisfying (1). Moreover, the norms in which data is measured can be weaken [6, Corollary 3 and Lemma 2]: there exist constants C > 0C>0 and alpha in(0,1]\alpha \in(0,1] such that
{:[(4)||grad u||_(L^(2)(B))+k||u||_(L^(2)(B)) <= Ck(||f||_(H^(-1)(Omega))+||g||_(L^(2)(omega)))^(alpha)(||f||_(H^(-1)(Omega))+||u||_(L^(2)(Omega)))^(1-alpha)],[ <= Ck(||f||_(L^(2)(Omega))+||g||_(L^(2)(omega)))^(alpha)(||f||_(L^(2)(Omega))+||u||_(L^(2)(Omega)))^(1-alpha)","]:}\begin{align*}
\|\nabla u\|_{L^{2}(B)}+k\|u\|_{L^{2}(B)} & \leq C k\left(\|f\|_{H^{-1}(\Omega)}+\|g\|_{L^{2}(\omega)}\right)^{\alpha}\left(\|f\|_{H^{-1}(\Omega)}+\|u\|_{L^{2}(\Omega)}\right)^{1-\alpha} \tag{4}\\
& \leq C k\left(\|f\|_{L^{2}(\Omega)}+\|g\|_{L^{2}(\omega)}\right)^{\alpha}\left(\|f\|_{L^{2}(\Omega)}+\|u\|_{L^{2}(\Omega)}\right)^{1-\alpha},
\end{align*}
for any u inH^(1)(Omega)u \in H^{1}(\Omega) satisfying (1). Note that the bound is robust in the L^(2)L^{2} norm, while the frequency dependence is linear for the H^(1)H^{1}-seminorm. Apart from the good dependence on the frequency, this kind of estimate is particularly suitable for numerical analysis since it can be directly applied to the error equation.
Remark 1. If the target set is the whole domain Omega\Omega, then a global stability estimate holds with the modulus of continuity being logarithmic |log(*)|^(-alpha)|\log (\cdot)|^{-\alpha} instead of Hölder-type |*|^(alpha)|\cdot|^{\alpha}. ◻\square
3. PHYSICS-INFORMED NEURAL NETWORKS (PINNS)
Let us first recall the setup of a feed-forward neural network, which is based on successively applying linear and nonlinear transformations to the inputs. Let N^(L)(x):R^(d_("in "))rarrR^(d_("out "))\mathcal{N}^{L}(\mathbf{x}): \mathbb{R}^{d_{\text {in }}} \rightarrow \mathbb{R}^{d_{\text {out }}} be an LL-layer neural network with ( L-1L-1 ) hidden layers and N_(ℓ)N_{\ell} neurons in the ℓ\ell-th layer ( N_(0)=d_("in "),N_(L)=d_("out ")N_{0}=d_{\text {in }}, N_{L}=d_{\text {out }} ). Let us denote the weight matrix and bias vector in the ℓ\ell-th layer by W^(ℓ)inR^(N_(ℓ)xxN_(ℓ-1))\boldsymbol{W}^{\ell} \in \mathbb{R}^{N_{\ell} \times N_{\ell-1}} and b^(ℓ)inR^(N_(ℓ))\mathbf{b}^{\ell} \in \mathbb{R}^{N_{\ell}}, respectively. For a nonlinear activation function sigma\sigma applied element wise, the feed-forward neural network is given by:
depends on the tuning parameter theta\theta. Training the network means using an optimization algorithm to find the parameters theta in Theta\theta \in \Theta that minimize a certain loss function L_(theta)\mathcal{L}_{\theta}.
Let us now briefly describe physics-informed neural networks (PINNs) as introduced in [16], based on feed-forward neural networks. The idea is to consider a loss function with two components: the residual of the differential equation and the data residual. For typical boundary value problems, the data residual contains boundary conditions, initial conditions, and any other available measurements of the solution. In our case, since no boundary conditions are given, the data residual will only involve the interior partial measurements.
To solve the ill-posed unique continuation problem (1) with PINNs we consider the residual of the Helmholtz equation
R^(@)_(theta):=-Deltau_(theta)-k^(2)u_(theta)-f" in "Omega\stackrel{\circ}{\mathcal{R}}_{\theta}:=-\Delta u_{\theta}-k^{2} u_{\theta}-f \text { in } \Omega
and the data residual
R^(@)_(theta,d):=u_(theta)-g" in "omega,\stackrel{\circ}{\mathcal{R}}_{\theta, d}:=u_{\theta}-g \text { in } \omega,
and aim to minimize the two residuals over the admissible set of tuning parameters Theta\Theta by considering the loss function
To approximate the integrals in the setting of a neural network, we introduce the following collocation points: T_(Omega):={x_(i)^(Omega)}_(i=1)^(|T_(Omega)|)\mathcal{T}_{\Omega}:=\left\{\mathbf{x}_{i}^{\Omega}\right\}_{i=1}^{\left|\mathcal{T}_{\Omega}\right|} for the whole domain Omega\Omega and T_(omega):={x_(i)^(omega)}_(i=1)^(|T_(omega)|)\mathcal{T}_{\omega}:=\left\{\mathbf{x}_{i}^{\omega}\right\}_{i=1}^{\left|\mathcal{T}_{\omega}\right|} for the data region omega sub Omega\omega \subset \Omega in which observations of the solution are given as {g_(i)(x^(omega))}_(i=1)^(|T_(omega)|)\left\{g_{i}\left(\mathbf{x}^{\omega}\right)\right\}_{i=1}^{\left|\mathcal{T}_{\omega}\right|}. The loss function for the PINN is hence defined as
{:(8)" find "theta^(**)in Theta" such that "theta^(**)=arg min_(theta in Theta)L_(theta)". ":}\begin{equation*}
\text { find } \theta^{*} \in \Theta \text { such that } \theta^{*}=\arg \min _{\theta \in \Theta} \mathcal{L}_{\theta} \text {. } \tag{8}
\end{equation*}
We will denote by
u^(**)=u_(theta^(**))u^{*}=u_{\theta^{*}}
the PINN solution obtained by this algorithm, see Fig. 2 for a sketch.
As the loss function is highly non-linear and non-convex, we will use the standard approach of minimizing by gradient-based methods such as ADAM [12] or L-BFGS [7], more details are given in Section 4.
Fig. 2. Representation of a PINN for solving (1), adapted from [8].
3.1. Estimating the generalization error. Such PINNs for unique continuation problems have been considered in [14] where the authors prove for the first time estimates for the generalization error by employing conditional stability estimates; details of the analysis and numerical examples were given for the Poisson, Stokes, heat and wave equations. In brief, one uses conditional stability estimates to bound the error in a target domain (generalization error) by the residuals, which are then bounded in terms of the training error and the number of training samples. We now adapt the analysis in [14, Theorem 2.4] to the case of the Helmholtz equation by using the stability results discussed in Section 2.1.
Theorem 2. Let f inC^(n-2)(Omega)f \in C^{n-2}(\Omega) and g inC^(n)(omega)g \in C^{n}(\omega), with continuous extensions of the functions and derivatives up to the boundaries, with n >= 2n \geq 2. Let u inH^(1)(Omega)u \in H^{1}(\Omega) be the solution of the unique continuation problem (1). Let u^(**)=u_(theta^(**))inC^(n)(Omega)u^{*}=u_{\theta^{*}} \in C^{n}(\Omega) be a PINN solution generated by training (8). Consider a target set B sub OmegaB \subset \Omega containing omega\omega satisfying the convexity condition in [6, Corollary 2]. Then there exist constants C_(st)(k) > 0C_{s t}(k)>0 and alpha in(0,1)\alpha \in(0,1) such that the generalization error
with constants C_(q)C_{q} and C_(qd)C_{q d} given by the quadrature bounds, and some tau,tau_(d) > 0\tau, \tau_{d}>0.
Proof. Consider the error hat(u)=u^(**)-u inH^(1)(Omega)\hat{u}=u^{*}-u \in H^{1}(\Omega) which satisfies
{:[-Delta hat(u)-k^(2) hat(u)=R^(@)_(theta^(**))","" in "Omega","],[ hat(u)=R^(@)_(theta^(**),d)","" in "omega","]:}\begin{aligned}
-\Delta \hat{u}-k^{2} \hat{u} & =\stackrel{\circ}{\mathcal{R}}_{\theta^{*}}, & & \text { in } \Omega, \\
\hat{u} & =\stackrel{\circ}{\mathcal{R}}_{\theta^{*}, d}, & & \text { in } \omega,
\end{aligned}
in a weak sense. Applying the conditional stability estimate (4) we obtain E_(G)(B) <=\mathcal{E}_{G}(B) \leq <= C_(st)(k)(||R^(@)_(theta^(**))||_(L^(2)(Omega))+||R^(@)_(theta^(**),d)||_(L^(2)(omega)))^(alpha)(||R^(@)_(theta^(**))||_(L^(2)(Omega))+||( hat(u))||_(L^(2)(Omega)))^(1-alpha)\leq C_{s t}(k)\left(\left\|\stackrel{\circ}{\mathcal{R}}_{\theta^{*}}\right\|_{L^{2}(\Omega)}+\left\|\stackrel{\circ}{\mathcal{R}}_{\theta^{*}, d}\right\|_{L^{2}(\omega)}\right)^{\alpha}\left(\left\|\stackrel{\circ}{\mathcal{R}}_{\theta^{*}}\right\|_{L^{2}(\Omega)}+\|\hat{u}\|_{L^{2}(\Omega)}\right)^{1-\alpha} <= C_(st)(k)(||"Å"_(theta^(**))||_(L^(2)(Omega))+||R^(@)_(theta^(**),d)||_(L^(2)(omega)))^(alpha)(||R^(@)_(theta^(**))||_(L^(2)(Omega))+||u||_(L^(2)(Omega))+||u^(**)||_(L^(2)(Omega)))^(1-alpha)\leq C_{s t}(k)\left(\left\|\AA_{\theta^{*}}\right\|_{L^{2}(\Omega)}+\left\|\stackrel{\circ}{\mathcal{R}}_{\theta^{*}, d}\right\|_{L^{2}(\omega)}\right)^{\alpha}\left(\left\|\stackrel{\circ}{\mathcal{R}}_{\theta^{*}}\right\|_{L^{2}(\Omega)}+\|u\|_{L^{2}(\Omega)}+\left\|u^{*}\right\|_{L^{2}(\Omega)}\right)^{1-\alpha}Å.
The conclusion follows by assuming quadrature rules with approximation bounds as in [14, Eqs (2.9) and (2.11)], where for some tau,tau_(d) > 0\tau, \tau_{d}>0 one has that
with constants C_(q)=C_(q)(||R^(@)_(theta^(**))||_(C^(n-2)(Omega)))C_{q}=C_{q}\left(\left\|\stackrel{\circ}{\mathcal{R}}_{\theta^{*}}\right\|_{C^{n-2}(\Omega)}\right) and C_(qd)=C_(qd)(||R^(@)_(theta^(**),d)||_(C^(n)(omega)))C_{q d}=C_{q d}\left(\left\|\stackrel{\circ}{\mathcal{R}}_{\theta^{*}, d}\right\|_{C^{n}(\omega)}\right). ◻\square
Remark 3. As discussed in Section 2.1, if the target domain BB is outside the convex hull of the measurement domain omega\omega, the constant C_(st)(k)C_{s t}(k) in the conditional stability estimate might depend exponentially on the wave number kk. In that case, the same behaviour will appear in the bound of the generalization error in Theorem 2, which indicates that obtaining good approximations for this ill-posed problem with PINNs can be very challenging for high frequencies. ◻\square
REMARK 4. If the target set is the whole domain Omega\Omega, then one can obtain a global bound on the generalization error by using a global logarithmic stability estimate with an implicit dependence on the wave number. ◻\square
4. NUMERICAL EXPERIMENTS
We present numerical experiments for the Helmholtz unique continuation problem (1) solved with physics-informed neural networks (PINNs) described in Section 3 and implemented using the open-source library DeepXDE [13]. The test case we will focus on has been considered as a benchmark test for this problem solved with primal-dual stabilized finite element methods: with a conforming discretization in [6] and with a high-order hybridized discontinuous
Galerkin method in [4]. It represents the Helmholtz version of the classical Hadamard example for ill-posed elliptic equations.
Let n inNn \in \mathbb{N} and consider the Cauchy problem
{[(13)Delta u+k^(2)u=0" in "Omega:=(0","1)xx(0","1)","],[u(x","0)=0" for "x in[0","1]","],[u_(y)(x","0)=sin(nx)" for "x in[0","1]","]:}\left\{\begin{align*}
\Delta u+k^{2} u & =0 & & \text { in } \Omega:=(0,1) \times(0,1), \tag{13}\\
u(x, 0) & =0 & & \text { for } x \in[0,1], \\
u_{y}(x, 0) & =\sin (n x) & & \text { for } x \in[0,1],
\end{align*}\right.
For such Hadamard-type solutions, we consider the interior datum g=u|_(omega)g=\left.u\right|_{\omega} and study two geometric configurations of omega\omega and BB : one with frequency-robust stability bounds and one exponentially sensitive to the frequency, as discussed in Section 2.1, namely
sketched in Fig. 1a.
To assess the effect of increasing the frequency, we will take exact solutions (14) with n=5,k=1n=5, k=1 and n=7,k=5n=7, k=5, both having sqrt(n^(2)-k^(2))=sqrt24\sqrt{n^{2}-k^{2}}=\sqrt{24}.
Hyper-parameters. The PINNs described in Section 3 need to be configured with the following hyper-parameters: number of hidden layers L-1L-1 (depth), number of neurons in each hidden layer ℓ\ell (width), learning rate lambda\lambda, activation function sigma\sigma. In order to find good configurations for these, we draw on previous numerical experiments for forward and inverse problems presented in the DeepXDE tutorials [13] and the numerical experiments for unique continuation subject to the Poisson equation in [14]. We also validate the choices by comparing them with the results given by the hyper-parameter optimization in [8], which uses Gaussian processes-based Bayesian optimization. For this we consider the search space L-1in[4,50],N_(ℓ)in[20,150],lambda in[10^(-4),10^(-2)],sigma in{sin,tanh}L-1 \in[4,50], N_{\ell} \in[20,150], \lambda \in\left[10^{-4}, 10^{-2}\right], \sigma \in \{\sin , \tanh \}. In these ways we obtain some optimal or near-optimal configurations to which the results presented below correspond. From different numerical experiments, it turns out that a good choice for the activation function is sigma=sin\sigma=\sin and for the learning rate lambda=10^(-3)\lambda=10^{-3}, which will be fixed from this point onwards.
Training. We sample N=N_(Omega)=N_(omega)N=N_{\Omega}=N_{\omega} points on Cartesian grids from the PDE domain Omega\Omega and the measurement domain omega\omega. We train the model for 50000 iterations with the ADAM optimizer [12] and then we train again with L-BFGS
[7], following [13]. Since PINNs approximations are obtained by non-convex minimization which in general does not have a unique solution, the method might converge to different solutions depending on the network's initial values. We use the standard strategy of training the PINNs with different random starting values for the optimizer and average the errors over 30 retrainings. We note that the smallest training error is considerably better.
Fig. 3. PINN solution and exact solution (14) with n=7,k=5n=7, k=5.
Results. We first show in Fig. 3 the exact solution uu for the Hadamard-type function (14) with n=7,k=5n=7, k=5, and the PINN solution u^(**)u^{*} when considering a network generated with N=400N=400 training points, L-1=4L-1=4 hidden layers and N_(ℓ)=24N_{\ell}=24 neurons in each layer. As discussed above, the activation function is sigma=sin\sigma=\sin and the learning rate lambda=10^(-3)\lambda=10^{-3}.
Table 1. Geometry with good stability (15). Relative percentage generalization errors in B,n=5B, n=5 and k=1k=1.
We study the efficienty of PINNs by considering the L^(2)L^{2} and H^(1)H^{1} relative percentage errors for the number of training points N=20^(2),40^(2),80^(2)N=20^{2}, 40^{2}, 80^{2} in the geometric configuration (15) with robust bounds in Theorem 2. Table 1 shows the errors in the target domain BB, while Table 2 shows the global errors. In both tables we observe similar behaviours for the error. Moreover, we notice that even for very few training points the generalization errors are small (around 3%3 \% ). Note that the reported errors correspond to averages over 30
retrainings. The generalization errors corresponding to the network with the smallest training error are typically smaller by a factor of 3 to 5 . The training error in these examples is in between 10^(-3)10^{-3} and 10^(-2)10^{-2}. The training of such a network is very fast, on a single Quadro GPU it takes about 1-2 minutes.
Table 3. Geometry with good stability (15). Relative percentage generalization errors in Omega,n=5\Omega, n=5 and k=1k=1. Perturbed data.
We also test the performance of the method when the measurements are perturbed with random values sampled from the normal distribution with zero mean and standard deviation 0.01 (representing approximately 2.5%2.5 \% of the average of the exact solution (14) with n=5,k=1n=5, k=1 considered in that example). The results given in Table 3 show that the numerical approximation is robust to such perturbations in data.
In Table 4 we keep the same (convex) geometric configuration but increase the frequency to k=5k=5, with n=7n=7 such that sqrt(n^(2)-k^(2))=sqrt24\sqrt{n^{2}-k^{2}}=\sqrt{24} as before. We observe that the approximation improves as the frequency increases: the errors decrease compared to Table 2. This is probably an artefact of the particular conditions of the experiment, but we note that such a surprising phenomenon for inverse Helmholtz problems has been previously noticed and theoretically discussed in [10]. However, when increasing the wave number k > 10k>10 we report that the PINNs no longer provide a good approximation to the solution.
Table 4. Geometry with good stability (15). Relative percentage generalization errors in Omega\Omega, n=7n=7 and k=5k=5.
We now consider continuation outside the convex hull of the data set in (16). We observe in Table 5 that even though the wave number is small k=1k=1, the PINNs fail to find a good approximation of the solution. This could be caused by a bad search space for hyper-parameter optimization or it could be an inherent limitation of vanilla PINNs.
Table 5. Geometry with bad stability (16). Relative percentage generalization errors in Omega,n=5\Omega, n=5 and k=1k=1.
REFERENCES
[1] G. Alessandrini, L. Rondi, E. Rosset, and S. Vessella, The stability for the Cauchy problem for elliptic equations, Inverse Problems, 25 (2009), p. 123004, https: //doi.org/10.1088/0266-5611/25/12/123004.
[2] S. M. Berge and E. Malinnikova, On the three ball theorem for solutions of the Helmholtz equation, Complex Anal. Synerg., 7 (2021), pp. 1-10, https://doi.org/10 .1007/s40627-021-00070-3.
[3] E. Burman, Stabilized finite element methods for nonsymmetric, noncoercive, and illposed problems. Part I: elliptic equations, SIAM J. Sci. Comput., 35 (2013), pp. A2752A2780, https://doi.org/10.1137/130916862.
[4] E. Burman, G. Delay, and A. Ern, A hybridized high-order method for unique continuation subject to the Helmholtz equation, SIAM J. Numer. Anal., 59 (2021), pp. 23682392, https://doi.org/10.1137/20M1375619.
[5] E. Burman, P. Hansbo, and M. G. Larson, Solving ill-posed control problems by stabilized finite element methods: an alternative to Tikhonov regularization, Inverse Problems, 34 (2018), p. 035004, https://doi.org/10.1088/1361-6420/aaa32b.
[6] E. Burman, M. Nechita, and L. Oksanen, Unique continuation for the Helmholtz equation using stabilized finite element methods, J. Math. Pures Appl., 129 (2019), pp. 1-22, https://doi.org/10.1016/j.matpur.2018.10.003.
[7] R. H. Byrd, P. Lu, J. Nocedal, and C. Zhu, A limited memory algorithm for bound constrained optimization, SIAM J. Sci. Comput., 16 (1995), pp. 1190-1208, https: //doi.org/10.1137/0916069.
[8] P. Escapil-Inchauspé and G. A. Ruz, Hyper-parameter tuning of physics-informed neural networks: Application to Helmholtz problems, arXiv preprint arXiv:2205.06704, (2022).
[9] T. G. Grossmann, U. J. Komorowska, J. Latz, and C.-B. Schönlieb, Can physics-informed neural networks beat the finite element method?, arXiv preprint arXiv:2302.04107, (2023).
[10] T. Hrycak and V. Isakov, Increased stability in the continuation of solutions to the Helmholtz equation, Inverse Problems, 20 (2004), pp. 697-712, https://doi.org/10.1 088/0266-5611/23/4/019.
[11] V. Isakov, Inverse problems for partial differential equations, vol. 127 of Applied Mathematical Sciences, Springer, 3rd ed., 2017, https://doi.org/10.1007/978-3-319-516 58-5.
[12] D. P. Kingma and J. Ba, Adam: A method for stochastic optimization, arXiv preprint arXiv:1412.6980, (2014).
[13] L. Lu, X. Meng, Z. Mao, and G. E. Karniadakis, DeepXDE: A deep learning library for solving differential equations, SIAM Rev., 63 (2021), pp. 208-228, https: //doi.org/10.1137/19M1274067.
[14] S. Mishra and R. Molinaro, Estimates on the generalization error of physicsinformed neural networks for approximating a class of inverse problems for PDEs, IMA J. Numer. Anal., 42 (2022), pp. 981-1022, https://doi.org/10.1093/imanum/drab032.
[15] M. Nechita, Unique continuation problems and stabilised finite element methods, PhD thesis, University College London, 2020, https://discovery.ucl.ac.uk/id/eprint/ 10113065/.
[16] M. Raissi, P. Perdikaris, and G. E. Karniadakis, Physics-informed neural networks: A deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations, J. Comp. Phys., 378 (2019), pp. 686-707, https://doi.org/10.1016/j.jcp.2018.10.045.
Received by the editors: February 21, 2023; accepted: June 29, 2023; published online: July 5, 2023.
*This work was supported by the project "The Development of Advanced and Applicative Research Competencies in the Logic of STEAM + Health" /POCU/993/6/13/153310, project co-financed by the European Social Fund through The Romanian Operational Programme Human Capital 2014-2020. ^(†){ }^{\dagger} Tiberiu Popoviciu Institute of Numerical Analysis, Romanian Academy, Cluj-Napoca, Romania, e-mail: mihai.nechita@ictp.acad.ro, orcid.org/0000-0002-6448-912X. ^(‡){ }^{\ddagger} Department of Mathematics, Babeş-Bolyai University, Cluj-Napoca, Romania.