Multivariate error function based neural network approximations
May 1st, 2014.
Here we present multivariate quantitative approximations of real and complex valued continuous multivariate functions on a box or
MSC. 41A17, 41A25, 41A30, 41A36.
Keywords. error function, multivariate neural network approximation, quasi-interpolation operator, Baskakov type operator, quadrature type operator, multivariate modulus of continuity, complex approximation, iterated approximation.
1 Introduction
The author in
[
2
]
and
[
3
]
, see chapters 2-5, was the first to establish neural network approximations to continuous functions with rates by very specifically defined neural network operators of Cardaliagnet-Euvrard and ”Squashing” types, by employing the modulus of continuity of the engaged function or its high order derivative, and producing very tight Jackson type inequalities. He treats there both the univariate and multivariate cases. The defining these operators ”bell-shaped” and ”squashing” functions are assumed to be of compact support. Also in
[
3
]
he gives the
For this article the author is motivated by the article [ 12 ] of Z. Chen and F. Cao, also by [ 4 ] , [ 5 ] , [ 6 ] , [ 7 ] , [ 8 ] , [ 9 ] , [ 10 ] , [ 13 ] , [ 14 ] .
The author here performs multivariate error function based neural network approximations to continuous functions over boxes or over the whole
The author here comes up with the ”right” precisely defined multivariate quasi-interpolation neural network operators related to boxes or
Feed-forward neural networks (FNNs) with one hidden layer, the only type of networks we deal with in this article, are mathematically expressed as
where for
2 Basics
We consider here the (Gauss) error special function ( [ 1 ] , [ 11 ] )
which is a sigmoidal type function and is a strictly increasing function.
It has the basic properties
We consider the activation function ( [ 10 ] )
which is an even function.
Next we follow
[
10
]
on
We need the important
[
10
]
Let
Denote by
[
10
]
Let
Also from [ 10 ] we get
at least for some
For large enough
We introduce
It has the properties:
,where
,hence
and
that is
is a multivariate density function.
Here
where
We obviously see that
For
In the last two sums the counting is over disjoint vector sets of
We treat
when
We have proved that
, ,
By Theorem 3 clearly we obtain
That is,
it holds
It is also clear that
, ,
Also we get that
for at least some
Let
We introduce and define the multivariate positive linear neural network operator (
For large enough
For convenience we call
That is
Hence
Consequently we derive
We will estimate the right hand side of (29).
For the last we need, for
It holds that
Similarly it is defined for
When
Also for
Again for
where
We put
Let fixed
Let
We denote
Call also
In this article we study the basic approximation properties of
3 Multidimensional Real Neural Network Approximations
Here we present a series of neural network approximations to a function given with rates.
We give
Let
1)
and
2)
We notice that
Thus
So that
Now using (29) we finish proof.
We continue with
Let
1)
2)
Given that
Hence
proving the claim.
We give
Let
1)
2)
Given that
Thus it holds
We observe that
proving the claim.
We also present
Let
1)
2)
Given that
proving the claim.
In the next we discuss high order of approximation by using the smoothness of
We give
Let
i)
ii)
iii)
iv) Assume
notice in the last the extremely high rate of convergence at
Then
for all
We have the multivariate Taylor’s formula
Notice
Furthermore
So we treat
Thus, we have for
where
We see that
Notice here that
We further see that
Conclusion: When
In general we notice that
We proved in general that
Next we see that
Consequently
We have established that
We observe that
The last says
Clearly
Thus (here
So we have proved that
for all
At last we observe
Putting all of the above together we prove theorem.
We make
Let
Clearly
Hence
We need
Let
Next we prove the continuity of
We will use the Weierstrass
(a)
(b)
then
Also we will use:
If
We will prove that
There always exists
Since
So for
and
Hence by Weierstrass
Since
Because
So for
and
Hence by Weierstrass
Since
So we proved that
Writing
we have it as a continuous function on
When
(there always exist
(For convenience call
Thus
Notice that the finite sum of continuous functions
is a continuous function.
The rest of the summands of
We will prove that
is continuous in
The continuous function
and
So by the Weierstrass
is uniformly and absolutely convergent. Therefore it is continuous on
Next we prove continuity on
Notice here that
and
So the double series under consideration is uniformly convergent and continuous. Clearly
Similarly reasoning one can prove easily now, but with more tedious work, that
By (25) it is obvious that
Call
Clearly then
etc.
Therefore we get
the contraction property.
Also we see that
Also
Here
Here
and
We give the condensed
Let
(i)
(ii)
For
pointwise and uniformly.
Next we do iterated neural network approximation (see also [ 9 ] ).
We make
Let
Then
That is
We give
We make
We give
Let
Clearly, we notice that the speed of convergence to the unit operator of the multiply iterated operator is not worse than the speed of
We continue with
4 Complex Multivariate Neural Network Approximations
We make
Let
Given that
where
We denote by
Here
We define
We observe that
and
We present
Let
i)
ii)
In the next we discuss high order of complex approximation by using the smoothness of
We give
Let
i)
ii)
iii)
iv) Assume
notice in the last the extremely high rate of convergence at
Bibliography
- 1
M. Abramowitz and I. A. Stegun, eds., Handbook of Mathematical Functions with Formulas, Graphs, and Mathematical Tables, New York, Dover Publications, 1972.
- 2
G. A. Anastassiou, Rate of convergence of some neural network operators to the unit-univariate case, J. Math. Anal. Appli., 212 (1997), pp. 237–262.
- 3
G. A. Anastassiou, Quantitative Approximations, Chapman&Hall/CRC, Boca Raton, New York, 2001.
- 4
G. A. Anastassiou, Inteligent Systems: Approximation by Artificial Neural Networks, Intelligent Systems Reference Library, vol. 19, Springer, Heidelberg, 2011.
- 5
G. A. Anastassiou, Univariate hyperbolic tangent neural network approximation, Mathematics and Computer Modelling, 53 (2011), pp. 1111–1132.
- 6
G. A. Anastassiou, Multivariate hyperbolic tangent neural network approximation, Computers and Mathematics 61 (2011), pp. 809–821.
- 7
G. A. Anastassiou, Multivariate sigmoidal neural network approximation, Neural Networks 24 (2011), pp. 378–386.
- 8
G. A. Anastassiou, Univariate sigmoidal neural network approximation, J. of Computational Analysis and Applications, vol. 14 (2012), no. 4, pp. 659–690.
- 9
G. A. Anastassiou, Approximation by neural networks iterates, Advances in Applied Mathematics and Approximation Theory, pp. 1–20, Springer Proceedings in Math. & Stat., Springer, New York, 2013, Eds. G. Anastassiou, O. Duman.
- 10
G. A. Anastassiou, Univariate error function based neural network approximation, submitted, 2014.
- 11
L. C. Andrews, Special Functions of Mathematics for Engineers, Second edition, Mc Graw-Hill, New York, 1992.
- 12
Z. Chen and F. Cao, The approximation operators with sigmoidal functions, Computers and Mathematics with Applications, 58 (2009), pp. 758–765.
- 13
D. Costarelli and R. Spigler, Approximation results for neural network operators activated by sigmoidal functions, Neural Networks 44 (2013), pp. 101–106.
- 14
D. Costarelli and R. Spigler, Multivariate neural network operators with sigmoidal activation functions, Neural Networks 48 (2013), pp. 72–77.
- 15
S. Haykin, Neural Networks: A Comprehensive Foundation (2 ed.), Prentice Hall, New York, 1998.
- 16
W. McCulloch and W. Pitts, A logical calculus of the ideas immanent in nervous activity, Bulletin of Mathematical Biophysics, 7 (1943), pp. 115–133.
- 17
T. M. Mitchell, Machine Learning, WCB-McGraw-Hill, New York, 1997.