143 related articles for article (PubMed ID: 33592434)
1. The Kolmogorov-Arnold representation theorem revisited.
Schmidt-Hieber J
Neural Netw; 2021 May; 137():119-126. PubMed ID: 33592434
[TBL] [Abstract][Full Text] [Related]
2. Error bounds for deep ReLU networks using the Kolmogorov-Arnold superposition theorem.
Montanelli H; Yang H
Neural Netw; 2020 Sep; 129():1-6. PubMed ID: 32473577
[TBL] [Abstract][Full Text] [Related]
3. Optimal approximation of piecewise smooth functions using deep ReLU neural networks.
Petersen P; Voigtlaender F
Neural Netw; 2018 Dec; 108():296-330. PubMed ID: 30245431
[TBL] [Abstract][Full Text] [Related]
4. Deep ReLU neural networks in high-dimensional approximation.
Dũng D; Nguyen VK
Neural Netw; 2021 Oct; 142():619-635. PubMed ID: 34392126
[TBL] [Abstract][Full Text] [Related]
5. Dimension independent bounds for general shallow networks.
Mhaskar HN
Neural Netw; 2020 Mar; 123():142-152. PubMed ID: 31869651
[TBL] [Abstract][Full Text] [Related]
6. On the approximation of functions by tanh neural networks.
De Ryck T; Lanthaler S; Mishra S
Neural Netw; 2021 Nov; 143():732-750. PubMed ID: 34482172
[TBL] [Abstract][Full Text] [Related]
7. On the Kolmogorov neural networks.
Ismayilova A; Ismailov VE
Neural Netw; 2024 Aug; 176():106333. PubMed ID: 38688072
[TBL] [Abstract][Full Text] [Related]
8. Approximation of smooth functionals using deep ReLU networks.
Song L; Liu Y; Fan J; Zhou DX
Neural Netw; 2023 Sep; 166():424-436. PubMed ID: 37549610
[TBL] [Abstract][Full Text] [Related]
9. Neural networks with ReLU powers need less depth.
Cabanilla KIM; Mohammad RZ; Lope JEC
Neural Netw; 2024 Apr; 172():106073. PubMed ID: 38159509
[TBL] [Abstract][Full Text] [Related]
10. Nonlinear approximation via compositions.
Shen Z; Yang H; Zhang S
Neural Netw; 2019 Nov; 119():74-84. PubMed ID: 31401528
[TBL] [Abstract][Full Text] [Related]
11. Simultaneous approximation of a smooth function and its derivatives by deep neural networks with piecewise-polynomial activations.
Belomestny D; Naumov A; Puchkin N; Samsonov S
Neural Netw; 2023 Apr; 161():242-253. PubMed ID: 36774863
[TBL] [Abstract][Full Text] [Related]
12. Two-hidden-layer feed-forward networks are universal approximators: A constructive approach.
Paluzo-Hidalgo E; Gonzalez-Diaz R; Gutiérrez-Naranjo MA
Neural Netw; 2020 Nov; 131():29-36. PubMed ID: 32739651
[TBL] [Abstract][Full Text] [Related]
13. Approximation properties of Gaussian-binary restricted Boltzmann machines and Gaussian-binary deep belief networks.
Gu L; Yang L; Zhou F
Neural Netw; 2022 Sep; 153():49-63. PubMed ID: 35700559
[TBL] [Abstract][Full Text] [Related]
14. Neural network approximation: Three hidden layers are enough.
Shen Z; Yang H; Zhang S
Neural Netw; 2021 Sep; 141():160-173. PubMed ID: 33906082
[TBL] [Abstract][Full Text] [Related]
15. On the approximation by single hidden layer feedforward neural networks with fixed weights.
Guliyev NJ; Ismailov VE
Neural Netw; 2018 Feb; 98():296-304. PubMed ID: 29301110
[TBL] [Abstract][Full Text] [Related]
16. On minimal representations of shallow ReLU networks.
Dereich S; Kassing S
Neural Netw; 2022 Apr; 148():121-128. PubMed ID: 35123261
[TBL] [Abstract][Full Text] [Related]
17. Approximation rates for neural networks with encodable weights in smoothness spaces.
Gühring I; Raslan M
Neural Netw; 2021 Feb; 134():107-130. PubMed ID: 33310376
[TBL] [Abstract][Full Text] [Related]
18. Neural network interpolation operators optimized by Lagrange polynomial.
Wang G; Yu D; Zhou P
Neural Netw; 2022 Sep; 153():179-191. PubMed ID: 35728337
[TBL] [Abstract][Full Text] [Related]
19. ReLU Networks Are Universal Approximators via Piecewise Linear or Constant Functions.
Huang C
Neural Comput; 2020 Nov; 32(11):2249-2278. PubMed ID: 32946706
[TBL] [Abstract][Full Text] [Related]
20. Approximation in shift-invariant spaces with deep ReLU neural networks.
Yang Y; Li Z; Wang Y
Neural Netw; 2022 Sep; 153():269-281. PubMed ID: 35763879
[TBL] [Abstract][Full Text] [Related]
[Next] [New Search]