139 related articles for article (PubMed ID: 34146968)
1. A deep network construction that adapts to intrinsic dimensionality beyond the domain.
Cloninger A; Klock T
Neural Netw; 2021 Sep; 141():404-419. PubMed ID: 34146968
[TBL] [Abstract][Full Text] [Related]
2. Optimal approximation of piecewise smooth functions using deep ReLU neural networks.
Petersen P; Voigtlaender F
Neural Netw; 2018 Dec; 108():296-330. PubMed ID: 30245431
[TBL] [Abstract][Full Text] [Related]
3. Low dimensional approximation and generalization of multivariate functions on smooth manifolds using deep ReLU neural networks.
Labate D; Shi J
Neural Netw; 2024 Jun; 174():106223. PubMed ID: 38458005
[TBL] [Abstract][Full Text] [Related]
4. Dimension independent bounds for general shallow networks.
Mhaskar HN
Neural Netw; 2020 Mar; 123():142-152. PubMed ID: 31869651
[TBL] [Abstract][Full Text] [Related]
5. On the capacity of deep generative networks for approximating distributions.
Yang Y; Li Z; Wang Y
Neural Netw; 2022 Jan; 145():144-154. PubMed ID: 34749027
[TBL] [Abstract][Full Text] [Related]
6. Deep ReLU neural networks in high-dimensional approximation.
Dũng D; Nguyen VK
Neural Netw; 2021 Oct; 142():619-635. PubMed ID: 34392126
[TBL] [Abstract][Full Text] [Related]
7. Theory of deep convolutional neural networks III: Approximating radial functions.
Mao T; Shi Z; Zhou DX
Neural Netw; 2021 Dec; 144():778-790. PubMed ID: 34688019
[TBL] [Abstract][Full Text] [Related]
8. Error bounds for deep ReLU networks using the Kolmogorov-Arnold superposition theorem.
Montanelli H; Yang H
Neural Netw; 2020 Sep; 129():1-6. PubMed ID: 32473577
[TBL] [Abstract][Full Text] [Related]
9. On minimal representations of shallow ReLU networks.
Dereich S; Kassing S
Neural Netw; 2022 Apr; 148():121-128. PubMed ID: 35123261
[TBL] [Abstract][Full Text] [Related]
10. Neural networks with ReLU powers need less depth.
Cabanilla KIM; Mohammad RZ; Lope JEC
Neural Netw; 2024 Apr; 172():106073. PubMed ID: 38159509
[TBL] [Abstract][Full Text] [Related]
11. Approximation in shift-invariant spaces with deep ReLU neural networks.
Yang Y; Li Z; Wang Y
Neural Netw; 2022 Sep; 153():269-281. PubMed ID: 35763879
[TBL] [Abstract][Full Text] [Related]
12. Nonlinear approximation via compositions.
Shen Z; Yang H; Zhang S
Neural Netw; 2019 Nov; 119():74-84. PubMed ID: 31401528
[TBL] [Abstract][Full Text] [Related]
13. Neural network approximation: Three hidden layers are enough.
Shen Z; Yang H; Zhang S
Neural Netw; 2021 Sep; 141():160-173. PubMed ID: 33906082
[TBL] [Abstract][Full Text] [Related]
14. Approximation of smooth functionals using deep ReLU networks.
Song L; Liu Y; Fan J; Zhou DX
Neural Netw; 2023 Sep; 166():424-436. PubMed ID: 37549610
[TBL] [Abstract][Full Text] [Related]
15. High-Dimensional Function Approximation With Neural Networks for Large Volumes of Data.
Andras P
IEEE Trans Neural Netw Learn Syst; 2018 Feb; 29(2):500-508. PubMed ID: 28129193
[TBL] [Abstract][Full Text] [Related]
16. A direct approach for function approximation on data defined manifolds.
Mhaskar HN
Neural Netw; 2020 Dec; 132():253-268. PubMed ID: 32927428
[TBL] [Abstract][Full Text] [Related]
17. Efficient Approximation of High-Dimensional Functions With Neural Networks.
Cheridito P; Jentzen A; Rossmannek F
IEEE Trans Neural Netw Learn Syst; 2022 Jul; 33(7):3079-3093. PubMed ID: 33513112
[TBL] [Abstract][Full Text] [Related]
18. Optimizing neural networks for medical data sets: A case study on neonatal apnea prediction.
Shirwaikar RD; Acharya U D; Makkithaya K; M S; Srivastava S; Lewis U LES
Artif Intell Med; 2019 Jul; 98():59-76. PubMed ID: 31521253
[TBL] [Abstract][Full Text] [Related]
19. ReLU Networks Are Universal Approximators via Piecewise Linear or Constant Functions.
Huang C
Neural Comput; 2020 Nov; 32(11):2249-2278. PubMed ID: 32946706
[TBL] [Abstract][Full Text] [Related]
20. Deep Network With Approximation Error Being Reciprocal of Width to Power of Square Root of Depth.
Shen Z; Yang H; Zhang S
Neural Comput; 2021 Mar; 33(4):1005-1036. PubMed ID: 33513325
[TBL] [Abstract][Full Text] [Related]
[Next] [New Search]