228 related articles for article (PubMed ID: 31178689)
1. Direct Feedback Alignment With Sparse Connections for Local Learning.
Crafton B; Parihar A; Gebhardt E; Raychowdhury A
Front Neurosci; 2019; 13():525. PubMed ID: 31178689
[TBL] [Abstract][Full Text] [Related]
2. Low-variance Forward Gradients using Direct Feedback Alignment and momentum.
Bacho F; Chu D
Neural Netw; 2024 Jan; 169():572-583. PubMed ID: 37956574
[TBL] [Abstract][Full Text] [Related]
3. Deep Supervised Learning Using Local Errors.
Mostafa H; Ramesh V; Cauwenberghs G
Front Neurosci; 2018; 12():608. PubMed ID: 30233295
[TBL] [Abstract][Full Text] [Related]
4. Deep Learning without Weight Symmetry.
Ji-An L; Benna MK
ArXiv; 2024 May; ():. PubMed ID: 38855537
[TBL] [Abstract][Full Text] [Related]
5. Spike-Train Level Direct Feedback Alignment: Sidestepping Backpropagation for On-Chip Training of Spiking Neural Nets.
Lee J; Zhang R; Zhang W; Liu Y; Li P
Front Neurosci; 2020; 14():143. PubMed ID: 32231513
[TBL] [Abstract][Full Text] [Related]
6. Hardware-Efficient On-line Learning through Pipelined Truncated-Error Backpropagation in Binary-State Networks.
Mostafa H; Pedroni B; Sheik S; Cauwenberghs G
Front Neurosci; 2017; 11():496. PubMed ID: 28932180
[TBL] [Abstract][Full Text] [Related]
7. Deep Learning With Asymmetric Connections and Hebbian Updates.
Amit Y
Front Comput Neurosci; 2019; 13():18. PubMed ID: 31019458
[TBL] [Abstract][Full Text] [Related]
8. GXNOR-Net: Training deep neural networks with ternary weights and activations without full-precision memory under a unified discretization framework.
Deng L; Jiao P; Pei J; Wu Z; Li G
Neural Netw; 2018 Apr; 100():49-58. PubMed ID: 29471195
[TBL] [Abstract][Full Text] [Related]
9. Accelerating DNN Training Through Selective Localized Learning.
Krithivasan S; Sen S; Venkataramani S; Raghunathan A
Front Neurosci; 2021; 15():759807. PubMed ID: 35087370
[TBL] [Abstract][Full Text] [Related]
10. Random synaptic feedback weights support error backpropagation for deep learning.
Lillicrap TP; Cownden D; Tweed DB; Akerman CJ
Nat Commun; 2016 Nov; 7():13276. PubMed ID: 27824044
[TBL] [Abstract][Full Text] [Related]
11. Biologically Plausible Training Mechanisms for Self-Supervised Learning in Deep Networks.
Tang M; Yang Y; Amit Y
Front Comput Neurosci; 2022; 16():789253. PubMed ID: 35386856
[TBL] [Abstract][Full Text] [Related]
12. Memristors for Neuromorphic Circuits and Artificial Intelligence Applications.
Miranda E; Suñé J
Materials (Basel); 2020 Feb; 13(4):. PubMed ID: 32093164
[TBL] [Abstract][Full Text] [Related]
13. Biologically plausible deep learning - But how far can we go with shallow networks?
Illing B; Gerstner W; Brea J
Neural Netw; 2019 Oct; 118():90-101. PubMed ID: 31254771
[TBL] [Abstract][Full Text] [Related]
14. Backpropagation Neural Tree.
Ojha V; Nicosia G
Neural Netw; 2022 May; 149():66-83. PubMed ID: 35193079
[TBL] [Abstract][Full Text] [Related]
15. Transformed ℓ
Ma R; Miao J; Niu L; Zhang P
Neural Netw; 2019 Nov; 119():286-298. PubMed ID: 31499353
[TBL] [Abstract][Full Text] [Related]
16. Event-based backpropagation can compute exact gradients for spiking neural networks.
Wunderlich TC; Pehle C
Sci Rep; 2021 Jun; 11(1):12829. PubMed ID: 34145314
[TBL] [Abstract][Full Text] [Related]
17. Cost-effective stochastic MAC circuits for deep neural networks.
Sim H; Lee J
Neural Netw; 2019 Sep; 117():152-162. PubMed ID: 31170575
[TBL] [Abstract][Full Text] [Related]
18. Backpropagation and the brain.
Lillicrap TP; Santoro A; Marris L; Akerman CJ; Hinton G
Nat Rev Neurosci; 2020 Jun; 21(6):335-346. PubMed ID: 32303713
[TBL] [Abstract][Full Text] [Related]
19. Deep learning in spiking neural networks.
Tavanaei A; Ghodrati M; Kheradpisheh SR; Masquelier T; Maida A
Neural Netw; 2019 Mar; 111():47-63. PubMed ID: 30682710
[TBL] [Abstract][Full Text] [Related]
20. Supervised Learning in Neural Networks: Feedback-Network-Free Implementation and Biological Plausibility.
Lin F
IEEE Trans Neural Netw Learn Syst; 2022 Dec; 33(12):7888-7898. PubMed ID: 34181554
[TBL] [Abstract][Full Text] [Related]
[Next] [New Search]