These tools will no longer be maintained as of December 31, 2024. Archived website can be found here. PubMed4Hh GitHub repository can be found here. Contact NLM Customer Service if you have questions.
9. On the computational power of winner-take-all. Maass W Neural Comput; 2000 Nov; 12(11):2519-35. PubMed ID: 11110125 [TBL] [Abstract][Full Text] [Related]
10. Reduction strategies for hierarchical multi-label classification in protein function prediction. Cerri R; Barros RC; P L F de Carvalho AC; Jin Y BMC Bioinformatics; 2016 Sep; 17(1):373. PubMed ID: 27627880 [TBL] [Abstract][Full Text] [Related]
11. Concept hierarchy memory model: a neural architecture for conceptual knowledge representation, learning, and commonsense reasoning. Tan AH; Soon HS Int J Neural Syst; 1996 Jul; 7(3):305-19. PubMed ID: 8891845 [TBL] [Abstract][Full Text] [Related]
12. Explaining synchrony in feed-forward networks: are McCulloch-Pitts neurons good enough? Nowotny T; Huerta R Biol Cybern; 2003 Oct; 89(4):237-41. PubMed ID: 14605888 [TBL] [Abstract][Full Text] [Related]
13. Extension of mixture-of-experts networks for binary classification of hierarchical data. Ng SK; McLachlan GJ Artif Intell Med; 2007 Sep; 41(1):57-67. PubMed ID: 17629686 [TBL] [Abstract][Full Text] [Related]
14. Controlling activity fluctuations in large, sparsely connected random networks. Smith AC; Wu XB; Levy WB Network; 2000 Feb; 11(1):63-81. PubMed ID: 10735529 [TBL] [Abstract][Full Text] [Related]
15. NeuCube: a spiking neural network architecture for mapping, learning and understanding of spatio-temporal brain data. Kasabov NK Neural Netw; 2014 Apr; 52():62-76. PubMed ID: 24508754 [TBL] [Abstract][Full Text] [Related]
16. The computational power of interactive recurrent neural networks. Cabessa J; Siegelmann HT Neural Comput; 2012 Apr; 24(4):996-1019. PubMed ID: 22295978 [TBL] [Abstract][Full Text] [Related]
17. A modular architecture for transparent computation in recurrent neural networks. Carmantini GS; Beim Graben P; Desroches M; Rodrigues S Neural Netw; 2017 Jan; 85():85-105. PubMed ID: 27814468 [TBL] [Abstract][Full Text] [Related]
18. A machine learning method for extracting symbolic knowledge from recurrent neural networks. Vahed A; Omlin CW Neural Comput; 2004 Jan; 16(1):59-71. PubMed ID: 15006023 [TBL] [Abstract][Full Text] [Related]
19. On the number of different dynamics in Boolean networks with deterministic update schedules. Aracena J; Demongeot J; Fanchon E; Montalva M Math Biosci; 2013 Apr; 242(2):188-94. PubMed ID: 23384974 [TBL] [Abstract][Full Text] [Related]
20. Identification of finite state automata with a class of recurrent neural networks. Won SH; Song I; Lee SY; Park CH IEEE Trans Neural Netw; 2010 Sep; 21(9):1408-21. PubMed ID: 20709639 [TBL] [Abstract][Full Text] [Related] [Next] [New Search]