BIOMARKERS

Molecular Biopsy of Human Tumors

- a resource for Precision Medicine *

157 related articles for article (PubMed ID: 15312837)

  • 1. Reliability of internal prediction/estimation and its application. I. Adaptive action selection reflecting reliability of value function.
    Sakaguchi Y; Takano M
    Neural Netw; 2004 Sep; 17(7):935-52. PubMed ID: 15312837
    [TBL] [Abstract][Full Text] [Related]  

  • 2. Adaptive importance sampling for value function approximation in off-policy reinforcement learning.
    Hachiya H; Akiyama T; Sugiayma M; Peters J
    Neural Netw; 2009 Dec; 22(10):1399-410. PubMed ID: 19216050
    [TBL] [Abstract][Full Text] [Related]  

  • 3. A parameter control method in reinforcement learning to rapidly follow unexpected environmental changes.
    Murakoshi K; Mizuno J
    Biosystems; 2004 Nov; 77(1-3):109-17. PubMed ID: 15527950
    [TBL] [Abstract][Full Text] [Related]  

  • 4. Ensemble algorithms in reinforcement learning.
    Wiering MA; van Hasselt H
    IEEE Trans Syst Man Cybern B Cybern; 2008 Aug; 38(4):930-6. PubMed ID: 18632380
    [TBL] [Abstract][Full Text] [Related]  

  • 5. [Mathematical models of decision making and learning].
    Ito M; Doya K
    Brain Nerve; 2008 Jul; 60(7):791-8. PubMed ID: 18646619
    [TBL] [Abstract][Full Text] [Related]  

  • 6. Goal-directed learning of features and forward models.
    Saeb S; Weber C; Triesch J
    Neural Netw; 2009; 22(5-6):586-92. PubMed ID: 19616917
    [TBL] [Abstract][Full Text] [Related]  

  • 7. Fuzzy-rule emulated networks, based on reinforcement learning for nonlinear discrete-time controllers.
    Treesatayapun C
    ISA Trans; 2008 Oct; 47(4):362-73. PubMed ID: 18675416
    [TBL] [Abstract][Full Text] [Related]  

  • 8. A spiking neural network model of an actor-critic learning agent.
    Potjans W; Morrison A; Diesmann M
    Neural Comput; 2009 Feb; 21(2):301-39. PubMed ID: 19196231
    [TBL] [Abstract][Full Text] [Related]  

  • 9. Incremental state aggregation for value function estimation in reinforcement learning.
    Mori T; Ishii S
    IEEE Trans Syst Man Cybern B Cybern; 2011 Oct; 41(5):1407-16. PubMed ID: 21632307
    [TBL] [Abstract][Full Text] [Related]  

  • 10. Reinforcement-learning-based output-feedback control of nonstrict nonlinear discrete-time systems with application to engine emission control.
    Shih P; Kaul BC; Jagannathan S; Drallmeier JA
    IEEE Trans Syst Man Cybern B Cybern; 2009 Oct; 39(5):1162-79. PubMed ID: 19336317
    [TBL] [Abstract][Full Text] [Related]  

  • 11. Adaptive learning via selectionism and Bayesianism, Part I: connection between the two.
    Zhang J
    Neural Netw; 2009 Apr; 22(3):220-8. PubMed ID: 19386469
    [TBL] [Abstract][Full Text] [Related]  

  • 12. A model for learning to segment temporal sequences, utilizing a mixture of RNN experts together with adaptive variance.
    Namikawa J; Tani J
    Neural Netw; 2008 Dec; 21(10):1466-75. PubMed ID: 18938059
    [TBL] [Abstract][Full Text] [Related]  

  • 13. Adaptive dynamic programming approach to experience-based systems identification and control.
    Lendaris GG
    Neural Netw; 2009; 22(5-6):822-32. PubMed ID: 19632087
    [TBL] [Abstract][Full Text] [Related]  

  • 14. Adaptive learning via selectionism and Bayesianism, Part II: the sequential case.
    Zhang J
    Neural Netw; 2009 Apr; 22(3):229-36. PubMed ID: 19395235
    [TBL] [Abstract][Full Text] [Related]  

  • 15. Reinforcement learning state estimator.
    Morimoto J; Doya K
    Neural Comput; 2007 Mar; 19(3):730-56. PubMed ID: 17298231
    [TBL] [Abstract][Full Text] [Related]  

  • 16. A neural learning classifier system with self-adaptive constructivism for mobile robot control.
    Hurst J; Bull L
    Artif Life; 2006; 12(3):353-80. PubMed ID: 16859445
    [TBL] [Abstract][Full Text] [Related]  

  • 17. Multiple model-based reinforcement learning.
    Doya K; Samejima K; Katagiri K; Kawato M
    Neural Comput; 2002 Jun; 14(6):1347-69. PubMed ID: 12020450
    [TBL] [Abstract][Full Text] [Related]  

  • 18. Composite adaptive control with locally weighted statistical learning.
    Nakanishi J; Farrell JA; Schaal S
    Neural Netw; 2005 Jan; 18(1):71-90. PubMed ID: 15649663
    [TBL] [Abstract][Full Text] [Related]  

  • 19. Improved Adaptive-Reinforcement Learning Control for morphing unmanned air vehicles.
    Valasek J; Doebbler J; Tandale MD; Meade AJ
    IEEE Trans Syst Man Cybern B Cybern; 2008 Aug; 38(4):1014-20. PubMed ID: 18632393
    [TBL] [Abstract][Full Text] [Related]  

  • 20. Machine learning approaches for estimation of prediction interval for the model output.
    Shrestha DL; Solomatine DP
    Neural Netw; 2006 Mar; 19(2):225-35. PubMed ID: 16530384
    [TBL] [Abstract][Full Text] [Related]  

    [Next]    [New Search]
    of 8.