These tools will no longer be maintained as of December 31, 2024. Archived website can be found here. PubMed4Hh GitHub repository can be found here. Contact NLM Customer Service if you have questions.


BIOMARKERS

Molecular Biopsy of Human Tumors

- a resource for Precision Medicine *

149 related articles for article (PubMed ID: 24808427)

  • 1. Policy improvement by a model-free Dyna architecture.
    Hwang KS; Lo CY
    IEEE Trans Neural Netw Learn Syst; 2013 May; 24(5):776-88. PubMed ID: 24808427
    [TBL] [Abstract][Full Text] [Related]  

  • 2. Model learning and knowledge sharing for a multiagent system with Dyna-Q learning.
    Hwang KS; Jiang WC; Chen YJ
    IEEE Trans Cybern; 2015 May; 45(5):964-76. PubMed ID: 25122850
    [TBL] [Abstract][Full Text] [Related]  

  • 3. Neuromuscular control of the point to point and oscillatory movements of a sagittal arm with the actor-critic reinforcement learning method.
    Golkhou V; Parnianpour M; Lucas C
    Comput Methods Biomech Biomed Engin; 2005 Apr; 8(2):103-13. PubMed ID: 16154874
    [TBL] [Abstract][Full Text] [Related]  

  • 4. Reinforcement learning to adaptive control of nonlinear systems.
    Hwang KS; Tan SW; Tsai MC
    IEEE Trans Syst Man Cybern B Cybern; 2003; 33(3):514-21. PubMed ID: 18238198
    [TBL] [Abstract][Full Text] [Related]  

  • 5. The role of multisensor data fusion in neuromuscular control of a sagittal arm with a pair of muscles using actor-critic reinforcement learning method.
    Golkhou V; Parnianpour M; Lucas C
    Technol Health Care; 2004; 12(6):425-38. PubMed ID: 15671597
    [TBL] [Abstract][Full Text] [Related]  

  • 6. Goal representation heuristic dynamic programming on maze navigation.
    Ni Z; He H; Wen J; Xu X
    IEEE Trans Neural Netw Learn Syst; 2013 Dec; 24(12):2038-50. PubMed ID: 24805221
    [TBL] [Abstract][Full Text] [Related]  

  • 7. Actor-critic-based optimal tracking for partially unknown nonlinear discrete-time systems.
    Kiumarsi B; Lewis FL
    IEEE Trans Neural Netw Learn Syst; 2015 Jan; 26(1):140-51. PubMed ID: 25312944
    [TBL] [Abstract][Full Text] [Related]  

  • 8. Model-Free Optimal Tracking Control via Critic-Only Q-Learning.
    Luo B; Liu D; Huang T; Wang D
    IEEE Trans Neural Netw Learn Syst; 2016 Oct; 27(10):2134-44. PubMed ID: 27416608
    [TBL] [Abstract][Full Text] [Related]  

  • 9. Reinforcement learning in continuous time and space.
    Doya K
    Neural Comput; 2000 Jan; 12(1):219-45. PubMed ID: 10636940
    [TBL] [Abstract][Full Text] [Related]  

  • 10. Improved Adaptive-Reinforcement Learning Control for morphing unmanned air vehicles.
    Valasek J; Doebbler J; Tandale MD; Meade AJ
    IEEE Trans Syst Man Cybern B Cybern; 2008 Aug; 38(4):1014-20. PubMed ID: 18632393
    [TBL] [Abstract][Full Text] [Related]  

  • 11. Efficient model learning methods for actor-critic control.
    Grondman I; Vaandrager M; Buşoniu L; Babuska R; Schuitema E
    IEEE Trans Syst Man Cybern B Cybern; 2012 Jun; 42(3):591-602. PubMed ID: 22156998
    [TBL] [Abstract][Full Text] [Related]  

  • 12. Actor-Critic Learning Control Based on -Regularized Temporal-Difference Prediction With Gradient Correction.
    Li L; Li D; Song T; Xu X
    IEEE Trans Neural Netw Learn Syst; 2018 Dec; 29(12):5899-5909. PubMed ID: 29993664
    [TBL] [Abstract][Full Text] [Related]  

  • 13. Humanoids Learning to Walk: A Natural CPG-Actor-Critic Architecture.
    Li C; Lowe R; Ziemke T
    Front Neurorobot; 2013; 7():5. PubMed ID: 23675345
    [TBL] [Abstract][Full Text] [Related]  

  • 14. Data-driven model reference control of MIMO vertical tank systems with model-free VRFT and Q-Learning.
    Radac MB; Precup RE; Roman RC
    ISA Trans; 2018 Feb; 73():227-238. PubMed ID: 29325777
    [TBL] [Abstract][Full Text] [Related]  

  • 15. A model of hippocampally dependent navigation, using the temporal difference learning rule.
    Foster DJ; Morris RG; Dayan P
    Hippocampus; 2000; 10(1):1-16. PubMed ID: 10706212
    [TBL] [Abstract][Full Text] [Related]  

  • 16. Controlling chaos by GA-based reinforcement learning neural network.
    Lin CT; Jou CP
    IEEE Trans Neural Netw; 1999; 10(4):846-59. PubMed ID: 18252582
    [TBL] [Abstract][Full Text] [Related]  

  • 17. A parameter control method in reinforcement learning to rapidly follow unexpected environmental changes.
    Murakoshi K; Mizuno J
    Biosystems; 2004 Nov; 77(1-3):109-17. PubMed ID: 15527950
    [TBL] [Abstract][Full Text] [Related]  

  • 18. Online learning of shaping rewards in reinforcement learning.
    Grześ M; Kudenko D
    Neural Netw; 2010 May; 23(4):541-50. PubMed ID: 20116208
    [TBL] [Abstract][Full Text] [Related]  

  • 19. Target Tracking Control of a Biomimetic Underwater Vehicle Through Deep Reinforcement Learning.
    Wang Y; Tang C; Wang S; Cheng L; Wang R; Tan M; Hou Z
    IEEE Trans Neural Netw Learn Syst; 2022 Aug; 33(8):3741-3752. PubMed ID: 33560993
    [TBL] [Abstract][Full Text] [Related]  

  • 20. Policy-Gradient and Actor-Critic Based State Representation Learning for Safe Driving of Autonomous Vehicles.
    Gupta A; Khwaja AS; Anpalagan A; Guan L; Venkatesh B
    Sensors (Basel); 2020 Oct; 20(21):. PubMed ID: 33105863
    [TBL] [Abstract][Full Text] [Related]  

    [Next]    [New Search]
    of 8.