These tools will no longer be maintained as of December 31, 2024. Archived website can be found here. PubMed4Hh GitHub repository can be found here. Contact NLM Customer Service if you have questions.


BIOMARKERS

Molecular Biopsy of Human Tumors

- a resource for Precision Medicine *

121 related articles for article (PubMed ID: 34648431)

  • 1. Adaptive Temporal Difference Learning With Linear Function Approximation.
    Sun T; Shen H; Chen T; Li D
    IEEE Trans Pattern Anal Mach Intell; 2022 Dec; 44(12):8812-8824. PubMed ID: 34648431
    [TBL] [Abstract][Full Text] [Related]  

  • 2. Reinforcement learning via kernel temporal difference.
    Bae J; Chhatbar P; Francis JT; Sanchez JC; Principe JC
    Annu Int Conf IEEE Eng Med Biol Soc; 2011; 2011():5662-5. PubMed ID: 22255624
    [TBL] [Abstract][Full Text] [Related]  

  • 3. Kernel-based least squares policy iteration for reinforcement learning.
    Xu X; Hu D; Lu X
    IEEE Trans Neural Netw; 2007 Jul; 18(4):973-92. PubMed ID: 17668655
    [TBL] [Abstract][Full Text] [Related]  

  • 4. Semi-Infinitely Constrained Markov Decision Processes and Provably Efficient Reinforcement Learning.
    Zhang L; Peng Y; Yang W; Zhang Z
    IEEE Trans Pattern Anal Mach Intell; 2024 May; 46(5):3722-3735. PubMed ID: 38163315
    [TBL] [Abstract][Full Text] [Related]  

  • 5. Value iteration for streaming data on a continuous space with gradient method in an RKHS.
    Liu J; Xu W; Wang Y; Lian H
    Neural Netw; 2023 Sep; 166():437-445. PubMed ID: 37566954
    [TBL] [Abstract][Full Text] [Related]  

  • 6. Deep Reinforcement Learning With Modulated Hebbian Plus Q-Network Architecture.
    Ladosz P; Ben-Iwhiwhu E; Dick J; Ketz N; Kolouri S; Krichmar JL; Pilly PK; Soltoggio A
    IEEE Trans Neural Netw Learn Syst; 2022 May; 33(5):2045-2056. PubMed ID: 34559664
    [TBL] [Abstract][Full Text] [Related]  

  • 7. Evaluating the TD model of classical conditioning.
    Ludvig EA; Sutton RS; Kehoe EJ
    Learn Behav; 2012 Sep; 40(3):305-19. PubMed ID: 22927003
    [TBL] [Abstract][Full Text] [Related]  

  • 8. Adaptive Learning Algorithm Convergence in Passive and Reactive Environments.
    Golden RM
    Neural Comput; 2018 Oct; 30(10):2805-2832. PubMed ID: 30021080
    [TBL] [Abstract][Full Text] [Related]  

  • 9. Integrating temporal difference methods and self-organizing neural networks for reinforcement learning with delayed evaluative feedback.
    Tan AH; Lu N; Xiao D
    IEEE Trans Neural Netw; 2008 Feb; 19(2):230-44. PubMed ID: 18269955
    [TBL] [Abstract][Full Text] [Related]  

  • 10. Reinforcement learning in continuous time and space.
    Doya K
    Neural Comput; 2000 Jan; 12(1):219-45. PubMed ID: 10636940
    [TBL] [Abstract][Full Text] [Related]  

  • 11. Actor-Critic Learning Control Based on -Regularized Temporal-Difference Prediction With Gradient Correction.
    Li L; Li D; Song T; Xu X
    IEEE Trans Neural Netw Learn Syst; 2018 Dec; 29(12):5899-5909. PubMed ID: 29993664
    [TBL] [Abstract][Full Text] [Related]  

  • 12. Kernel Temporal Difference based Reinforcement Learning for Brain Machine Interfaces
    Shen X; Zhang X; Wang Y
    Annu Int Conf IEEE Eng Med Biol Soc; 2021 Nov; 2021():6721-6724. PubMed ID: 34892650
    [TBL] [Abstract][Full Text] [Related]  

  • 13. Convergence of the RMSProp deep learning method with penalty for nonconvex optimization.
    Xu D; Zhang S; Zhang H; Mandic DP
    Neural Netw; 2021 Jul; 139():17-23. PubMed ID: 33662649
    [TBL] [Abstract][Full Text] [Related]  

  • 14. Reinforcement learning in continuous time and space: interference and not ill conditioning is the main problem when using distributed function approximators.
    Baddeley B
    IEEE Trans Syst Man Cybern B Cybern; 2008 Aug; 38(4):950-6. PubMed ID: 18632383
    [TBL] [Abstract][Full Text] [Related]  

  • 15. Risk-sensitive reinforcement learning.
    Shen Y; Tobia MJ; Sommer T; Obermayer K
    Neural Comput; 2014 Jul; 26(7):1298-328. PubMed ID: 24708369
    [TBL] [Abstract][Full Text] [Related]  

  • 16. Reinforcement learning for partially observable dynamic processes: adaptive dynamic programming using measured output data.
    Lewis FL; Vamvoudakis KG
    IEEE Trans Syst Man Cybern B Cybern; 2011 Feb; 41(1):14-25. PubMed ID: 20350860
    [TBL] [Abstract][Full Text] [Related]  

  • 17. Off-Policy Prediction Learning: An Empirical Study of Online Algorithms.
    Ghiassian S; Rafiee B; Sutton RS
    IEEE Trans Neural Netw Learn Syst; 2024 Jun; PP():. PubMed ID: 38857133
    [TBL] [Abstract][Full Text] [Related]  

  • 18. Online Model-Free n-Step HDP With Stability Analysis.
    Al-Dabooni S; Wunsch DC
    IEEE Trans Neural Netw Learn Syst; 2020 Apr; 31(4):1255-1269. PubMed ID: 31251198
    [TBL] [Abstract][Full Text] [Related]  

  • 19. Sigmoid-weighted linear units for neural network function approximation in reinforcement learning.
    Elfwing S; Uchibe E; Doya K
    Neural Netw; 2018 Nov; 107():3-11. PubMed ID: 29395652
    [TBL] [Abstract][Full Text] [Related]  

  • 20. Minibatch Recursive Least Squares Q-Learning.
    Zhang C; Song Q; Meng Z
    Comput Intell Neurosci; 2021; 2021():5370281. PubMed ID: 34659393
    [TBL] [Abstract][Full Text] [Related]  

    [Next]    [New Search]
    of 7.