These tools will no longer be maintained as of December 31, 2024. Archived website can be found here. PubMed4Hh GitHub repository can be found here. Contact NLM Customer Service if you have questions.


PUBMED FOR HANDHELDS

Journal Abstract Search


243 related items for PubMed ID: 29993664

  • 1. Actor-Critic Learning Control Based on -Regularized Temporal-Difference Prediction With Gradient Correction.
    Li L, Li D, Song T, Xu X.
    IEEE Trans Neural Netw Learn Syst; 2018 Dec; 29(12):5899-5909. PubMed ID: 29993664
    [Abstract] [Full Text] [Related]

  • 2.
    ; . PubMed ID:
    [No Abstract] [Full Text] [Related]

  • 3.
    ; . PubMed ID:
    [No Abstract] [Full Text] [Related]

  • 4. Reinforcement learning in continuous time and space.
    Doya K.
    Neural Comput; 2000 Jan; 12(1):219-45. PubMed ID: 10636940
    [Abstract] [Full Text] [Related]

  • 5. Meta attention for Off-Policy Actor-Critic.
    Huang J, Huang W, Lan L, Wu D.
    Neural Netw; 2023 Jun; 163():86-96. PubMed ID: 37030278
    [Abstract] [Full Text] [Related]

  • 6. Efficient Actor-Critic Algorithm with Hierarchical Model Learning and Planning.
    Zhong S, Liu Q, Fu Q.
    Comput Intell Neurosci; 2016 Jun; 2016():4824072. PubMed ID: 27795704
    [Abstract] [Full Text] [Related]

  • 7.
    ; . PubMed ID:
    [No Abstract] [Full Text] [Related]

  • 8.
    ; . PubMed ID:
    [No Abstract] [Full Text] [Related]

  • 9.
    ; . PubMed ID:
    [No Abstract] [Full Text] [Related]

  • 10.
    ; . PubMed ID:
    [No Abstract] [Full Text] [Related]

  • 11. A policy iteration approach to online optimal control of continuous-time constrained-input systems.
    Modares H, Naghibi Sistani MB, Lewis FL.
    ISA Trans; 2013 Sep; 52(5):611-21. PubMed ID: 23706414
    [Abstract] [Full Text] [Related]

  • 12.
    ; . PubMed ID:
    [No Abstract] [Full Text] [Related]

  • 13. Mild Policy Evaluation for Offline Actor-Critic.
    Huang L, Dong B, Lu J, Zhang W.
    IEEE Trans Neural Netw Learn Syst; 2023 Sep 07; PP():. PubMed ID: 37676802
    [Abstract] [Full Text] [Related]

  • 14.
    ; . PubMed ID:
    [No Abstract] [Full Text] [Related]

  • 15.
    ; . PubMed ID:
    [No Abstract] [Full Text] [Related]

  • 16. Efficient model learning methods for actor-critic control.
    Grondman I, Vaandrager M, Buşoniu L, Babuska R, Schuitema E.
    IEEE Trans Syst Man Cybern B Cybern; 2012 Jun 07; 42(3):591-602. PubMed ID: 22156998
    [Abstract] [Full Text] [Related]

  • 17. Distributional Soft Actor-Critic: Off-Policy Reinforcement Learning for Addressing Value Estimation Errors.
    Duan J, Guan Y, Li SE, Ren Y, Sun Q, Cheng B.
    IEEE Trans Neural Netw Learn Syst; 2022 Nov 07; 33(11):6584-6598. PubMed ID: 34101599
    [Abstract] [Full Text] [Related]

  • 18. Deep Multi-Critic Network for accelerating Policy Learning in multi-agent environments.
    Hook J, Silva V, Kondoz A.
    Neural Netw; 2020 Aug 07; 128():97-106. PubMed ID: 32446194
    [Abstract] [Full Text] [Related]

  • 19. Reinforcement learning solution for HJB equation arising in constrained optimal control problem.
    Luo B, Wu HN, Huang T, Liu D.
    Neural Netw; 2015 Nov 07; 71():150-8. PubMed ID: 26356598
    [Abstract] [Full Text] [Related]

  • 20. Deep Deterministic Policy Gradient With Compatible Critic Network.
    Wang D, Hu M.
    IEEE Trans Neural Netw Learn Syst; 2023 Aug 07; 34(8):4332-4344. PubMed ID: 34653007
    [Abstract] [Full Text] [Related]


    Page: [Next] [New Search]
    of 13.