These tools will no longer be maintained as of December 31, 2024. Archived website can be found here. PubMed4Hh GitHub repository can be found here. Contact NLM Customer Service if you have questions.
4. Improving Exploration in Actor-Critic With Weakly Pessimistic Value Estimation and Optimistic Policy Optimization. Li F; Fu M; Chen W; Zhang F; Zhang H; Qu H; Yi Z IEEE Trans Neural Netw Learn Syst; 2024 Jul; 35(7):8783-8796. PubMed ID: 36306289 [TBL] [Abstract][Full Text] [Related]
5. Stochastic Integrated Actor-Critic for Deep Reinforcement Learning. Zheng J; Kurt MN; Wang X IEEE Trans Neural Netw Learn Syst; 2024 May; 35(5):6654-6666. PubMed ID: 36256721 [TBL] [Abstract][Full Text] [Related]
6. Actor-Critic Learning Control With Regularization and Feature Selection in Policy Gradient Estimation. Li L; Li D; Song T; Xu X IEEE Trans Neural Netw Learn Syst; 2021 Mar; 32(3):1217-1227. PubMed ID: 32324571 [TBL] [Abstract][Full Text] [Related]
7. Actor-Critic With Synthesis Loss for Solving Approximation Biases. Guo BW; Chao F; Chang X; Shang C; Shen Q IEEE Trans Cybern; 2024 Sep; 54(9):5323-5336. PubMed ID: 38700970 [TBL] [Abstract][Full Text] [Related]
8. Meta attention for Off-Policy Actor-Critic. Huang J; Huang W; Lan L; Wu D Neural Netw; 2023 Jun; 163():86-96. PubMed ID: 37030278 [TBL] [Abstract][Full Text] [Related]
9. Network Architecture for Optimizing Deep Deterministic Policy Gradient Algorithms. Zhang H; Xu J; Zhang J; Liu Q Comput Intell Neurosci; 2022; 2022():1117781. PubMed ID: 36438689 [TBL] [Abstract][Full Text] [Related]
10. Mild Policy Evaluation for Offline Actor-Critic. Huang L; Dong B; Lu J; Zhang W IEEE Trans Neural Netw Learn Syst; 2023 Sep; PP():. PubMed ID: 37676802 [TBL] [Abstract][Full Text] [Related]
11. The Actor-Dueling-Critic Method for Reinforcement Learning. Wu M; Gao Y; Jung A; Zhang Q; Du S Sensors (Basel); 2019 Mar; 19(7):. PubMed ID: 30935035 [TBL] [Abstract][Full Text] [Related]
12. Relative Entropy Regularized Sample-Efficient Reinforcement Learning With Continuous Actions. Shang Z; Li R; Zheng C; Li H; Cui Y IEEE Trans Neural Netw Learn Syst; 2023 Nov; PP():. PubMed ID: 37943648 [TBL] [Abstract][Full Text] [Related]
13. Reinforcement learning in continuous time and space. Doya K Neural Comput; 2000 Jan; 12(1):219-45. PubMed ID: 10636940 [TBL] [Abstract][Full Text] [Related]
14. De-Pessimism Offline Reinforcement Learning via Value Compensation. Huang Z; Zhao J; Sun S IEEE Trans Neural Netw Learn Syst; 2024 Aug; PP():. PubMed ID: 39178073 [TBL] [Abstract][Full Text] [Related]
15. Boosting On-Policy Actor-Critic With Shallow Updates in Critic. Li L; Zhu Y IEEE Trans Neural Netw Learn Syst; 2024 Apr; PP():. PubMed ID: 38619961 [TBL] [Abstract][Full Text] [Related]
16. Distributional generative adversarial imitation learning with reproducing kernel generalization. Zhou Y; Lu M; Liu X; Che Z; Xu Z; Tang J; Zhang Y; Peng Y; Peng Y Neural Netw; 2023 Aug; 165():43-59. PubMed ID: 37276810 [TBL] [Abstract][Full Text] [Related]
17. Distributional Policy Gradient With Distributional Value Function. Liu Q; Li Y; Shi X; Lin K; Liu Y; Lou Y IEEE Trans Neural Netw Learn Syst; 2024 Apr; PP():. PubMed ID: 38669170 [TBL] [Abstract][Full Text] [Related]
18. Offline Reinforcement Learning With Behavior Value Regularization. Huang L; Dong B; Xie W; Zhang W IEEE Trans Cybern; 2024 Jun; 54(6):3692-3704. PubMed ID: 38669164 [TBL] [Abstract][Full Text] [Related]
19. Realistic Actor-Critic: A framework for balance between value overestimation and underestimation. Li S; Tang Q; Pang Y; Ma X; Wang G Front Neurorobot; 2022; 16():1081242. PubMed ID: 36699950 [TBL] [Abstract][Full Text] [Related]