These tools will no longer be maintained as of December 31, 2024. Archived website can be found here. PubMed4Hh GitHub repository can be found here. Contact NLM Customer Service if you have questions.
3. An immediate-return reinforcement learning for the atypical Markov decision processes. Pan Z; Wen G; Tan Z; Yin S; Hu X Front Neurorobot; 2022; 16():1012427. PubMed ID: 36582302 [TBL] [Abstract][Full Text] [Related]
4. Scaling Up Q-Learning via Exploiting State-Action Equivalence. Lyu Y; Côme A; Zhang Y; Talebi MS Entropy (Basel); 2023 Mar; 25(4):. PubMed ID: 37190372 [TBL] [Abstract][Full Text] [Related]
5. Optimization of anemia treatment in hemodialysis patients via reinforcement learning. Escandell-Montero P; Chermisi M; Martínez-Martínez JM; Gómez-Sanchis J; Barbieri C; Soria-Olivas E; Mari F; Vila-Francés J; Stopper A; Gatti E; Martín-Guerrero JD Artif Intell Med; 2014 Sep; 62(1):47-60. PubMed ID: 25091172 [TBL] [Abstract][Full Text] [Related]
6. Sample Efficient Deep Reinforcement Learning With Online State Abstraction and Causal Transformer Model Prediction. Lan Y; Xu X; Fang Q; Hao J IEEE Trans Neural Netw Learn Syst; 2024 Nov; 35(11):16574-16588. PubMed ID: 37581972 [TBL] [Abstract][Full Text] [Related]
8. Hierarchical approximate policy iteration with binary-tree state space decomposition. Xu X; Liu C; Yang SX; Hu D IEEE Trans Neural Netw; 2011 Dec; 22(12):1863-77. PubMed ID: 21990333 [TBL] [Abstract][Full Text] [Related]
9. Kernel-based least squares policy iteration for reinforcement learning. Xu X; Hu D; Lu X IEEE Trans Neural Netw; 2007 Jul; 18(4):973-92. PubMed ID: 17668655 [TBL] [Abstract][Full Text] [Related]
11. Sample-Efficient Reinforcement Learning for Linearly-Parameterized MDPs with a Generative Model. Wang B; Yan Y; Fan J Adv Neural Inf Process Syst; 2021 Dec; 34():16671-16685. PubMed ID: 36168331 [TBL] [Abstract][Full Text] [Related]
12. An Experimental Safety Response Mechanism for an Autonomous Moving Robot in a Smart Manufacturing Environment Using Q-Learning Algorithm and Speech Recognition. Kiangala KS; Wang Z Sensors (Basel); 2022 Jan; 22(3):. PubMed ID: 35161688 [TBL] [Abstract][Full Text] [Related]
13. Energy-efficient and damage-recovery slithering gait design for a snake-like robot based on reinforcement learning and inverse reinforcement learning. Bing Z; Lemke C; Cheng L; Huang K; Knoll A Neural Netw; 2020 Sep; 129():323-333. PubMed ID: 32593929 [TBL] [Abstract][Full Text] [Related]
15. Local Alignment of DNA Sequence Based on Deep Reinforcement Learning. Song YJ; Cho DH IEEE Open J Eng Med Biol; 2021; 2():170-178. PubMed ID: 35402982 [No Abstract] [Full Text] [Related]
16. On Practical Robust Reinforcement Learning: Adjacent Uncertainty Set and Double-Agent Algorithm. Hwang U; Hong S IEEE Trans Neural Netw Learn Syst; 2024 Apr; PP():. PubMed ID: 38619960 [TBL] [Abstract][Full Text] [Related]
17. Self-learning adaptive traffic signal control for real-time safety optimization. Essa M; Sayed T Accid Anal Prev; 2020 Oct; 146():105713. PubMed ID: 32823035 [TBL] [Abstract][Full Text] [Related]
18. MOO-MDP: An Object-Oriented Representation for Cooperative Multiagent Reinforcement Learning. Da Silva FL; Glatt R; Costa AHR IEEE Trans Cybern; 2019 Feb; 49(2):567-579. PubMed ID: 29990289 [TBL] [Abstract][Full Text] [Related]
19. Online learning of shaping rewards in reinforcement learning. Grześ M; Kudenko D Neural Netw; 2010 May; 23(4):541-50. PubMed ID: 20116208 [TBL] [Abstract][Full Text] [Related]
20. Benchmarking for Bayesian Reinforcement Learning. Castronovo M; Ernst D; Couëtoux A; Fonteneau R PLoS One; 2016; 11(6):e0157088. PubMed ID: 27304891 [TBL] [Abstract][Full Text] [Related] [Next] [New Search]