These tools will no longer be maintained as of December 31, 2024. Archived website can be found here. PubMed4Hh GitHub repository can be found here. Contact NLM Customer Service if you have questions.
7. Folic acid supplementation and malaria susceptibility and severity among people taking antifolate antimalarial drugs in endemic areas. Crider K; Williams J; Qi YP; Gutman J; Yeung L; Mai C; Finkelstain J; Mehta S; Pons-Duran C; Menéndez C; Moraleda C; Rogers L; Daniels K; Green P Cochrane Database Syst Rev; 2022 Feb; 2(2022):. PubMed ID: 36321557 [TBL] [Abstract][Full Text] [Related]
8. Learning-Based DoS Attack Power Allocation in Multiprocess Systems. Huang M; Ding K; Dey S; Li Y; Shi L IEEE Trans Neural Netw Learn Syst; 2023 Oct; 34(10):8017-8030. PubMed ID: 35167483 [TBL] [Abstract][Full Text] [Related]
9. On the Q statistic with constant weights for standardized mean difference. Bakbergenuly I; Hoaglin DC; Kulinskaya E Br J Math Stat Psychol; 2022 Nov; 75(3):444-465. PubMed ID: 35094381 [TBL] [Abstract][Full Text] [Related]
10. Improving Exploration in Actor-Critic With Weakly Pessimistic Value Estimation and Optimistic Policy Optimization. Li F; Fu M; Chen W; Zhang F; Zhang H; Qu H; Yi Z IEEE Trans Neural Netw Learn Syst; 2024 Jul; 35(7):8783-8796. PubMed ID: 36306289 [TBL] [Abstract][Full Text] [Related]
11. Reduction of bias in estimating the frequency of recessive genes. Huether CA; Murphy EA Am J Hum Genet; 1980 Mar; 32(2):212-22. PubMed ID: 7386457 [TBL] [Abstract][Full Text] [Related]
12. Scaling Up Q-Learning via Exploiting State-Action Equivalence. Lyu Y; Côme A; Zhang Y; Talebi MS Entropy (Basel); 2023 Mar; 25(4):. PubMed ID: 37190372 [TBL] [Abstract][Full Text] [Related]
13. Value Iteration Networks with Double Estimator for Planetary Rover Path Planning. Jin X; Lan W; Wang T; Yu P Sensors (Basel); 2021 Dec; 21(24):. PubMed ID: 34960508 [TBL] [Abstract][Full Text] [Related]
14. Human-in-the-Loop Reinforcement Learning in Continuous-Action Space. Luo B; Wu Z; Zhou F; Wang BC IEEE Trans Neural Netw Learn Syst; 2024 Nov; 35(11):15735-15744. PubMed ID: 37418406 [TBL] [Abstract][Full Text] [Related]
15. A Maximum Divergence Approach to Optimal Policy in Deep Reinforcement Learning. Yang Z; Qu H; Fu M; Hu W; Zhao Y IEEE Trans Cybern; 2023 Mar; 53(3):1499-1510. PubMed ID: 34478393 [TBL] [Abstract][Full Text] [Related]
16. Realistic Actor-Critic: A framework for balance between value overestimation and underestimation. Li S; Tang Q; Pang Y; Ma X; Wang G Front Neurorobot; 2022; 16():1081242. PubMed ID: 36699950 [TBL] [Abstract][Full Text] [Related]
17. Parameterized MDPs and Reinforcement Learning Problems-A Maximum Entropy Principle-Based Framework. Srivastava A; Salapaka SM IEEE Trans Cybern; 2022 Sep; 52(9):9339-9351. PubMed ID: 34406959 [TBL] [Abstract][Full Text] [Related]
18. Approximate Policy-Based Accelerated Deep Reinforcement Learning. Wang X; Gu Y; Cheng Y; Liu A; Chen CLP IEEE Trans Neural Netw Learn Syst; 2020 Jun; 31(6):1820-1830. PubMed ID: 31398131 [TBL] [Abstract][Full Text] [Related]
19. Q-ADER: An Effective Q-Learning for Recommendation With Diminishing Action Space. Li F; Qu H; Zhang L; Fu M; Chen W; Yi Z IEEE Trans Neural Netw Learn Syst; 2024 Jul; PP():. PubMed ID: 39012739 [TBL] [Abstract][Full Text] [Related]
20. Context transfer in reinforcement learning using action-value functions. Mousavi A; Nadjar Araabi B; Nili Ahmadabadi M Comput Intell Neurosci; 2014; 2014():428567. PubMed ID: 25610457 [TBL] [Abstract][Full Text] [Related] [Next] [New Search]