These tools will no longer be maintained as of December 31, 2024. Archived website can be found here. PubMed4Hh GitHub repository can be found here. Contact NLM Customer Service if you have questions.


BIOMARKERS

Molecular Biopsy of Human Tumors

- a resource for Precision Medicine *

126 related articles for article (PubMed ID: 38983756)

  • 1. Actor-Critic Alignment for Offline-to-Online Reinforcement Learning.
    Yu Z; Zhang X
    Proc Mach Learn Res; 2023 Jul; 202():40452-40474. PubMed ID: 38983756
    [TBL] [Abstract][Full Text] [Related]  

  • 2. Mild Policy Evaluation for Offline Actor-Critic.
    Huang L; Dong B; Lu J; Zhang W
    IEEE Trans Neural Netw Learn Syst; 2023 Sep; PP():. PubMed ID: 37676802
    [TBL] [Abstract][Full Text] [Related]  

  • 3. Offline Reinforcement Learning With Behavior Value Regularization.
    Huang L; Dong B; Xie W; Zhang W
    IEEE Trans Cybern; 2024 Jun; 54(6):3692-3704. PubMed ID: 38669164
    [TBL] [Abstract][Full Text] [Related]  

  • 4. Meta attention for Off-Policy Actor-Critic.
    Huang J; Huang W; Lan L; Wu D
    Neural Netw; 2023 Jun; 163():86-96. PubMed ID: 37030278
    [TBL] [Abstract][Full Text] [Related]  

  • 5. The Actor-Dueling-Critic Method for Reinforcement Learning.
    Wu M; Gao Y; Jung A; Zhang Q; Du S
    Sensors (Basel); 2019 Mar; 19(7):. PubMed ID: 30935035
    [TBL] [Abstract][Full Text] [Related]  

  • 6. De-Pessimism Offline Reinforcement Learning via Value Compensation.
    Huang Z; Zhao J; Sun S
    IEEE Trans Neural Netw Learn Syst; 2024 Aug; PP():. PubMed ID: 39178073
    [TBL] [Abstract][Full Text] [Related]  

  • 7. A priority experience replay actor-critic algorithm using self-attention mechanism for strategy optimization of discrete problems.
    Sun Y; Yang B
    PeerJ Comput Sci; 2024; 10():e2161. PubMed ID: 38983226
    [TBL] [Abstract][Full Text] [Related]  

  • 8. Improving Exploration in Actor-Critic With Weakly Pessimistic Value Estimation and Optimistic Policy Optimization.
    Li F; Fu M; Chen W; Zhang F; Zhang H; Qu H; Yi Z
    IEEE Trans Neural Netw Learn Syst; 2024 Jul; 35(7):8783-8796. PubMed ID: 36306289
    [TBL] [Abstract][Full Text] [Related]  

  • 9. Robust Actor-Critic With Relative Entropy Regulating Actor.
    Cheng Y; Huang L; Chen CLP; Wang X
    IEEE Trans Neural Netw Learn Syst; 2023 Nov; 34(11):9054-9063. PubMed ID: 35286268
    [TBL] [Abstract][Full Text] [Related]  

  • 10. Offline Reward Perturbation Boosts Distributional Shift in Online RL.
    Yu Z; Kang S; Zhang X
    Uncertain Artif Intell; 2024 Jul; 2024():. PubMed ID: 39006853
    [TBL] [Abstract][Full Text] [Related]  

  • 11. Distributional Soft Actor-Critic: Off-Policy Reinforcement Learning for Addressing Value Estimation Errors.
    Duan J; Guan Y; Li SE; Ren Y; Sun Q; Cheng B
    IEEE Trans Neural Netw Learn Syst; 2022 Nov; 33(11):6584-6598. PubMed ID: 34101599
    [TBL] [Abstract][Full Text] [Related]  

  • 12. Stochastic Integrated Actor-Critic for Deep Reinforcement Learning.
    Zheng J; Kurt MN; Wang X
    IEEE Trans Neural Netw Learn Syst; 2024 May; 35(5):6654-6666. PubMed ID: 36256721
    [TBL] [Abstract][Full Text] [Related]  

  • 13. Adaptive bias-variance trade-off in advantage estimator for actor-critic algorithms.
    Chen Y; Zhang F; Liu Z
    Neural Netw; 2024 Jan; 169():764-777. PubMed ID: 37981458
    [TBL] [Abstract][Full Text] [Related]  

  • 14. An actor-critic framework based on deep reinforcement learning for addressing flexible job shop scheduling problems.
    Zhao C; Deng N
    Math Biosci Eng; 2024 Jan; 21(1):1445-1471. PubMed ID: 38303472
    [TBL] [Abstract][Full Text] [Related]  

  • 15. SMONAC: Supervised Multiobjective Negative Actor-Critic for Sequential Recommendation.
    Zhou F; Luo B; Wu Z; Huang T
    IEEE Trans Neural Netw Learn Syst; 2023 Oct; PP():. PubMed ID: 37788188
    [TBL] [Abstract][Full Text] [Related]  

  • 16. Improving Offline Reinforcement Learning With In-Sample Advantage Regularization for Robot Manipulation.
    Ma C; Yang D; Wu T; Liu Z; Yang H; Chen X; Lan X; Zheng N
    IEEE Trans Neural Netw Learn Syst; 2024 Sep; PP():. PubMed ID: 39302799
    [TBL] [Abstract][Full Text] [Related]  

  • 17. Representation Learning and Reinforcement Learning for Dynamic Complex Motion Planning System.
    Zhou C; Huang B; Franti P
    IEEE Trans Neural Netw Learn Syst; 2024 Aug; 35(8):11049-11063. PubMed ID: 37028017
    [TBL] [Abstract][Full Text] [Related]  

  • 18. Boosting On-Policy Actor-Critic With Shallow Updates in Critic.
    Li L; Zhu Y
    IEEE Trans Neural Netw Learn Syst; 2024 Apr; PP():. PubMed ID: 38619961
    [TBL] [Abstract][Full Text] [Related]  

  • 19. Offline Model-Based Adaptable Policy Learning for Decision-Making in Out-of-Support Regions.
    Chen XH; Luo FM; Yu Y; Li Q; Qin Z; Shang W; Ye J
    IEEE Trans Pattern Anal Mach Intell; 2023 Dec; 45(12):15260-15274. PubMed ID: 37725727
    [TBL] [Abstract][Full Text] [Related]  

  • 20. Reinforcement learning in continuous time and space.
    Doya K
    Neural Comput; 2000 Jan; 12(1):219-45. PubMed ID: 10636940
    [TBL] [Abstract][Full Text] [Related]  

    [Next]    [New Search]
    of 7.