Analysis and improvement of policy gradient estimation (Q448295): Difference between revisions

From MaRDI portal
Import240304020342 (talk | contribs)
Set profile property.
ReferenceBot (talk | contribs)
Changed an Item
 
(2 intermediate revisions by 2 users not shown)
Property / full work available at URL
 
Property / full work available at URL: https://doi.org/10.1016/j.neunet.2011.09.005 / rank
 
Normal rank
Property / OpenAlex ID
 
Property / OpenAlex ID: W2148053762 / rank
 
Normal rank
Property / Wikidata QID
 
Property / Wikidata QID: Q51513131 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Q4533363 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Using Expectation-Maximization for Reinforcement Learning / rank
 
Normal rank
Property / cites work
 
Property / cites work: Q4692508 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Q3093234 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Q4427427 / rank
 
Normal rank
Property / cites work
 
Property / cites work: 10.1162/1532443041827907 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Q2769922 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Approximate gradient methods in policy-space optimization of Markov reward processes / rank
 
Normal rank
Property / cites work
 
Property / cites work: Simple statistical gradient-following algorithms for connectionist reinforcement learning / rank
 
Normal rank

Latest revision as of 15:37, 5 July 2024

scientific article
Language Label Description Also known as
English
Analysis and improvement of policy gradient estimation
scientific article

    Statements

    Analysis and improvement of policy gradient estimation (English)
    0 references
    0 references
    0 references
    0 references
    0 references
    30 August 2012
    0 references
    0 references
    reinforcement learning
    0 references
    policy gradients
    0 references
    policy gradients with parameter-based exploration
    0 references
    variance reduction
    0 references
    0 references
    0 references