Optimal Exploration–Exploitation in a Multi-armed Bandit Problem with Non-stationary Rewards (Q5113912): Difference between revisions

From MaRDI portal
Importer (talk | contribs)
Created a new Item
 
Normalize DOI.
 
(7 intermediate revisions by 7 users not shown)
Property / DOI
 
Property / DOI: 10.1287/stsy.2019.0033 / rank
Normal rank
 
Property / describes a project that uses
 
Property / describes a project that uses: AdaBoost.MH / rank
 
Normal rank
Property / MaRDI profile type
 
Property / MaRDI profile type: MaRDI publication profile / rank
 
Normal rank
Property / OpenAlex ID
 
Property / OpenAlex ID: W2962821829 / rank
 
Normal rank
Property / arXiv ID
 
Property / arXiv ID: 1405.3316 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Finite-time analysis of the multiarmed bandit problem / rank
 
Normal rank
Property / cites work
 
Property / cites work: The Nonstochastic Multiarmed Bandit Problem / rank
 
Normal rank
Property / cites work
 
Property / cites work: Learning and Strategic Pricing / rank
 
Normal rank
Property / cites work
 
Property / cites work: Q3809068 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Restless Bandits, Linear Programming Relaxations, and a Primal-Dual Index Heuristic / rank
 
Normal rank
Property / cites work
 
Property / cites work: Non-Stationary Stochastic Optimization / rank
 
Normal rank
Property / cites work
 
Property / cites work: An analog of the minimax theorem for vector payoffs / rank
 
Normal rank
Property / cites work
 
Property / cites work: Regret Analysis of Stochastic and Nonstochastic Multi-armed Bandit Problems / rank
 
Normal rank
Property / cites work
 
Property / cites work: Dynamic Assortment with Demand Learning for Seasonal Consumer Goods / rank
 
Normal rank
Property / cites work
 
Property / cites work: Prediction, Learning, and Games / rank
 
Normal rank
Property / cites work
 
Property / cites work: Regret in the on-line decision problem / rank
 
Normal rank
Property / cites work
 
Property / cites work: A decision-theoretic generalization of on-line learning and an application to boosting / rank
 
Normal rank
Property / cites work
 
Property / cites work: On Upper-Confidence Bound Policies for Switching Bandit Problems / rank
 
Normal rank
Property / cites work
 
Property / cites work: Q4197923 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Q4692329 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Q4057976 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Q3245635 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Q5396640 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Asymptotically efficient adaptive allocation rules / rank
 
Normal rank
Property / cites work
 
Property / cites work: Regret bounds for restless Markov bandits / rank
 
Normal rank
Property / cites work
 
Property / cites work: Some aspects of the sequential design of experiments / rank
 
Normal rank
Property / cites work
 
Property / cites work: Q2934090 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Arm-acquiring bandits / rank
 
Normal rank
Property / cites work
 
Property / cites work: Q3815845 / rank
 
Normal rank
Property / Wikidata QID
 
Property / Wikidata QID: Q126855665 / rank
 
Normal rank
Property / DOI
 
Property / DOI: 10.1287/STSY.2019.0033 / rank
 
Normal rank
links / mardi / namelinks / mardi / name
 

Latest revision as of 15:58, 30 December 2024

scientific article; zbMATH DE number 7213023
Language Label Description Also known as
English
Optimal Exploration–Exploitation in a Multi-armed Bandit Problem with Non-stationary Rewards
scientific article; zbMATH DE number 7213023

    Statements

    Optimal Exploration–Exploitation in a Multi-armed Bandit Problem with Non-stationary Rewards (English)
    0 references
    0 references
    0 references
    0 references
    18 June 2020
    0 references
    multi-armed bandit
    0 references
    exploration/exploitation
    0 references
    nonstationary
    0 references
    dynamic oracle
    0 references
    minimax regret
    0 references
    dynamic regret
    0 references

    Identifiers