Decision-making in the face of uncertainty is a significant challenge in machine learning, and the multi-armed bandit model is a commonly used framework to address it. This comprehensive and rigorous introduction to the multi-armed bandit problem examines all the major settings, including stochastic, adversarial, and Bayesian frameworks. A focus on both mathematical intuition and carefully worked proofs makes this an excellent reference for established researchers and a helpful resource for graduate students in computer ...
Read More
Decision-making in the face of uncertainty is a significant challenge in machine learning, and the multi-armed bandit model is a commonly used framework to address it. This comprehensive and rigorous introduction to the multi-armed bandit problem examines all the major settings, including stochastic, adversarial, and Bayesian frameworks. A focus on both mathematical intuition and carefully worked proofs makes this an excellent reference for established researchers and a helpful resource for graduate students in computer science, engineering, statistics, applied mathematics and economics. Linear bandits receive special attention as one of the most useful models in applications, while other chapters are dedicated to combinatorial bandits, ranking, non-stationary problems, Thompson sampling and pure exploration. The book ends with a peek into the world beyond bandits with an introduction to partial monitoring and learning in Markov decision processes.
Read Less
Add this copy of Bandit Algorithms to cart. $52.44, new condition, Sold by Ria Christie Books rated 5.0 out of 5 stars, ships from Uxbridge, MIDDLESEX, UNITED KINGDOM, published 2020 by Cambridge University Press.
Add this copy of Bandit Algorithms to cart. $52.43, new condition, Sold by GreatBookPricesUK5 rated 4.0 out of 5 stars, ships from Castle Donington, DERBYSHIRE, UNITED KINGDOM, published 2020 by Cambridge University Press.
Choose your shipping method in Checkout. Costs may vary based on destination.
Seller's Description:
New. Contains: Unspecified. Includes unspecified. Intended for professional and scholarly audience. In Stock. 100% Money Back Guarantee. Brand New, Perfect Condition, allow 4-14 business days for standard shipping. To Alaska, Hawaii, U.S. protectorate, P.O. box, and APO/FPO addresses allow 4-28 business days for Standard shipping. No expedited shipping. All orders placed with expedited shipping will be cancelled. Over 3, 000, 000 happy customers.
Add this copy of Bandit Algorithms to cart. $61.32, new condition, Sold by Booksplease rated 3.0 out of 5 stars, ships from Southport, MERSEYSIDE, UNITED KINGDOM, published 2020 by Cambridge University Press.
Add this copy of Bandit Algorithms to cart. $55.23, new condition, Sold by Ingram Customer Returns Center rated 5.0 out of 5 stars, ships from NV, USA, published 2020 by Cambridge University Press.
Add this copy of Bandit Algorithms to cart. $68.54, new condition, Sold by GreatBookPrices rated 4.0 out of 5 stars, ships from Columbia, MD, UNITED STATES, published 2020 by Cambridge University Press.
Choose your shipping method in Checkout. Costs may vary based on destination.
Seller's Description:
New. Contains: Unspecified. Includes unspecified. Intended for professional and scholarly audience. In Stock. 100% Money Back Guarantee. Brand New, Perfect Condition, allow 4-14 business days for standard shipping. To Alaska, Hawaii, U.S. protectorate, P.O. box, and APO/FPO addresses allow 4-28 business days for Standard shipping. No expedited shipping. All orders placed with expedited shipping will be cancelled. Over 3, 000, 000 happy customers.
Add this copy of Bandit Algorithms (Hardback Or Cased Book) to cart. $68.55, new condition, Sold by BargainBookStores rated 4.0 out of 5 stars, ships from Grand Rapids, MI, UNITED STATES, published 2020 by Cambridge University Press.