Undiscounted Bandit Games

 0 Người đánh giá. Xếp hạng trung bình 0

Tác giả: Godfrey Keller, Sven Rady

Ngôn ngữ: eng

Ký hiệu phân loại: 649.55 Play with toys

Thông tin xuất bản: 2019

Mô tả vật lý:

Bộ sưu tập: Metadata

ID: 163424

 Comment: 31 pages, 5 figures
  no changes relative to previous version
  full bibliographic detailsWe analyze undiscounted continuous-time games of strategic experimentation with two-armed bandits. The risky arm generates payoffs according to a L\'{e}vy process with an unknown average payoff per unit of time which nature draws from an arbitrary finite set. Observing all actions and realized payoffs, plus a free background signal, players use Markov strategies with the common posterior belief about the unknown parameter as the state variable. We show that the unique symmetric Markov perfect equilibrium can be computed in a simple closed form involving only the payoff of the safe arm, the expected current payoff of the risky arm, and the expected full-information payoff, given the current belief. In particular, the equilibrium does not depend on the precise specification of the payoff-generating processes.
Tạo bộ sưu tập với mã QR

THƯ VIỆN - TRƯỜNG ĐẠI HỌC CÔNG NGHỆ TP.HCM

ĐT: (028) 36225755 | Email: tt.thuvien@hutech.edu.vn

Copyright @2024 THƯ VIỆN HUTECH