IIIT-Delhi Institutional Repository

Browsing by Subject "Upper confidence Bound"

Browsing by Subject "Upper confidence Bound"

Sort by: Order: Results:

  • Pareek, Animesh; Darak, Sumit Jagdish (Advisor) (IIIT-Delhi, 2023-11-29)
    The Multi Armed Bandit (MAB) problem in the Reinforcement Learning field refers to the problem of allocating resources in certain choices for achieving an overall goal with the objective of maximizing expected gain. Our ...

Search Repository

Browse

My Account