IIIT-Delhi Institutional Repository

Browsing by Subject "Early-Exiting"

Browsing by Subject "Early-Exiting"

Sort by: Order: Results:

  • Pareek, Animesh; Darak, Sumit Jagdish (Advisor) (IIIT-Delhi, 2023-11-29)
    The Multi Armed Bandit (MAB) problem in the Reinforcement Learning field refers to the problem of allocating resources in certain choices for achieving an overall goal with the objective of maximizing expected gain. Our ...

Search Repository

Browse

My Account