Cookies on this website

We use cookies to ensure that we give you the best experience on our website. If you click 'Accept all cookies' we'll assume that you are happy to receive all cookies and you won't see this message again. If you click 'Reject all non-essential cookies' only necessary cookies providing core functionality such as security, network management, and accessibility will be enabled. Click 'Find out more' for information on how to change your cookie settings.

ABSTRACT

 

In natural environments, animals must effectively allocate their choices across many concurrently available resources when foraging. This is a complex decision-making process not fully captured by existing models. In this talk I will describe a novel paradigm that we developed, in which untrained, water-restricted mice were free to sample from six options rewarded at a range of deterministic intervals and positioned around the walls of a large (~2m) arena. Mice exhibited rapid learning, matching their choices to integrated reward ratios across six options within the first session. To develop a mechanistic description of this learning, we constructed a reinforcement learning model inspired by foraging theory. In combination with a dynamic, global (across all options) learning rate, this model was able to accurately reproduce mouse learning and decision-making. Finally, I will discuss results of fiber photometry recordings in the nucleus accumbens core (NAcC) and dorsomedial striatum (DMS). We found that NAcC, but not DMS, dopamine more closely reflected this global learning rate than local error-based updating. Altogether, our results provide insight into the neural substrate of a learning algorithm that allows mice to rapidly exploit multiple options when foraging in large spatial environments.