Volume 84, Issue 3 pp. 581-617

LINEAR-NONLINEAR-POISSON MODELS OF PRIMATE CHOICE DYNAMICS

Greg S. Corrado

Corresponding Author

Greg S. Corrado

HOWARD HUGHES MEDICAL INSTITUTE, STANFORD UNIVERSITY SCHOOL OF MEDICINE, AND MASSACHUSETTS INSTITUTE OF TECHNOLOGY

Howard Hughes Medical Institute and Stanford University School of Medicine.

Department of Neurobiology, Stanford University, D200 Fairchild Building, 299 Campus Drive West, Stanford, California 94309 (e-mail: [email protected]).Search for more papers by this author
Leo P. Sugrue

Leo P. Sugrue

HOWARD HUGHES MEDICAL INSTITUTE, STANFORD UNIVERSITY SCHOOL OF MEDICINE, AND MASSACHUSETTS INSTITUTE OF TECHNOLOGY

Howard Hughes Medical Institute and Stanford University School of Medicine.

Search for more papers by this author
H. Sebastian Seung

H. Sebastian Seung

HOWARD HUGHES MEDICAL INSTITUTE, STANFORD UNIVERSITY SCHOOL OF MEDICINE, AND MASSACHUSETTS INSTITUTE OF TECHNOLOGY

Howard Hughes Medical Institute and Massachusetts Institute of Technology.

Search for more papers by this author
William T. Newsome

William T. Newsome

HOWARD HUGHES MEDICAL INSTITUTE, STANFORD UNIVERSITY SCHOOL OF MEDICINE, AND MASSACHUSETTS INSTITUTE OF TECHNOLOGY

Howard Hughes Medical Institute and Stanford University School of Medicine.

Search for more papers by this author
First published: 26 February 2013
Citations: 128

Abstract

The equilibrium phenomenon of matching behavior traditionally has been studied in stationary environments. Here we attempt to uncover the local mechanism of choice that gives rise to matching by studying behavior in a highly dynamic foraging environment. In our experiments, 2 rhesus monkeys (Macacca mulatta) foraged for juice rewards by making eye movements to one of two colored icons presented on a computer monitor, each rewarded on dynamic variable-interval schedules. Using a generalization of Wiener kernel analysis, we recover a compact mechanistic description of the impact of past reward on future choice in the form of a Linear-Nonlinear-Poisson model. We validate this model through rigorous predictive and generative testing. Compared to our earlier work with this same data set, this model proves to be a better description of choice behavior and is more tightly correlated with putative neural value signals. Refinements over previous models include hyperbolic (as opposed to exponential) temporal discounting of past rewards, and differential (as opposed to fractional) comparisons of option value. Through numerical simulation we find that within this class of strategies, the model parameters employed by animals are very close to those that maximize reward harvesting efficiency.

The full text of this article hosted at iucr.org is unavailable due to technical difficulties.