Use this URL to cite or link to this record in EThOS:
Title: Individual decision making, reinforcement learning and myopic behaviour
Author: Pastore, Alvin
ISNI:       0000 0004 7964 5142
Awarding Body: University of Sheffield
Current Institution: University of Sheffield
Date of Award: 2019
Availability of Full Text:
Access from EThOS:
Access from Institution:
Individuals use their cognitive abilities to make decisions, with the ultimate goal of improving their status. Decisions outcomes are used to learn the association between the decisions which lead to good results and those resulting in punishing outcomes. These associations might not be easily inferable because of environmental complexity or noisy feedback. Tasks in which outcomes probabilities are known are termed "decisions under risk". Researchers have consistently showed that people are risk averse when choosing among options featuring gains, while they are risk seeking when making decisions about options featuring losses. When the probabilities of the options are not clearly stated the task is known as "decisions under ambiguity". In this type of task individuals face an exploration-exploitation trade off: to maximise their profit they need to choose the best option but at the same time they need to discover which option leads to the best outcome by trial-and-error. The process of knowledge acquisition by interaction with the environment is called adaptive learning. Evidence from literature points in the direction of unskilled investors behaviour being consistent with naive reinforcement learning, simply adjusting their preference for which option to choose based on its recent outcomes. Experimental data from a binary choice task and a quasi-field scenario is used to test a combination of Reinforcement Learning and Prospect Theory. Both the investigations include reinforcement learning models featuring specific parameters which can be tuned to describe individual learning decision-making strategies. The first part is focused on integrating the two computational models, the second on testing it on a more realistic scenario. The results indicate that the combination of Reinforcement Learning and Prospect Theory could be a descriptive account of decision- making in binary decision tasks. A two-state space configuration, together with a non- saturating reward function appears to be the best setup to capture behaviour in said task. Moreover, analysing the parameters of the models it becomes evident that payoff variability has an impact on speed of learning and randomness of choice. The same modelling approach fails to capture behaviour in a more complex task, indicating that more complex models might be needed to provide a computational account of decisions from experience in non-trivial tasks.
Supervisor: Vasilaki, Eleni ; Stafford, Tom ; Marshall, James Sponsor: Not available
Qualification Name: Thesis (Ph.D.) Qualification Level: Doctoral
EThOS ID:  DOI: Not available