TY - JOUR
T1 - Investigating Transfer Learning in Noisy Environments
T2 - A Study of Predecessor and Successor Features in Spatial Learning Using a T-Maze
AU - Seo, Incheol
AU - Lee, Hyunsu
N1 - Publisher Copyright:
© 2024 by the authors.
PY - 2024/10
Y1 - 2024/10
N2 - In this study, we investigate the adaptability of artificial agents within a noisy T-maze that use Markov decision processes (MDPs) and successor feature (SF) and predecessor feature (PF) learning algorithms. Our focus is on quantifying how varying the hyperparameters, specifically the reward learning rate ((Formula presented.)) and the eligibility trace decay rate ((Formula presented.)), can enhance their adaptability. Adaptation is evaluated by analyzing the hyperparameters of cumulative reward, step length, adaptation rate, and adaptation step length and the relationships between them using Spearman’s correlation tests and linear regression. Our findings reveal that an (Formula presented.) of 0.9 consistently yields superior adaptation across all metrics at a noise level of 0.05. However, the optimal setting for (Formula presented.) varies by metric and context. In discussing these results, we emphasize the critical role of hyperparameter optimization in refining the performance and transfer learning efficacy of learning algorithms. This research advances our understanding of the functionality of PF and SF algorithms, particularly in navigating the inherent uncertainty of transfer learning tasks. By offering insights into the optimal hyperparameter configurations, this study contributes to the development of more adaptive and robust learning algorithms, paving the way for future explorations in artificial intelligence and neuroscience.
AB - In this study, we investigate the adaptability of artificial agents within a noisy T-maze that use Markov decision processes (MDPs) and successor feature (SF) and predecessor feature (PF) learning algorithms. Our focus is on quantifying how varying the hyperparameters, specifically the reward learning rate ((Formula presented.)) and the eligibility trace decay rate ((Formula presented.)), can enhance their adaptability. Adaptation is evaluated by analyzing the hyperparameters of cumulative reward, step length, adaptation rate, and adaptation step length and the relationships between them using Spearman’s correlation tests and linear regression. Our findings reveal that an (Formula presented.) of 0.9 consistently yields superior adaptation across all metrics at a noise level of 0.05. However, the optimal setting for (Formula presented.) varies by metric and context. In discussing these results, we emphasize the critical role of hyperparameter optimization in refining the performance and transfer learning efficacy of learning algorithms. This research advances our understanding of the functionality of PF and SF algorithms, particularly in navigating the inherent uncertainty of transfer learning tasks. By offering insights into the optimal hyperparameter configurations, this study contributes to the development of more adaptive and robust learning algorithms, paving the way for future explorations in artificial intelligence and neuroscience.
KW - T-maze transfer learning
KW - hyperparameter tuning
KW - noisy observation
KW - predecessor features
KW - reinforcement learning
KW - robustness
UR - http://www.scopus.com/inward/record.url?scp=85206495214&partnerID=8YFLogxK
U2 - 10.3390/s24196419
DO - 10.3390/s24196419
M3 - Article
C2 - 39409459
AN - SCOPUS:85206495214
SN - 1424-8220
VL - 24
JO - Sensors
JF - Sensors
IS - 19
M1 - 6419
ER -