TY - GEN
T1 - Transfer Learning and Curriculum Learning in Sokoban
AU - Yang, Zhao
AU - Preuss, Mike
AU - Plaat, Aske
PY - 2022
Y1 - 2022
N2 - Transfer learning can speed up training in machine learning, and is regularly used in classification tasks. It reuses prior knowledge from other tasks to pre-train networks for new tasks. In reinforcement learning, learning actions for a behavior policy that can be applied to new environments is still a challenge, especially for tasks that involve much planning. Sokoban is a challenging puzzle game. It has been used widely as a benchmark in planning-based reinforcement learning. In this paper, we show how prior knowledge improves learning in Sokoban tasks. We find that reusing feature representations learned previously can accelerate learning new, more complex, instances. In effect, we show how curriculum learning, from simple to complex tasks, works in Sokoban. Furthermore, feature representations learned in simpler instances are more general, and thus lead to positive transfers towards more complex tasks, but not vice versa. We have also studied which part of the knowledge is most important for transfer to succeed, and identify which layers should be used for pre-training (Codes we used for this work can be found at https://github.com/yangzhao-666/TLCLS ).
AB - Transfer learning can speed up training in machine learning, and is regularly used in classification tasks. It reuses prior knowledge from other tasks to pre-train networks for new tasks. In reinforcement learning, learning actions for a behavior policy that can be applied to new environments is still a challenge, especially for tasks that involve much planning. Sokoban is a challenging puzzle game. It has been used widely as a benchmark in planning-based reinforcement learning. In this paper, we show how prior knowledge improves learning in Sokoban tasks. We find that reusing feature representations learned previously can accelerate learning new, more complex, instances. In effect, we show how curriculum learning, from simple to complex tasks, works in Sokoban. Furthermore, feature representations learned in simpler instances are more general, and thus lead to positive transfers towards more complex tasks, but not vice versa. We have also studied which part of the knowledge is most important for transfer to succeed, and identify which layers should be used for pre-training (Codes we used for this work can be found at https://github.com/yangzhao-666/TLCLS ).
UR - http://www.scopus.com/inward/record.url?scp=85123444482&partnerID=8YFLogxK
U2 - 10.1007/978-3-030-93842-0_11
DO - 10.1007/978-3-030-93842-0_11
M3 - Conference contribution
SN - 9783030938413
T3 - Communications in Computer and Information Science
SP - 187
EP - 200
BT - Artificial Intelligence and Machine Learning - 33rd Benelux Conference on Artificial Intelligence, BNAIC/Benelearn 2021, Revised Selected Papers
A2 - Leiva, L.A.
A2 - Pruski, C.
A2 - Markovich, R.
A2 - Najjar, A.
A2 - Schommer, C.
PB - Springer Science and Business Media Deutschland GmbH
T2 - 33rd Benelux Conference on Artificial Intelligence, BNAIC/ BENELEARN 2021
Y2 - 10 November 2021 through 12 November 2021
ER -