Playtesting: What is Beyond Personas


Ariyurek S., Sürer E., Betin Can A.

IEEE TRANSACTIONS ON GAMES, vol.15, no.3, pp.348-359, 2023 (SCI-Expanded) identifier identifier

  • Publication Type: Article / Article
  • Volume: 15 Issue: 3
  • Publication Date: 2023
  • Doi Number: 10.1109/tg.2022.3165882
  • Journal Name: IEEE TRANSACTIONS ON GAMES
  • Journal Indexes: Science Citation Index Expanded (SCI-EXPANDED)
  • Page Numbers: pp.348-359
  • Keywords: Automated playtesting, play persona, player modeling, reinforcement learning (RL)
  • Middle East Technical University Affiliated: Yes

Abstract

Playtesting is an essential step in the game design process. Game designers use the feedback from playtests to refine their designs. Game designers may employ procedural personas to automate the playtesting process. In this article, we present two approaches to improve automated playtesting. First, we propose developing persona, which allows a persona to progress to different goals. In contrast, the procedural persona is fixed to a single goal. Second, a human playtester knows which paths she has tested before, and during the consequent tests, she may test different paths. However, reinforcement learning (RL) agents disregard these previous paths. We propose a novel methodology that we refer to as alternative path finder (APF). We train APF with previous paths and employ APF during the training of an RL agent. APF modulates the reward structure of the environment, while preserving the agent's goal. When evaluated, the agent generates a different trajectory that achieves the same goal. We use the general video game artificial intelligence and VizDoom frameworks to test our proposed methodologies. We use proximal policy optimization RL agent during experiments. First, we compare the playtest data generated by developing and procedural persona. Our experiments show that developing persona provides better insight into the game and how different players would play. Second, we present the alternative paths found using APF and argue why traditional RL agents cannot learn those paths.