Deep Reinforcement Learning in agent-based model AgriPoliS to simulate strategic land market interactions
Saved in:
| Title: | Deep Reinforcement Learning in agent-based model AgriPoliS to simulate strategic land market interactions |
|---|---|
| Authors: | Dong, Changxing, Njiru, Ruth Dionisia Gicuku, Appel, Franziska |
| Publisher Information: | Berlin: Berlin Universities Publishing, 2025. |
| Publication Year: | 2025 |
| Subject Terms: | message queue, Deep Reinforcement Learning, ddc:630, AgriPoliS, policy gradient algorithm, agent-based model |
| Description: | AgriPoliS (Agricultural Policy Simulator) is an Agent-Based Model for simulating the dynamic evolution of agricultural regions with a particular emphasis on the structural transformations influenced by economic, ecological, and societal factors. This study introduces a Reinforcement Learning (RL) framework to empower agents within AgriPoliS with strategic decision-making capabilities, specifically in the context of land market bidding. Traditional agents within AgriPoliS make decisions through Mixed Integer Programming (MIP) optimizations, inherently limited by myopic considerations of the current year's conditions. In this framework one agent within AgriPoliS is defined as the RL agent, with AgriPoliS itself serving as the learning environment. The RL agent's bidding coefficient are formulated as actions, with the state space comprising various agent and regional variables, including liquidity, farm endowments, remaining land, and neighbouring farms. We adopt the cumulative equity capital of the agent at the end of the simulation as the reward from the environment, reflecting a holistic perspective towards stable and sustainable development. Utilizing a policy gradient algorithm, we update the RL agent's policy network to optimize bidding strategies. Training results demonstrate that the RL-enhanced agent consistently outperforms its non-learning counterpart, exhibiting strategically stable bidding behavior. To assess the robustness of the algorithm, extensive training runs are conducted with varying hyperparameters, including policy network initialization, learning rate, and exploration noise levels. Our findings underscore the efficacy of RL in enhancing strategic decision-making in agricultural land markets, showcasing superior performance compared to fixed bidding strategies. |
| Document Type: | Article |
| Language: | English |
| DOI: | 10.14279/eceasst.v83.2621 |
| Access URL: | https://eceasst.org/index.php/eceasst/article/view/2621 https://hdl.handle.net/10419/312031 |
| Rights: | CC BY |
| Accession Number: | edsair.od......1687..85da55d0502f1a10057e4b17fb7ebdf2 |
| Database: | OpenAIRE |
| Abstract: | AgriPoliS (Agricultural Policy Simulator) is an Agent-Based Model for simulating the dynamic evolution of agricultural regions with a particular emphasis on the structural transformations influenced by economic, ecological, and societal factors. This study introduces a Reinforcement Learning (RL) framework to empower agents within AgriPoliS with strategic decision-making capabilities, specifically in the context of land market bidding. Traditional agents within AgriPoliS make decisions through Mixed Integer Programming (MIP) optimizations, inherently limited by myopic considerations of the current year's conditions. In this framework one agent within AgriPoliS is defined as the RL agent, with AgriPoliS itself serving as the learning environment. The RL agent's bidding coefficient are formulated as actions, with the state space comprising various agent and regional variables, including liquidity, farm endowments, remaining land, and neighbouring farms. We adopt the cumulative equity capital of the agent at the end of the simulation as the reward from the environment, reflecting a holistic perspective towards stable and sustainable development. Utilizing a policy gradient algorithm, we update the RL agent's policy network to optimize bidding strategies. Training results demonstrate that the RL-enhanced agent consistently outperforms its non-learning counterpart, exhibiting strategically stable bidding behavior. To assess the robustness of the algorithm, extensive training runs are conducted with varying hyperparameters, including policy network initialization, learning rate, and exploration noise levels. Our findings underscore the efficacy of RL in enhancing strategic decision-making in agricultural land markets, showcasing superior performance compared to fixed bidding strategies. |
|---|---|
| DOI: | 10.14279/eceasst.v83.2621 |
Nájsť tento článok vo Web of Science