Результаты исследований: Научные публикации в периодических изданиях › статья › Рецензирование
Reinforcement learning-based optimal control for stochastic opinion dynamics. / Chen, Yajin; Gao, Hongwei; Mazalov, Vladimir V.; Liu, Yanshan.
в: Scientific Reports, 06.03.2026.Результаты исследований: Научные публикации в периодических изданиях › статья › Рецензирование
}
TY - JOUR
T1 - Reinforcement learning-based optimal control for stochastic opinion dynamics
AU - Chen, Yajin
AU - Gao, Hongwei
AU - Mazalov, Vladimir V.
AU - Liu, Yanshan
PY - 2026/3/6
Y1 - 2026/3/6
N2 - This paper proposes a integrated framework for optimal control of opinion dynamics in social networks, addressing three progressively challenging scenarios: Model-based stochastic control, where agent interactions follow known probability distributions, enabling analytical optimal policies; Model-free Reinforcement Learning (RL), where interaction randomness has unknown distributions but system dynamics are preserved; Data-driven RL for unknown systems, where time-varying network dynamics (with stochasticity constraints) are fully unknown, requiring purely observational learning. By designing an RL control framework grounded in convex quadratic optimization, we bridge model-based control and data-driven learning, offering new insights for social network manipulation and multi-agent coordination. Numerical simulations demonstrate the framework’s effectiveness.
AB - This paper proposes a integrated framework for optimal control of opinion dynamics in social networks, addressing three progressively challenging scenarios: Model-based stochastic control, where agent interactions follow known probability distributions, enabling analytical optimal policies; Model-free Reinforcement Learning (RL), where interaction randomness has unknown distributions but system dynamics are preserved; Data-driven RL for unknown systems, where time-varying network dynamics (with stochasticity constraints) are fully unknown, requiring purely observational learning. By designing an RL control framework grounded in convex quadratic optimization, we bridge model-based control and data-driven learning, offering new insights for social network manipulation and multi-agent coordination. Numerical simulations demonstrate the framework’s effectiveness.
U2 - 10.1038/s41598-026-42646-1
DO - 10.1038/s41598-026-42646-1
M3 - Article
JO - Scientific Reports
JF - Scientific Reports
SN - 2045-2322
ER -
ID: 150015116