An Edge Server Placement Method Based on Reinforcement Learning |
| |
Authors: | Fei Luo Shuai Zheng Weichao Ding Joel Fuentes Yong Li |
| |
Affiliation: | 1.School of Information Science and Engineering, East China University of Science and Technology, Shanghai 200237, China; (F.L.); (S.Z.);2.Department of Computer Science and Information Technologies, Universidad del Bío-Bío, Chillán 3780000, Chile; |
| |
Abstract: | In mobile edge computing systems, the edge server placement problem is mainly tackled as a multi-objective optimization problem and solved with mixed integer programming, heuristic or meta-heuristic algorithms, etc. These methods, however, have profound defect implications such as poor scalability, local optimal solutions, and parameter tuning difficulties. To overcome these defects, we propose a novel edge server placement algorithm based on deep q-network and reinforcement learning, dubbed DQN-ESPA, which can achieve optimal placements without relying on previous placement experience. In DQN-ESPA, the edge server placement problem is modeled as a Markov decision process, which is formalized with the state space, action space and reward function, and it is subsequently solved using a reinforcement learning algorithm. Experimental results using real datasets from Shanghai Telecom show that DQN-ESPA outperforms state-of-the-art algorithms such as simulated annealing placement algorithm (SAPA), Top-K placement algorithm (TKPA), K-Means placement algorithm (KMPA), and random placement algorithm (RPA). In particular, with a comprehensive consideration of access delay and workload balance, DQN-ESPA achieves up to 13.40% and 15.54% better placement performance for 100 and 300 edge servers respectively. |
| |
Keywords: | edge computing markov decision process reinforcement learning access delay workload balance |
|
|