Reinforcement Learning for Energy Optimization with 5G Communications in Vehicular Social Networks
Citations

WEB OF SCIENCE

21
Citations

SCOPUS

27

초록

Increased data traffic resulting from the increase in the deployment of connected vehicles has become relevant in vehicular social networks (VSNs). To provide efficient communication between connected vehicles, researchers have studied device-to-device (D2D) communication. D2D communication not only reduces the energy consumption and loads of the system but also increases the system capacity by reusing cellular resources. However, D2D communication is highly affected by interference and therefore requires interference-management techniques, such as mode selection and power control. To make an optimal mode selection and power control, it is necessary to apply reinforcement learning that considers a variety of factors. In this paper, we propose a reinforcement-learning technique for energy optimization with fifth-generation communication in VSNs. To achieve energy optimization, we use centralized Q-learning in the system and distributed Q-learning in the vehicles. The proposed algorithm learns to maximize the energy efficiency of the system by adjusting the minimum signal-to-interference plus noise ratio to guarantee the outage probability. Simulations were performed to compare the performance of the proposed algorithm with that of the existing mode-selection and power-control algorithms. The proposed algorithm performed the best in terms of system energy efficiency and achievable data rate.

키워드

5GD2D communicationvehicle-to-vehicle communicationmode selectionpower controlvehicular social networkMODE SELECTIOND2D COMMUNICATIONS
제목
Reinforcement Learning for Energy Optimization with 5G Communications in Vehicular Social Networks
저자
Park, HyebinLim, Yujin
DOI
10.3390/s20082361
발행일
2020-04
유형
Article
저널명
Sensors
20
8