Deep reinforcement learning to enhance the energy-efficient performance of UAV-enabled F-RAN

Fog radio access network (F-RAN) is suitable for Internet of things applications of national important industries, such as pipeline network monitoring in wide area.However, the performance of the F-RAN based on the territorial fog access point will be affected greatly by the complicated territorial...

Full description

Saved in:
Bibliographic Details
Main Authors: Haibo MEI, Kun YANG, Xinyu FAN
Format: Article
Language:zho
Published: China InfoCom Media Group 2021-06-01
Series:物联网学报
Subjects:
Online Access:http://www.wlwxb.com.cn/zh/article/doi/10.11959/j.issn.2096-3750.2021.00234/
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1841533800502263808
author Haibo MEI
Kun YANG
Xinyu FAN
author_facet Haibo MEI
Kun YANG
Xinyu FAN
author_sort Haibo MEI
collection DOAJ
description Fog radio access network (F-RAN) is suitable for Internet of things applications of national important industries, such as pipeline network monitoring in wide area.However, the performance of the F-RAN based on the territorial fog access point will be affected greatly by the complicated territorial environment.This causes F-RAN not able to provide fog access service in a timely and effectively manner.To this problem, the research was proposed to utilize low altitude UAV as the fog access point to realize air ground edge communication and fog computing, which has attracted enormous research interests.How to use deep reinforcement learning (DRL) to improve the energy efficiency of UAV fog access point and extend the mission time of UAV were discussed.Deep reinforcement learning can ensure the UAV fog access point to adjust the configuration strategy timely of air ground communication and computing, including resource optimization, dynamic task offloading and caching.DRL can also optimize the UAV trajectory in 3-D space, and improve the overall performance of UAV enabled fog access network.The innovation of the research lies in the comprehensive discussion of the main optimization problems to be solved in the UAV-enabled F-RAN using DRL.The technical details were also summarized to solve the related optimization problems.Finally, the technical challenges and future research directions of the application of DRL in the UAV-enabled F-RAN were discussed.
format Article
id doaj-art-38fbee53fd4b447185b6d1ce2fcca248
institution Kabale University
issn 2096-3750
language zho
publishDate 2021-06-01
publisher China InfoCom Media Group
record_format Article
series 物联网学报
spelling doaj-art-38fbee53fd4b447185b6d1ce2fcca2482025-01-15T02:53:38ZzhoChina InfoCom Media Group物联网学报2096-37502021-06-015485959649942Deep reinforcement learning to enhance the energy-efficient performance of UAV-enabled F-RANHaibo MEIKun YANGXinyu FANFog radio access network (F-RAN) is suitable for Internet of things applications of national important industries, such as pipeline network monitoring in wide area.However, the performance of the F-RAN based on the territorial fog access point will be affected greatly by the complicated territorial environment.This causes F-RAN not able to provide fog access service in a timely and effectively manner.To this problem, the research was proposed to utilize low altitude UAV as the fog access point to realize air ground edge communication and fog computing, which has attracted enormous research interests.How to use deep reinforcement learning (DRL) to improve the energy efficiency of UAV fog access point and extend the mission time of UAV were discussed.Deep reinforcement learning can ensure the UAV fog access point to adjust the configuration strategy timely of air ground communication and computing, including resource optimization, dynamic task offloading and caching.DRL can also optimize the UAV trajectory in 3-D space, and improve the overall performance of UAV enabled fog access network.The innovation of the research lies in the comprehensive discussion of the main optimization problems to be solved in the UAV-enabled F-RAN using DRL.The technical details were also summarized to solve the related optimization problems.Finally, the technical challenges and future research directions of the application of DRL in the UAV-enabled F-RAN were discussed.http://www.wlwxb.com.cn/zh/article/doi/10.11959/j.issn.2096-3750.2021.00234/unmanned aerial vehiclefog radio access networkdeep reinforcement learningtrajectory designnetwork configuration
spellingShingle Haibo MEI
Kun YANG
Xinyu FAN
Deep reinforcement learning to enhance the energy-efficient performance of UAV-enabled F-RAN
物联网学报
unmanned aerial vehicle
fog radio access network
deep reinforcement learning
trajectory design
network configuration
title Deep reinforcement learning to enhance the energy-efficient performance of UAV-enabled F-RAN
title_full Deep reinforcement learning to enhance the energy-efficient performance of UAV-enabled F-RAN
title_fullStr Deep reinforcement learning to enhance the energy-efficient performance of UAV-enabled F-RAN
title_full_unstemmed Deep reinforcement learning to enhance the energy-efficient performance of UAV-enabled F-RAN
title_short Deep reinforcement learning to enhance the energy-efficient performance of UAV-enabled F-RAN
title_sort deep reinforcement learning to enhance the energy efficient performance of uav enabled f ran
topic unmanned aerial vehicle
fog radio access network
deep reinforcement learning
trajectory design
network configuration
url http://www.wlwxb.com.cn/zh/article/doi/10.11959/j.issn.2096-3750.2021.00234/
work_keys_str_mv AT haibomei deepreinforcementlearningtoenhancetheenergyefficientperformanceofuavenabledfran
AT kunyang deepreinforcementlearningtoenhancetheenergyefficientperformanceofuavenabledfran
AT xinyufan deepreinforcementlearningtoenhancetheenergyefficientperformanceofuavenabledfran