TY - JOUR
T1 - μ-DDRL
T2 - A QoS-Aware Distributed Deep Reinforcement Learning Technique for Service Offloading in Fog Computing Environments
AU - Goudarzi, Mohammad
AU - Rodriguez, Maria A.
AU - Sarvi, Majid
AU - Buyya, Rajkumar
N1 - Publisher Copyright:
© 2008-2012 IEEE.
PY - 2024/1
Y1 - 2024/1
N2 - Fog and Edge computing extend cloud services to the proximity of end users, allowing many Internet of Things (IoT) use cases, particularly latency-critical applications. Smart devices, such as traffic and surveillance cameras, often do not have sufficient resources to process computation-intensive and latency-critical services. Hence, the constituent parts of services can be offloaded to nearby Edge/Fog resources for processing and storage. However, making offloading decisions for complex services in highly stochastic and dynamic environments is an important, yet difficult task. Recently, Deep Reinforcement Learning (DRL) has been used in many complex service offloading problems; however, existing techniques are most suitable for centralized environments, and their convergence to the best-suitable solutions is slow. In addition, constituent parts of services often have predefined data dependencies and quality of service constraints, which further intensify the complexity of service offloading. To solve these issues, we propose a distributed DRL technique following the actor-critic architecture based on Asynchronous Proximal Policy Optimization (APPO) to achieve efficient and diverse distributed experience trajectory generation. Also, we employ PPO clipping and V-trace techniques for off-policy correction for faster convergence to the most suitable service offloading solutions. The results obtained demonstrate that our technique converges quickly, offers high scalability and adaptability, and outperforms its counterparts by improving the execution time of heterogeneous services.
AB - Fog and Edge computing extend cloud services to the proximity of end users, allowing many Internet of Things (IoT) use cases, particularly latency-critical applications. Smart devices, such as traffic and surveillance cameras, often do not have sufficient resources to process computation-intensive and latency-critical services. Hence, the constituent parts of services can be offloaded to nearby Edge/Fog resources for processing and storage. However, making offloading decisions for complex services in highly stochastic and dynamic environments is an important, yet difficult task. Recently, Deep Reinforcement Learning (DRL) has been used in many complex service offloading problems; however, existing techniques are most suitable for centralized environments, and their convergence to the best-suitable solutions is slow. In addition, constituent parts of services often have predefined data dependencies and quality of service constraints, which further intensify the complexity of service offloading. To solve these issues, we propose a distributed DRL technique following the actor-critic architecture based on Asynchronous Proximal Policy Optimization (APPO) to achieve efficient and diverse distributed experience trajectory generation. Also, we employ PPO clipping and V-trace techniques for off-policy correction for faster convergence to the most suitable service offloading solutions. The results obtained demonstrate that our technique converges quickly, offers high scalability and adaptability, and outperforms its counterparts by improving the execution time of heterogeneous services.
KW - Deep Reinforcement Learning (DRL)
KW - Fog/Edge Computing
KW - Internet of Things (IoT)
KW - QoS- aware service Offloading
UR - http://www.scopus.com/inward/record.url?scp=85177088728&partnerID=8YFLogxK
U2 - 10.1109/TSC.2023.3332308
DO - 10.1109/TSC.2023.3332308
M3 - Article
AN - SCOPUS:85177088728
SN - 1939-1374
VL - 17
SP - 47
EP - 59
JO - IEEE Transactions on Services Computing
JF - IEEE Transactions on Services Computing
IS - 1
ER -