Open Access
ARTICLE
Priority-Aware Resource Allocation for VNF Deployment in Service Function Chains Based on Graph Reinforcement Learning
1 Department of Software Convergence, Soonchunhyang University, Asan, 31538, Republic of Korea
2 School of Digital Technologies, American University of Phnom Penh, Phnom Penh, 12106, Cambodia
3 Department of Computer Software Engineering, Soonchunhyang University, Asan, 31538, Republic of Korea
* Corresponding Author: Seokhoon Kim. Email:
# These authors contributed equally to this work
(This article belongs to the Special Issue: Graph Neural Networks: Methods and Applications in Graph-related Problems)
Computers, Materials & Continua 2025, 83(2), 1649-1665. https://doi.org/10.32604/cmc.2025.062716
Received 25 December 2024; Accepted 06 March 2025; Issue published 16 April 2025
Abstract
Recently, Network Functions Virtualization (NFV) has become a critical resource for optimizing capability utilization in the 5G/B5G era. NFV decomposes the network resource paradigm, demonstrating the efficient utilization of Network Functions (NFs) to enable configurable service priorities and resource demands. Telecommunications Service Providers (TSPs) face challenges in network utilization, as the vast amounts of data generated by the Internet of Things (IoT) overwhelm existing infrastructures. IoT applications, which generate massive volumes of diverse data and require real-time communication, contribute to bottlenecks and congestion. In this context, Multi-access Edge Computing (MEC) is employed to support resource and priority-aware IoT applications by implementing Virtual Network Function (VNF) sequences within Service Function Chaining (SFC). This paper proposes the use of Deep Reinforcement Learning (DRL) combined with Graph Neural Networks (GNN) to enhance network processing, performance, and resource pooling capabilities. GNN facilitates feature extraction through Message-Passing Neural Network (MPNN) mechanisms. Together with DRL, Deep Q-Networks (DQN) are utilized to dynamically allocate resources based on IoT network priorities and demands. Our focus is on minimizing delay times for VNF instance execution, ensuring effective resource placement, and allocation in SFC deployments, offering flexibility to adapt to real-time changes in priority and workload. Simulation results demonstrate that our proposed scheme outperforms reference models in terms of reward, delay, delivery, service drop ratios, and average completion ratios, proving its potential for IoT applications.Keywords
The Internet of Things (IoT) has increased rapidly in recent years, as it has birthed an immense chance to transition from a conventional economy to a digital economy. As in an intrinsically distributed data-centric system, IoT can gather and provide an enormous volume of data that can be processed, interpreted, and analyzed into useful information and acknowledgment. Hence, in the context of IoT-based applications, it involves creating huge and vast amounts of data that lead to resource constraints, increasingly vast resource capacity in network performance, and resource congestion in terms of the urgent business on time. However, Software-Defined Networking (SDN) and Network Functions Virtualization (NFV) were proposed to provision treated as the key enabler technologies to address dynamic and agile IoT networks [1–4]. SDN creates a programmable network by centralizing the control plan and data plan functionality from physical routing devices, while NFV decouples various network resources to compose the Network Functions (NFs) from physical network devices, running them as virtualization technologies [5,6]. By adopting these technologies, SDN and NFV-enabled IoT networks considerably simplify the deployment of NFs and improve network performance and service.
Upon closely providing the network resource management and orchestration of all data utilization, Service Function Chaining (SFC) is used to indicate an ordered or sequential collection of Virtual Network Functions (VNFs) to create service functions [6]. SFC propagates ordered NFs from ingress to egress (e.g., VNF1, VNF2, VNF3, and VNF-n); meanwhile, SFC requests must be rounded by specifying the typed and listed VNFs in the specified order. On the other hand, the more appropriate representation is the VNF-Forwarding Graph (VNF-FG), which is proposed by the European Telecommunications Standards Institute (ETSI) [7,8]. In fact, Multi-access Edge Computing (MEC) architecture provides benefits in conducting the proliferation of IoT networks. MEC provides efficient resource utilization and management mechanisms by eliminating computation intensively to enhance the end user of the resource capacities in terms of computation and communication on latency. However, placing and allocating virtual resources in network resource management in MEC servers is essential for responding and adapting to the satisfaction of service derived in complex network environments. Otherwise, VNFI employs resource computing techniques to optimize resource placement and allocation, offering scalable, adaptable, and flexible real-time solutions. However, efficient resource management is impacted by heterogeneous workloads to meet user demands for dynamic and flexible service. Due to limitation of availability of those resource orchestration and computation for diversity applications under MEC architecture, placement, and allocation of resources adaptively lead to many challenges. These challenges arise, including resource constraints, latency, and interoperability issues, during compute-intensive servers across heterogeneous IoT networks on applications. The distributed significant environment of multi-service demands, and multi-vendor applications adds complexity to allocating and placing the resource utilization.
This paper focuses on leveraging intelligent resource allocation and placement based on service priority demands-driven resource management in IoT networks. By integrating Deep Reinforcement Learning (DRL) based on Graph Neural Network (GNN) paradigms which use the graph structure of Message-Passing Neural Network (MPNN) for embedding the nodes and link states of VNF [9–11], GNN solves the scalability problem caused by MPNN variants in VNF-FG size. Moreover, DRL is implemented to determine the policy charging on SDNC for adaptively aligning the optimal decision service provisioning in the network fluctuation on IoT network. Fig. 1 leverages DRL based on GNN to increase near-optimal decision-making in the dynamic service provisioning process. SDN controller captures the traffic flow in distributed IoT networks in routing tables by using OpenFlow protocol for managing the flow entries. Meanwhile, NFVO handles orchestrating and managing the resource instant for computing workload of IoT networks. On the other hand, DRL-GNN framework ensures efficient techniques to align the controller for adaptive self-configuration on VNF placement and resource allocation. Our main contributions are summarized as follows:

Figure 1: Message-passing neural network embedding with SDNC/NFVO architecture
• We propose DRL based on the GNN framework to support service requests simultaneously from MEC resources, maximizing the SFC acceptance ratio while minimizing the MEC workload. By addressing the complexities of resource allocation and VNF placement in SFC modules, our methods ensure priority-driven resource management in IoT networks, optimizing resource utilization based on service priorities. Once, NFVO abstracts the SFC placement in all the resource allocations to address and guarantee various network services’ QoS requirements while offering minimum delay and resource usage.
• Leveraging a DRL-GNN algorithm, namely a priority-aware VNF deployment algorithm (PVD-GDR), to set the threshold outperformance and satisfy IoT applications. The PVD-GDR framework enables the capability to ensure resource preservation on VNF nodes and link embedding, which utilizes a message-passing variant to propagate the feature information to its neighbors until a stable equilibrium is reached optimally. We employ DQN algorithms to provide the capabilities to learn optimal policies in dynamic settings for efficient priority-aware resource orchestration and adaptive resource utilization in SFC deployment in assisted control data flow. Hence, DQN ensures that the update of value function in SFC deployment and make flexibility decision for SFC requests arriving at real-time in difference setting according to the optimal value function.
• Lastly, our scheme shows the performance matric, which compares the reference scheme in terms of reward, end-to-end task execution delay of requested service, delivery ratios of requested service, task drop ratios of requested service, and average completion ratios.
The remainder of the paper is aligned as follows: in Section 2, we illustrate the related work and the current work. Section 3 describes the network system formulation in terms of computation and network resource utilization. Section 4 presents DRL-GNN for service priority on IoT demands. Lastly, the proposed scheme indicates its performance compared to the reference scheme.
2 Preliminary of Priority/Demanding in Resource Utilization
In this section, many studies have investigated resource placement and solved optimization problems. However, efficient allocation and resource placement of the limited resource is the most impact-critical component in MEC and its efficiency for orchestrating resource to satisfaction on different QoS. Hence, minimum delay and resource cost are priorities for the end goal. Additionally, we comprehensively gather previous studies on resource placement in MEC.
2.1 Enabling Resource Utilization Based on DRL
The utilization of resources plays a crucial role in demonstrating the efficiency of resource allocation and computation. SFC deployment directly impacts specific network functions to achieve network service targets. Moreover, compiling SFC formations for multiple sequential applications is a critical area that requires timely processing [12–15]. DRL has been leveraged to address the challenges of SFC placement, particularly in terms of service flexibility and diversity. However, existing DRL approaches struggle with handling a dynamic number of objectives and suffer from unreliable training time consumption. To overcome this, researchers have utilized divided sub-problem weight vectors to establish the relationship between solution positioning and placement. By introducing a related zone decomposition model, the weighted vectors are trained in parallel, leading to time savings and improved model quality. Furthermore, the dynamic objective placement method ensures efficient training for high-dimensional models. In [16] with multi-objective optimization problem was transformed into a single-objective Mixed Integer Linear Programming (MILP) problem, aiming to maximize the acceptable flow rate while minimizing energy costs for multiple service chains. A polynomial-time algorithm, based on linear relaxation and rounding, to approximate the optimal solution of MILP. This approach achieved a weakly Pareto-optimal solution resulting in balancing acceptable flow rate and energy cost, while ensuring easy implementation. Similarly, studies [17,18] utilized the DRL capability to address the challenges in SFC placement in terms of service flexibility and diversity. With the existing DRL lacking the ability in solving a dynamic number of objectives and unreliable training time consumption, the author takes the advantage of divided sub-problem’s weight vectors to show the relation toward the solution position of placement. By proposing related zone decomposition (MQDRL), the weighted vectors are trained in parallel, resulting in time-saving and improving the quality of models. Then, the dynamic objective placement method ensures the training with the high-dimension model. As a result, the proposal achieved an improvement of the service acceptance ratio and the hyper-volume value in multi-objective SFC placement. Reference [19] categorized two sub-classes of the task according to their real-time (type-0) and non-real-time (type-1) requirements according to the priority level and latency demand. The author proposed two-phase solutions. First, the PARS scheme prioritizes type-0 users and allows network operators to assist the resource over the performance of SFC resource allocation. In the next phase, an adaptive algorithm, with two parts, was created; the first part identifies the best candidate node for deploying VNF, then the second part evaluates the best path connecting identified nodes using Dijkstra’s minimum cost path-finding algorithm.
In [20], the authors addressed the resource competition of utilizing collaborative edges by proposing a prioritized queuing scheme with task dependency (PQTD) to allow the task with high priority to jump the queue, which addressed the resource competition in the edge with the drawback of implementing algorithms first come first served (FCFS) and other benchmark schemes. In addition, a joint DAG-queue delay model was proposed to address the complication in delay change due to queue-jumping by analyzing the chain reaction of delay changes. As a multi-task assignment problem, PQTD with Monte Carlo Tree Search algorithm is proposed to maximize the average satisfaction degree framework and avoid local optimization dilemmas. To address the resource constraints in IoT devices, the study in [21] introduces IRATS, a PPO-based intelligent algorithm for priority- and deadline-aware resource allocation and task scheduling, aimed at enhancing computational workloads and communication capabilities. By modeling resource allocation as an MDP, IRATS minimizes task waiting time and delay, using a multi-level queue scheduler to prioritize tasks.
2.2 DRL-GNN Method-Based Resource Utilization in IoT
Although deep learning (DL) can effectively capture patterns in Euclidean data, the complexity of network topology and data generation in IoT networks presents significant challenges, as conventional DL algorithms struggle to handle non-Euclidean data effectively. For that reason, researchers found that GNN is one of the up-to-date methods to extract network properties in network environments and utilize in SFC deployments according to superiority of GNN in graph features extraction.
For instance, Zhang et al. [22] employed GNNs alongside a multitask regression layer to analyze VNF-FG topologies and predict future resource requirements. Similarly, reference [23] explored resource allocation and task scheduling for edge machine learning, integrating heuristic algorithms, GNNs, and Multi-Agent Reinforcement Learning (MARL). While heuristic approaches lack scalability in dynamic environments, GNN-based models effectively capture task-device dependencies. Furthermore, MARL enhances decentralized decision-making, allowing multiple agents to collaboratively optimize resource usage. Building on these advancements, our work combines GNNs and MARL to overcome limitations in scalability and adaptability, achieving enhanced task placement and resource efficiency. Additionally, reference [24] proposed DRL-based on GNN approach to optimize resource consumption in elastic optical data center. This reinforcement learning method integrated Graph Convolutional Networks (GCNs) with temporal-difference learning. This method leverages GCNs to capture and provide detailed insights into the network state, aiming to maximize the long-term average revenue while enabling real-time deployment decisions for SFC requests.
In summary, existing research has primarily focused on resource utilization and placement. While previous studies have made significant contributions to their respective domains [25–29], challenges remain in addressing the diverse critical service priorities of real-time computation in IoT networks. Specifically, there is a need to achieve convergence between autonomous resource allocation, placement, and the mapping of service criticality.
In this paper, we investigate how to handle priority-demanding resource management using an effective solution which utilizes the GNN model for effective extraction of network in IoT network. MPNN executes through three steps: hidden layer, aggregation, and readout in network topologies. Subsequently, our model integrates with DQN-enabled NFV architecture to ensure efficient priority-aware resource orchestration and adaptive resource utilization in SFC deployment. This work has two objectives minimizing short delay time on executing VNF instances and ensuring resource utilization properly for resource placement in SFC deployment.
3 System Model and Problem Formulation
At this stage, NFV architecture integrated with DRL-GNN are given to address DQN with agent policy formulation for determining service pooling and orchestrating resource. Therefore, forwarding decisions and rules obey the optimal SFC orchestration policy. In VNF-FG, descriptor is cooperated with leveraging the Topology and Orchestration Specifications for Cloud Applications (TOSCA) template in NFV [30]. We introduce the system model, including the substrate physical resource in the MEC server of the network model, SFC request model, and resource allocation. Then, we aim to offer a formulation method to represent our proposed scheme, including constraints, and minimize the execution cost and delay in system architecture. The key notation of our system model is in Table 1.

We use the undirected graph to represent the physical resource.
We assume the set of service requests is denoted by
Decision Variables
In this case, optimized time is divided into multiple sequential discrete time slots. It provides optimal VNF placement and virtual link mapping, minimizing cost, resource usage, and delay.
• Binary variable
• Binary variable
The node ensures the maximum capacity of a physical MEC server is not eclipsed by the resources in VNFs. Every server possesses limited computational resources, such as CPU, memory, and disk components. The resource demands of the VNFs are added up for all service requests that are assigned to servers. This constraint prevents overloaded physical nodes from receiving too many requests while protecting the quality of services and the system’s overall stability. Resource demand of a specific VNF
In a particular defined bandwidth constraint, a set of links denoted as
Communication captures the delay caused by data transmission over physical links. Hence, communication delay for virtual links relies on the propagation delay from ingress to egress by chaining the VNF nodes. Otherwise, we need to make sure in ensuring that the service request meets the latency requirement.
Now, we need to make sure that computation delay for processing a VNF on a MEC server ensures that VNFs are executed within reasonable time limits. Eq. (4) expresses the computation allocation in MEC for executing
Minimizing the total execution delay is a key objective, which achieved by optimally allocating sufficient computational resources, ensuring adequate bandwidth for the NFV-enabled
The objective is to minimize the total cost and delay in embedding service requests. The total cost includes computing resources, bandwidth usage, and propagation delay. The delay encompasses computation and communication delays for VNFs.
4 Graph Reinforcement Learning-Based Solution in SFC
In this section, we present the design of DRL-GNN algorithms to address the proper handling of resource allocation and VNF placement for provisioning resource priority on demands. Firstly, we introduce the GNN framework that interacts with network environments. GNN embedded the service functionality in VNF nodes as a directed graph by leveraging Message Passing Neural Network (MPNN) variant. Meanwhile, the DRL framework is feasibly defined as the agent policy to select resource utilization regarding the VNF-FG placement problem. It significantly decomposes the creation of a realistic networking environment to make a policy charging for network configuration. GNN gathers network states and readout as node embedding, link level, and MEC graph. After gaining output from the GNN operation, DRL will define the action, and DQN will decompose the target q-network and online network as in Fig. 2.

Figure 2: An architecture of DRL-GNN embedding in SDNC-NFVO for service priority
4.1 Priority-Awareness-Based Approach
We study the significance of resource management MEC servers-assisted in SDN/NFV controller that enhances the orchestration and management flow entries of IoT workload. SDN/NFV-assisted MEC environment, where it allocates the resource cost, link, bandwidth, and required resource to instantiate per VNFs. We propose leveraging a GNN-enabled DRL algorithm, namely a priority-aware VNF deployment algorithm (PVD-GDR).
State spaces indicate the significance of gathering which consists of
The decision-maker and flow selection,
The evaluation of agent and SDNC/NFVO control policy from overall open flow that maximize the long-term reward expectation by experience from each iteration. In our proposed method, the reward, denoted as
4.1.4 Optimal Policy Selection
Q-value and loss optimization concentrate on improving long-term goals and adjusting parameters for deep neural networks that act as function approximators. The target network utilizes replay batch parameters throughout the training process. Nevertheless, the loss values can vary during each training session, sometimes higher or lower than the intended level. The loss is calculated using the mean squared error (MSE), which assesses the squared difference between the target Q-value
4.2 GNN-Based Resource Embedding
In primary GNN, an input feature was received from SDN/NFV-assisted MEC entities. GNN model defines three phases: message passing, aggregation and update, and readout. Resource capacities (e.g., create the sequential of VNFs and VLs), which appended into hidden state
4.3 DQN-Based MEC Resource Pooling
For network environments and agent initiatives, frameworks are leveraged to design and implement including the primary algorithm of DRL-GNN, which uses DQN in online networks and target network optimization. Algorithm 1 presents the PVD-GDR function in approximating

5.1 Simulation Parameter and Specification
Simulation parameters and specifications in network scenarios are used to define the operation in terms of resource utilization and optimization for intended resource management to achieve satisfaction based on resource priority and demands. Hence, the parameter is conducted with primary parameters, and hyperparameters used in this softwarization and virtualization framework are described in Table 2. Specifically, Mininet and TensorFlow are used to conduct our proposed.

• Mininet and mini-NFV: it creates realistic network topologies and resource management for network environments [31,32]. Ryu creates the flow manager over OVS [33]. The number of edge devices is 300 connected to 5 MEC servers and the maximum delay on a link is allowable, which ranges from 5–15 ms per ISF (inference service function). MEC servers are utilized to handle computation in closed-to IoT devices which allocate the resource for computing tasks to address the latency and avoid the traffic congestion. Hence, the placement of resources in instantiating VNF is 4 in each service request. Setting upper-bound bandwidth is 20 MHz [34].
• TensorFlow setup: it provides the design for relevance for running GNN and DRL models by defining the functionality to initialize the environment [35,36]. The episode is set to 500, the replay memory size is set to 50,000, and the batch size is set to (64, 128, 256). The learning rate, discount factor, hidden layer, and exploration rate are set to 0.001, 0.95, (32–256), and 0.5, respectively.
With the above given simulation scenarios, we dive to address our management approach and reference approach in the controller to evaluate the performance. Network performance is conducted in several key performance indicators such as the cumulative reward evaluation on metrics, end-to-end task execution delay of requested service, delivery ratio tasks of requested service, and task drop ratio of requested service.
In our proposed scheme, the suggested approach collects state features and the experienced path to dynamically modify critical weight to address the unpredictability of congestion and performance fluctuations. The designated simulation parameters are conducted with simulation scenarios randomly from normal to heavy congestion. To demonstrate the performance efficiency of the proposed approach, three reference schemes are chosen for experimentation and comparison as follows:
• Load-balancing shows about dedicating network selection which calculates the target network selection and workloads evenly across available resources to optimize system performance and prevent resource overloading. This approach assigns VNFs to SFC deployment based on their priority levels, ensuring critical functions are handled with higher preference while balancing the overall system load. However, it cannot overcome the priorities constraints of VNFs in priority demands.
• Greedy approach is a conventional approach that leads to finding the optimal policy in terms of cost or resource allocation. While greedy approach can be effective for smaller networks or scenarios with less complexity, it may struggle in large-scale, dynamic environments.
• Deep Q-learning (DQL) indicates the sole reinforcement learning-based SDNC method, where a single Q-learning algorithm is utilized to manage the MEC network resource. The agent acquires knowledge via trial and error, refining its policy by repeatedly querying through the Q-learning process (for instance, by employing q-table).
5.2 Performance Evaluation and Discussion
This sub-section presents the results of both the proposed and reference schemes based on five key performance metrics. Each scheme adheres to the simulation configuration for the overall topologies. Fig. 3a demonstrates how the agent explored and exploited. The evaluation illustrates how each algorithm performed throughout the episodes in terms of effectively exploring and exploiting resources. In the initial episodes, the greedy algorithm recorded a negative reward of −20.51, considerably lower than DQL and load-balancing which achieved −10.01 and −15.12, respectively. Furthermore, the PVD-GDR algorithm exhibited a positive reward of 1.29, showcasing its ability to adapt to resource exploration and exploitation early on. At the final episode, PVD-GDR reaches a cumulative reward of 40.08, indicating optimal exploitation and effective prioritization of SFC deployments under varying resource constraints. PVD-GDR reaches a reward of 40.08, which is 33.07, 27.69, and 18.74 of load-balancing, greedy, and DQL, respectively. The PVD-GDR algorithm successfully balances the trade-offs among costs, latency, and resource utilization while adapting to congestion states. Its multi-faceted weighted reward guarantees that priority-aware deployment of SFCs can manage complex, real-time IoT service requests with different priority classes. The algorithm reduces non-optimal orchestrations, ensuring reliable and scalable SFC deployments for various use cases, especially in settings that require dynamic adaptability and efficient resource allocation. In Fig. 3b, in the initial stages, the load-balancing algorithm achieved a delay of 12.02 ms, which was slightly higher than greedy of 11.19 ms and DQL of 11.16 ms. PVD-GDR algorithm shown delay of 10.12 ms, reflecting its ability to manage task execution more efficiently from the outset. As the simulation progressed, PVD-GDR continued to maintain relatively stable and lower delays. At the final simulation timeslot, PVD-GDR algorithm excels in reducing the task execution delay by effectively balancing resource utilization and delay optimization. This makes PVD-GDR the most effective and reliable choice for managing real-time SFC in a dynamic and large-scale network environment. Fig. 3c presents in processing for real-time applications on IoT devices, QCI establishes the necessary upper-limit tolerable delays lower than those for lower critical application. In the initial stages, PVD-GDR achieved a delivery ratio of 99.9985% at 50 ms, PVD-GDR maintained a high delivery ratio, reaching 99.9001% at 500 ms. This result is greater than load-balancing, greedy, and DQL of 0.1675%, 0.0804%, and 0.2802%, respectively. The proposed PVD-GDR algorithm excels in delivering high success rates for service tasks due to its proactive resource management, ensuring that VNFs are appropriately positioned and efficiently allocated. Fig. 3d shows the task drop ratio of requested service where network states are configured in succession based on increasing congestion levels and the complexity of application tasks. We assess the flexibility of the agent by examining the variations in results in relation to the diversity of intense tasks and congestion conditions. Results demonstrate that PVD-GDR significantly outperforms load-balancing, greedy, and DQL in minimizing the task drop ratio over increasing simulation times. While greedy struggles with consistently high task drops and DQL shows moderate improvement, PVD-GDR ensures minimal task drops even under increasing workloads. This highlights its robustness and reliability for efficient task allocation in dynamic and resource-constrained environments.

Figure 3: Result of proposed and reference schemes on (a) cumulative reward evaluation metric, (b) end-to-end execution delay, (c) delivery ratio, and (d) task drop ratio of requested service
Fig. 4 presents the average completion ratios of the requested service tasks throughout the simulation timeslots. PVD-GDR continues to demonstrate exceptional performance, reaching 99.88% at the final timeslot, which is higher than load-balancing, greedy, and DQL, thereby maintaining a steady and high completion ratio throughout. The superior performance of PVD-GDR can be attributed to its efficient handling of resource allocation, ensuring that tasks are completed successfully even under dynamic network conditions. PVD-GDR optimizes the completion of service tasks, making it an ideal choice for environments that require high reliability and low-latency performance.

Figure 4: Average completion ratio of requested services
This paper presented a priority-aware resource allocation for VNF deployment in MEC environments for IoT devices. The proposed scheme utilized MEC servers and SDN/NFV to capture traffic flow control and manage the resource orchestration to achieve scalable resource computation. We proposed DRL-GNN algorithm called the Priority-Aware VNF Deployment Algorithm (PVD-GDR) integrated with SDN/NFV architecture. Additionally, GNN is leveraged to execute the network environment in three phases: hidden layer, aggregation, and readout. On the other hand, we designed resource allocation and placement in SFC deployment to execute massive resource constraints in IoT services aligned with their priority level based on DQN. This PVD-GDR approach effectively prioritizes high-priority tasks while minimizing the impact on lower-priority tasks, ensuring QoS for critical applications. Finally, we conduct a simulation to evaluate efficiency of PVD-GDR, demonstrating its superior performance compared to reference schemes. Our proposed enhances efficient resource utilization and optimizes for diversity service priority on IoT device demands. In future work, we aim to extend simulation to an MEC-enabled O-RAN system, supporting diversity platforms. Additionally, we explore MARL to handle different criticalities in IoT networks.
Acknowledgement: Not applicable.
Funding Statement: This work was supported by Institute of Information & Communications Technology Planning and Evaluation (IITP) grant funded by the Korean government (MSIT) (No. RS-2022-00167197, Development of Intelligent 5G/6G Infrastructure Technology for the Smart City); in part by the National Research Foundation of Korea (NRF), Ministry of Education, through the Basic Science Research Program under Grant NRF-2020R1I1A3066543; in part by BK21 FOUR (Fostering Outstanding Universities for Research) under Grant 5199990914048; and in part by the Soonchunhyang University Research Fund.
Author Contributions: The authors confirm contribution to the paper as follows: study conception and design: Seyha Ros, Seungwoo Kang, Seokhoon Kim; data collection: Seyha Ros, Seungwoo Kang, Taikuong Iv; analysis and interpretation of results: Seyha Ros, Prohim Tam, Inseok Song; draft manuscript preparation: Seyha Ros, Taikuong Iv, Seungwoo Kang, Inseok Song, Prohim Tam, Seokhoon Kim. All authors reviewed the results and approved the final version of the manuscript.
Availability of Data and Materials: Not applicable.
Ethics Approval: Not applicable.
Conflicts of Interest: The authors declare no conflicts of interest to report regarding the present study.
References
1. Mostafavi S, Hakami V, Sanaei M. Quality of service provisioning in network function virtualization: a survey. Computing. 2021;103(5):917–91. doi:10.1007/s00607-021-00925-x. [Google Scholar] [CrossRef]
2. Yi B, Wang X, Li K, Das SK, Huang M. A comprehensive survey of network function virtualization. Comput Netw. 2018;133(2):212–62. doi:10.1016/j.comnet.2018.01.021. [Google Scholar] [CrossRef]
3. Shahraki A, Taherkordi A, Haugen Ø, Eliassen F. A survey and future directions on clustering: from WSNs to IoT and modern networking paradigms. IEEE Trans Netw Serv Manag. 2021;18(2):2242–74. doi:10.1109/TNSM.2020.3035315. [Google Scholar] [CrossRef]
4. Bizanis N, Kuipers FA. SDN and virtualization solutions for the Internet of Things: a survey. IEEE Access. 2016;4:5591–606. doi:10.1109/ACCESS.2016.2607786. [Google Scholar] [CrossRef]
5. Siddiqui S, Hameed S, Shah SA, Ahmad I, Aneiba A, Draheim D, et al. Toward software-defined networking-based IoT frameworks: a systematic literature review, taxonomy, open challenges and prospects. IEEE Access. 2022;10(20):70850–901. doi:10.1109/ACCESS.2022.3188311. [Google Scholar] [CrossRef]
6. Bera S, Misra S, Vasilakos AV. Software-defined networking for Internet of Things: a survey. IEEE Internet Things J. 2017;4(6):1994–2008. doi:10.1109/JIOT.2017.2746186. [Google Scholar] [CrossRef]
7. ETSI GS NFV 002 V1.2.1 (2014-12). Network functions virtualisation (NFV). In: Architectural framework. Nice, France: ETSI; 2014. [Google Scholar]
8. Abderrahmane A, Drid H, Behaz A. A survey of controller placement problem in SDN-IoT network. Int J Networked Distrib Comput. 2024;12(2):170–84. doi:10.1007/s44227-024-00035-y. [Google Scholar] [CrossRef]
9. Alam I, Sharif K, Li F, Latif Z, Karim MM, Biswas S, et al. A survey of network virtualization techniques for Internet of Things using SDN and NFV. ACM Comput Surv. 2021;53(2):1–40. doi:10.1145/3379444. [Google Scholar] [CrossRef]
10. Cao B, Zhang J, Liu X, Sun Z, Cao W, Nowak RM, et al. Edge-cloud resource scheduling in space-air–ground-integrated networks for Internet of vehicles. IEEE Internet Things J. 2022;9(8):5765–72. doi:10.1109/JIOT.2021.3065583. [Google Scholar] [CrossRef]
11. Hantouti H, Benamar N, Taleb T. Service function chaining in 5G & beyond networks: challenges and open research issues. IEEE Netw. 2020;34(4):320–7. doi:10.1109/MNET.001.1900554. [Google Scholar] [CrossRef]
12. Yang S, Li F, Trajanovski S, Yahyapour R, Fu X. Recent advances of resource allocation in network function virtualization. IEEE Trans Parallel Distrib Syst. 2021;32(2):295–314. doi:10.1109/TPDS.2020.3017001. [Google Scholar] [CrossRef]
13. Ros S, Tam P, Song I, Kang S, Kim S. Handling efficient VNF placement with graph-based reinforcement learning for SFC fault tolerance. Electronics. 2024;13(13):2552. doi:10.3390/electronics13132552. [Google Scholar] [CrossRef]
14. Tam P, Math S, Kim S. Priority-aware resource management for adaptive service function chaining in real-time intelligent IoT services. Electronics. 2022;11(19):2976. doi:10.3390/electronics11192976. [Google Scholar] [CrossRef]
15. Farkiani B, Bakhshi B, Ali MirHassani S, Wauters T, Volckaert B, De Turck F. Prioritized deployment of dynamic service function chains. IEEE/ACM Trans Netw. 2021;29(3):979–93. doi:10.1109/TNET.2021.3055074. [Google Scholar] [CrossRef]
16. Jang I, Suh D, Pack S, Dán G. Joint optimization of service function placement and flow distribution for service function chaining. IEEE J Sel Areas Commun. 2017;35(11):2532–41. doi:10.1109/JSAC.2017.2760162. [Google Scholar] [CrossRef]
17. Chen Z, Xiong B, Chen X, Min G, Li J. Joint computation offloading and resource allocation in multi-edge smart communities with personalized federated deep reinforcement learning. IEEE Trans Mob Comput. 2024;23(12):11604–19. doi:10.1109/TMC.2024.3396511. [Google Scholar] [CrossRef]
18. Zhou C, Zhao B, Tang F, Han B, Wang B. Dynamic multi-objective service function chain placement based on deep reinforcement learning. IEEE Trans Netw Serv Manag. 2024;PP(99):1. doi:10.1109/TNSM.2024.3446248. [Google Scholar] [CrossRef]
19. Pati PS, Sonu M, Datta R, Krishnaswamy D, Singhal C. PARS: a priority-aware resource sharing scheme for services with differential QoS requirements in B5G edge network. IEEE Netw Lett. 2023;5(2):105–9. doi:10.1109/LNET.2023.3256087. [Google Scholar] [CrossRef]
20. Cai Q, Zhou Y, Liu L, Qi Y, Shi J. Prioritized assignment with task dependency in collaborative mobile edge computing. IEEE Trans Mob Comput. 2024;23(12):13505–21. doi:10.1109/TMC.2024.3427380. [Google Scholar] [CrossRef]
21. Jamil B, Ijaz H, Shojafar M, Munir K. IRATS: a DRL-based intelligent priority and deadline-aware online resource allocation and task scheduling algorithm in a vehicular fog network. Ad Hoc Netw. 2023;141(7):103090. doi:10.1016/j.adhoc.2023.103090. [Google Scholar] [CrossRef]
22. Zhang J, Liu Y, Li Z, Lu Y. Forecast-assisted service function chain dynamic deployment for SDN/NFV-enabled cloud management systems. IEEE Syst J. 2023;17(3):4371–82. doi:10.1109/JSYST.2023.3263865. [Google Scholar] [CrossRef]
23. Li Y, Zhang X, Zeng T, Duan J, Wu C, Wu D, et al. Task placement and resource allocation for edge machine learning: a GNN-based multi-agent reinforcement learning paradigm. IEEE Trans Parallel Distrib Syst. 2023;34(12):3073–89. doi:10.1109/TPDS.2023.3313779. [Google Scholar] [CrossRef]
24. Li B, Zhu Z. GNN-based hierarchical deep reinforcement learning for NFV-oriented online resource orchestration in elastic optical DCIs. J Light Technol. 2022;40(4):935–46. doi:10.1109/JLT.2021.3125974. [Google Scholar] [CrossRef]
25. Pan P, Fan Q, Wang S, Li X, Li J, Shi W. GCN-TD: a learning-based approach for service function chain deployment on the fly. In: Proceedings of the IEEE Global Communications Conference (GLOBECOM); 2020 Dec 7–11; Taipei, Taiwan. p. 1–6. doi:10.1109/globecom42002.2020.9322359. [Google Scholar] [CrossRef]
26. Huang W, Li S, Wang S, Li H. An improved adaptive service function chain mapping method based on deep reinforcement learning. Electronics. 2023;12(6):1307. doi:10.3390/electronics12061307. [Google Scholar] [CrossRef]
27. Shokrnezhad M, Taleb T. ORIENT: a priority-aware energy-efficient approach for latency-sensitive applications in 6G. arXiv:2402.06931. 2024. [Google Scholar]
28. Khemani B, Patil S, Kotecha K, Tanwar S. A review of graph neural networks: concepts, architectures, techniques, challenges, datasets, applications, and future directions. J Big Data. 2024;11(1):18. doi:10.1186/s40537-023-00876-4. [Google Scholar] [CrossRef]
29. Tam P, Song I, Kang S, Ros S, Kim S. Graph neural networks for intelligent modelling in network management and orchestration: a survey on communications. Electronics. 2022;11(20):3371. doi:10.3390/electronics11203371. [Google Scholar] [CrossRef]
30. Di Martino B, Cretella G, Esposito A. Defining cloud services workflow: a comparison between TOSCA and OpenStack hot. In: Proceedings of the 2015 Ninth International Conference on Complex, Intelligent, and Software Intensive Systems; 2015 Jul 8–10; Santa Catarina, Brazil. p. 541–6. doi:10.1109/CISIS.2015.80. [Google Scholar] [CrossRef]
31. Lantz B, O’Connor B. A mininet-based virtual testbed for distributed SDN development. SIGCOMM Comput Commun Rev. 2015;45(4):365–6. doi:10.1145/2829988.2790030. [Google Scholar] [CrossRef]
32. Castillo-Lema J, Venancio Neto A, de Oliveira F, Takeo Kofuji S. Mininet-nfv: evolving Mininet with oasis Tosca nvf profiles Towards Reproducible nfv Prototyping. In: Proceedings of the IEEE Conference on Network Softwarization (NetSoft); 2019 Jun 24–28; Paris, France. p. 506–12. doi:10.1109/netsoft.2019.8806686. [Google Scholar] [CrossRef]
33. Asadollahi S, Goswami B, Sameer M. Ryu controller’s scalability experiment on software defined networks. In: Proceedings of the IEEE International Conference on Current Trends in Advanced Computing (ICCTAC); 2018 Feb 1–2; Bangalore, India. p. 1–5. doi:10.1109/ICCTAC.2018.8370397. [Google Scholar] [CrossRef]
34. ETSI TS 138 101-1 V15.2.0 (2018-07). 5G; NR; User Equipment (UE) radio transmission and reception. In: Part 1: range 1 standalone (3GPP TS 38.101-1 version 15.2.0 Release 15). Nice, France: ETSI; 2018. [Google Scholar]
35. Ferludin O, Eigenwillig A, Blais M, Zelle D, Pfeifer J, Sanchez-Gonzalez A, et al. TF-GNN: graph neural networks in TensorFlow. arXiv:2207.03522. 2022. [Google Scholar]
36. Ajay Rao P, Navaneesh Kumar B, Cadabam S, Praveena T. Distributed deep reinforcement learning using TensorFlow. In: Proceedings of the 2017 International Conference on Current Trends in Computer, Electrical, Electronics and Communication (CTCEEC); 2017 Sep 8–9; Mysore, India. p. 171–4. doi:10.1109/CTCEEC.2017.8455196. [Google Scholar] [CrossRef]
Cite This Article
Copyright © 2025 The Author(s). Published by Tech Science Press.This work is licensed under a Creative Commons Attribution 4.0 International License , which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.


Submit a Paper
Propose a Special lssue
View Full Text
Download PDF
Downloads
Citation Tools