Next Article in Journal
Research on Longitudinal Thermoelastic Waves in an Orthotropic Anisotropic Hollow Cylinder Based on the Thermoelastic Theory of Green–Naghdi
Previous Article in Journal
DFD-SLAM: Visual SLAM with Deep Features in Dynamic Environment
Previous Article in Special Issue
A Two-Stage Automatic Container Code Recognition Method Considering Environmental Interference
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

User-Centric Internet of Things and Controlled Service Scheduling Scheme for a Software-Defined Network

1
Department of Computer Engineering, Aligarh Muslim University, Aligarh 202002, India
2
School of Computing, Gachon University, Seongnam 13120, Republic of Korea
3
Department of Zoology, College of Science, King Saud University, Riyadh 11451, Saudi Arabia
*
Author to whom correspondence should be addressed.
Submission received: 27 April 2024 / Revised: 30 May 2024 / Accepted: 3 June 2024 / Published: 6 June 2024

Abstract

:
Mobile users can access vital real-time services through wireless paradigms like software-defined network (SDN) topologies and the Internet of Things. Point-of-contact-based infrastructures and dynamic user densities increase resource access and service-sharing concurrency. Thus, controlling power consumption and network and device congestion becomes a major issue for SDN-based IoT applications. This paper uses the Controlled Service Scheduling Scheme (CS3) to address the challenge of simultaneous scheduling and power allocation. The suggested approach uses deep recurrent learning and probabilistic balancing for power allocation and service distribution during user-centric concurrent sharing intervals. The SDN control plane decides how much power to use for service delivery while forecasting user service demands directs the scheduling interval allocation. Power management is under the control plane of the SDN, whereas service distribution is under the data plane. Power-to-service requirements are evaluated probabilistically, and updates for both aircraft are obtained via the deep learning model. This allocation serves as the basis for training the learning model to alleviate power deficits across succeeding intervals. The simulation experiments are modeled using the Contiki Cooja simulator, where 200 mobile users are placed. The proposed plan delivers a 14.9% high-service distribution for various users, 18.29% less delay, 13.34% less failure, 5.54% less downtime, and 18.68% less power consumption.

1. Introduction

The collaborative Internet of Things (IoT) technology venture enhances scalability and mobility support for compatible devices and users. IoT application services must correctly schedule network traffic and distribution intervals to prevent latency-less service distributions. Software-defined network (SDN) traffic classification capabilities and network virtualizations in IoT services provide more reliable scheduling. As a result, the distribution and widespread use of the IoT platform is managed without sacrificing service quality [1,2]. In addition, scheduling is performed for services customized to a given application to minimize resource stagnation and user waiting times. The SDN correctly distinguishes between user requirements, service availability, and distribution using software-based controls and control plane instructions [3]. Essential power requirements must be met for the IoT environment’s gadgets to operate smoothly. As a result, power management and allocation became important study areas for IoT service deployments [4]. Power distribution, conservation, and harvesting are essential components of energy management. Based on IoT and SDN concepts, unnecessary power distribution and allocations are limited [5]. Power control and management extend their useful lives by preserving the devices’ remaining energy. As a result, the device’s responsiveness and widespread use are preserved, making the smooth dissemination of IoT services more dependable [6]. Resource and user-centric power management and allocation techniques have been proposed for large-scale applications that require energy-efficient services. Consequently, IoT power management is integrated with resource distribution and allocation procedures, necessitating machine or SDN-based optimization [7].
The OpenFlow Protocol facilitates communication between the controller and the data plane in an SDN-based cloud environment. It decouples the network from individual switches and lets administrators remotely control routing tables and centralized packet-switching choices. This allows switches and data centers to be programmed independently [8]. Instead of a central server or cloud, predictions and data analysis are increasingly used at the edge to draw conclusions and take preventative measures [9]. The IoT is now heavily ingrained in our everyday lives and significantly impacts many aspects of our lifestyle, including how we drive, shop, and even decide what to eat and avoid, preserving our health. Flexible, agile, and adaptive IoT architecture is required due to the variety of applications associated with the IoT [10]. The design of an IoT architecture based on the SDN incorporates intelligent management functions by separating the control plane from the data plane to satisfy this requirement [11]. Nevertheless, most of the current studies in this sector focus on a particular area, which encourages researchers and engineers to provide comprehensive guidance on growth and direction for the future [12]. By switching from an open-loop to a closed-loop approach, the suggested strategy seeks to improve resource allocation efficiency and user experience in IoT networks by putting the user at the center of the network [13]. To improve traffic control and maintainability, the IoT application has recently included virtualization of resources and network control using software-defined networking policies. To enable the adjustments, though, the IoT and software-defined networking needs must coincide [14]. The IoT environment consists of heterogeneous devices, applications, and communication modes for aiding large-scale user-specific services. The computation, analytical, and visualization features are either built-in or acquired from different platforms [15]. Mobile customers can now receive real-time services on wireless networks thanks to the convergence of the IoT and SDN. The centralized control provided by the SDN optimizes resource allocation, guaranteeing reliable connectivity and effective communication between servers and IoT devices. Because of this integration, service delivery is enhanced by dynamically changing network characteristics in response to customer requests. The SDN also improves security via centralized management and permits control over power usage, which makes networks more adaptable, scalable, and effective. Social benefits of the proposed methods enable more personalized and efficient services for IoT users, leading to increased convenience and improved quality of life. This research can contribute to economic growth and efficiency by reducing network congestion and optimizing resource allocation, ultimately lowering operational costs for service providers and potentially stimulating innovation and investment in IoT technologies. The novel aspects of this research utilize state-of-the-art IoT and a software-defined network (SDN) that might include new algorithms, methodologies, or frameworks developed to address challenges in user-centric service scheduling within SDN environments.
The main contributions and novelty of this paper are:
  • To address the challenge of simultaneous scheduling and power allocation utilizing the Controlled Service Scheduling Scheme (CS3).
  • To use deep recurrent learning and probabilistic balancing for power allocation and service distribution during user-centric concurrent sharing intervals.
  • To train the learning model to alleviate power deficits across succeeding intervals.
  • To assess the efficacy of the proposed design by using matrices such as power consumption, service distribution ratio, delay, failures, and outages.
The remaining part of the study includes Section 2, covering the literature review; Section 3, covering the topics of the proposed study along with performance outcomes; and finally, the paper concludes in Section 4, along with future studies.

2. Literature Review

Sun et al. [16] efficiently implement resource allocation and compute offloading to reduce request time and energy usage. Reliable processing between IoT devices is approximated through computation offloading and resource allocation. The IoT-based fog-cloud environment reduces the cost and time required for this work. Consequently, the various sets of IoT-based devices experience higher energy consumption.
Iterative searching-based job offloading in multi-access edge computing for effective ultra-dense IoT devices was proposed by Guo et al. [17]. In this research, task offloading addresses the complexity of processing and transmission power. Offloading is performed on numerous edge servers to estimate power allocation efficiently. The computer task is identified for energy-conscious task offloading, and the power allocation is estimated.
The development of the delay-based allocation of workloads (DBWA) method addresses the problems of environmentally friendly and delay-guaranteed task allocation [18]. The Lyapunov drift-plus-penalty theory generates the latency guarantee in edge cloud systems based on the IoT. Enhancing green computing’s energy consumption is the goal of this effort. In the IoT platform, the quality of service is preserved between the edge servers.
A dynamic microservices scheduling technique is presented for computational complexity in mobile edge computing. The network delay is reduced by increasing the number of IoT-based services utilized. Scheduling, in addition to queuing, Ref. [19] improves the pace of energy consumption. This paper aims to increase energy efficiency for the best available resources in mobile edge computing.
A multiple-input multiple-output non-orthogonal multiple access system was presented by Wang et al. [20] to convey the signal to the access control with high connectivity. This paper aims to enhance ultra-low latency and reliability. Zero-forcing beamforming, a beamforming optimization approach, is introduced for power allocation. The IoT network addresses the best local solution by optimizing the technique.
Ref. [21] describes a two-stage dimensional matching method for estimating the joint power control and time allocation platform. Here, the optimal allocation is determined to address the joint choice of channels and peer discovery. This method examines three problem-matching scenarios: linear programming, nonlinear fractional computer programming, and alternating optimization. By matching the priority-based data, this technique reduces its complexity.
A three-tier design is suggested, which raises the inquiry from the source and implements scheduling for the IoT devices. Mixed-integer programming is developed with the optimization model to shorten the deadline. The round-robin and scheduling of priorities are estimated using a genetic algorithm. The entire processing is created in [22] and assessed at the specified time.
Effective computation offloading and allocation of resources are introduced in [23] to reduce energy usage and request time. Computation offload and resource allocation are estimated for accurate processing between the IoT devices. This approach reduces the time and cost in an IoT-based fog-cloud scenario. As a result, the energy consumption of the various IoT-based devices is rising.
Singh et al. [24] proposed a systematic review examining how wireless sensor networks have evolved from simple sensory monitoring to sophisticated, service-oriented applications. By comparing various architectural styles and technological eras, it provides an understanding of how technology has advanced. It also outlines the directions for future research to direct additional creativity and investigation in this quickly developing sector.
Paulson et al. [25] suggested architecture prioritizes resource provisioning while ensuring the network’s quality of service requirements are met. Jitter, latency, and throughput were compared between the SDN network strategy and the conventional method. The SDN-based IoT network increases network efficiency by lowering network overheads caused by frequent communication between the nodes and the controllers, according to the latency, delay, and throughput performance findings.
Huang et al. [26] proposed a digital twin-based user-centric resource management plan to lower resource use while increasing user satisfaction. First, the study creates a framework for resource reservations facilitated by user data transfer. Second, the study formulated a joint non-convex issue for reserving bandwidth and computing resources. To further increase user happiness, the study will look at the joint optimization of resource allocation and segment-level cache order based on distilled data from user data transfer.
Pervej et al. [27] presented a novel software-defined user-centric radio access technology solution that divides power-hungry access points into virtual cells and offers dependable, all-around connectivity to mobile-connected vehicles. Due to the heterogeneous preferences of the connected vehicles, the study introduced a preference–popularity trade-off into each content request, but also used content prefetching at the edge server with multiple classes in the content catalog to ensure fast decision making for mission-critical operations and uninterrupted onboard entertainment. The suggested method outperforms the current starting points regarding content delivery latency, deadline violation, and cache hit ratio.
Abir et al. [28] proposed a thorough analysis of software-defined unmanned aerial vehicle networks and architecture, emphasizing their needs, opportunities, and uses. In the context of 6G networking, it highlights how crucial it is to integrate the SDN and unmanned aerial vehicle-enabled networks since they provide superior flexibility, scalability, dependability, and efficient connectivity compared to traditional methods. The study also addresses practical issues with proposed networks, including security, flexibility, interference control, interoperability, and lack of standards.
Raeisi, M. and Sesay, A. B. [29] a novel mobility management feature that decreases the number of handovers in the vehicle-to-network service, hence improving the performance of 5G for high-speed road users, such as connected autonomous vehicles, in small cells. Additionally, the paper suggested a novel cell reselection process for high-speed users under our mobility management function’s control using the vehicular frequency reuse scheme and in the RRC_Connected (Radio Resource Control) state. Compared to the conventional scheme, computer simulations demonstrate that the suggested vehicular frequency reuse approach can minimize the total amount of handovers (handover rate) for users by more than 99%.
Murshed, M. [30] proposed several methods for enhancing connection stability through effective decision-making during handover. First, M-FiVH, a modified probabilistic technique, was developed to improve network stability and minimize 5G handovers. Later, an adaptive learning strategy used connectivity-oriented SARSA reinforcement learning for user-centric virtual cell management to enable effective handover decisions. The study presented an analysis and comparison of several methods and showed that our suggested methods outperform the others regarding network connection.
Wenbo Wang et al. [31] suggested the Proactive Manufacturing Resources Assignment (PMRA) method for the Smart Factory to Production Performance Prediction. First, a smart factory’s resources are equipped with distributed control capacity and made smart using modern IIoT and CPS technologies. In this scenario, resources at the cloud center and those at the edge may work together dynamically. A second technique is suggested for reliably predicting production status in the future by extracting real-time production information and using a real-time colored Petri net (RCPN) to analyze key production performance indicators (KPPIs).
Based on the survey, there are several issues with existing models in attaining high power consumption, service distribution ratio, delay, failures, and outages. Hence, this study proposes the Controlled Service Scheduling Scheme (CS3) for controlling power consumption and network and device congestion for SDN-based IoT applications.

3. Proposed Scheduling Scheme

3.1. Problem Statement

There is a lack of an overall framework that provides all of the fundamental tools to manage end-to-end resources and traffic, and the existing solutions continue to be insufficient (they only take into account latency). Although the Internet of Things (IoT) platform is very complicated, heterogeneous, and large, there is a prominent lack of cognitive processes that may reduce the amount of human involvement in the quality-of-service management process. A decrease in quality of service (QoS) levels in service provisioning is caused by the absence of mobility-aware latency-constrained service management at the edge. This is particularly relevant for an increase in communication latency. Because of the use of architectures that were developed for the Internet before the development of wireless technologies, service provisioning is affected negatively. At that time, most of the Internet was made up of static nodes, resulting in less fluctuation in the topology, and applications were not sensitive to delays. To reduce interruptions and latency for users of services at the edge, this research aims to examine problems and theories connected to service management. Similar to the previous mark, this research aims to reduce the need for service migrations and context transfers prompted by user mobility. This is because these events substantially influence the quality of service. Regarding addressing mobility-related difficulties at the network’s edge, this research focuses primarily on using software-defined networking (SDN). Allowing for the fine-grained and active management of communication flows at the edge, the global view of network entities made possible by this networking paradigm may make it easier to implement innovative solutions.
Real-time service handling is provided to mobile users in the IoT environment by deploying the SDN. Here, concurrent service sharing is estimated for dynamic users on heterogeneous platforms. This work determines a point-of-contact-based infrastructure for concurrent processing in SDN-based IoT applications. The scope of this work is to avoid latency, failure, and outage by increasing the service distribution to the valid user. Thus, concurrent processing is estimated for the mobile devices in IoT applications concerning the SDN. A formal representation of the CS3 is presented in Figure 1. The simulation experiments are modeled using the Contiki Cooja simulator, where 200 mobile users are placed. An IoT cloud architecture with seven service providers, each with 1 TB of storage, is used for service distribution. The transmit power of the devices is set as 30 dBm, and it operates in 60 scheduling intervals.
Figure 1 presents the proposed scheme in the IoT-SDN platform. The SDN is responsible for service scheduling and power management using this scheme’s data and control plane. It operates between the users and resources throughout scheduling, queuing, and distribution processes. Moreover, the power allocation for the resources is governed by scheduling and prior distribution.

3.2. Problem Definition

This work addresses scheduling and power allocation in SDN-based IoT applications. This is resolved by introducing the CS3 in this article. In this case, the user estimates the cumulative congestion and power management. The power allocation and service distribution are processed to the network and analyzed by the scheduling approach. The point-of-contact-based infrastructure is used to derive the congestion-free service transmission.
Joint scheduling and power allocation are addressed here by proposing probabilistic balancing and deep recurrent learning. In this work, the SDN is derived for the service distribution, whereas the control plane indicates power management. In this evaluation step, the service is shared by deploying the concurrent processing. The following equation derives the service sharing that deploys power management and service handling.
μ ( h a ) = r 0 v 0 M s × s u + i d × N + r n X a 0 + r 0 a 0 × b e t × c k
The determination is made for service sharing between the user and the devices in SDN-based IoT applications. Here, the problem definition is addressed by estimating the congestion in the data transfer to the other devices. In this computation step, the concurrent processing is derived that deploys service sharing and resource allocation. In this evaluation step, concurrent processing is carried out for the service distribution among the dynamic user densities and point-of-contact-based infrastructure. Here, the service sharing is performed for the valid user by deploying the reliable distribution to the device in an SDN network-based IoT application. The service sharing is performed for the other devices in the network, and it is represented as h a ; the determination is denoted as μ . In this process, the distribution is estimated for scheduling.
Scheduling is performed for the devices in the IoT application termed as S . In this processing step, congestion is addressed and decreased, and it is represented as N ; the power allocation is estimated for the user and the device is termed as a 0 . The user and the number of users in the network are defined as r 0   a n d   r n , and the device and number of devices are represented as v 0   a n d   v n , respectively; service distribution is termed as i d . The access is provided to a valid user, and it is denoted as c k   a n d   V ; balancing is performed for the data transmission, and it is represented as b . The time is calculated for a better examination of data analysis, referred to as e t ; an examination is performed for the congestion in data forwarding, and it is denoted as X . The resources in the network are estimated for the sharing represented as s u ; the power management is denoted as M . The following equation determines resource access for the requested user-centric method.
μ c k , s u = V r 0 N × X + v n × S a 0 + r 0 + h a V + r e t
Resource access is provided to the valid end-user, who deploys concurrent service sharing in the IoT application. Here, the SDN is used to deploy service sharing, and power allocation is estimated for reliable data computation. The transmission is carried out for the network devices and estimates the scheduling for the user. In this computation step, power management is performed to derive the concurrent service handling from the end-user. The examination is conducted for the service distribution and power allocation method for the end-user, and it is denoted as r . The valid user accesses the information by forwarding it from the power management system.
Scheduling is performed to estimate the better processing of devices in the SDN and provides the power allocation for further processing. Here, the resource is derived for the requested user and the balancing access is estimated. The power management decides whether the access is forwarded if the user requests a particular service. Based on this process, the access is forwarded to the required user and the scheduling is evaluated as v n × S a 0 + r 0 . Sharing is performed by posting to Equation (1a); resource access is shown in Equation (1b). Scheduling is performed based on the queuing technique discussed in the section below from these two equations.

3.3. Scheduling

Scheduling is based on the user and device requests for power in the SDN. In this derivation, if power is estimated for reliable processing, then the point-of-centric-based infrastructure is created for resource sharing. Resource sharing is conducted by deploying the scheduling with the existing and the current state of the process. The task and the power allocation are queued based on the usage and congestion in the network. This method performs scheduling by deploying the SDN-based IoT application, which is determined for better service distribution. The following equation estimates the scheduling process for the concurrent service handling in the SDN.
S = h a × r 0 a 0 + v 0 M + i d + q 0 × r e t a 0 = r 0 × i d + r e t i d = r h a + N h a = r w c + M
Scheduling is performed by deploying the queuing process, which deploys resource sharing to the end-user at the mentioned time interval. Here, the access is forwarded to the valid user in the IoT network, who estimates the reliable processing by handling the data. Power management is conducted to distribute the appropriate data to the appropriate resources in the SDN. The end-user is responsible for forwarding the service based on valid access generation. If the access generated is valid, the service is forwarded to the requested user at the mentioned time interval. In this computation step, the queueing is conducted by allocating the service to the devices denoted as q 0 ; forwarding the service is represented as w c . The scheduling process is illustrated in Figure 2.
Queuing is performed for the number of resources in the IoT application and deploys the power management system. The power allocation balances the service between the user and the end-user. From this evaluation step, scheduling is carried out by determining the sharing of services, and it is represented as h a × r 0 a 0 . In this evaluation step, service sharing to the end-user is conducted by examining SDN access. The point-of-contact-based infrastructure distributes the number of end-user requests from the devices. Here, scheduling is carried out by indicating the queuing process for the valid user in the SDN; this power allocation is performed in the equation below.
a 0 = s u w c + X × q 0 + r 0 D V × S = S × r 0 + v 0 × r M X × q 0 e t
In Equation (3), scheduling is performed by deploying appropriate service handling to the end-users. In this computation step, the user is used to define the congestion in the data transmission. Here, the queuing is conducted for the scheduling devices and determines the detection process, and it is represented as D . Queuing is performed for the power allocation process that deploys power management and estimates better processing. Here, scheduling is conducted for the number of devices in the SDN used to evaluate better detection. In this derivation, power allocation is examined for resource sharing to the end-user, and the detection of congestion in the SDN is termed as D .
ScheduleSDNPower manages power scheduling and allocation within an SDN using Algorithm 1. It queues service requests, verifies device access, and computes scheduling parameters based on several parameters. Using a sophisticated formula incorporating forwarding, congestion detection, and service sharing, the power distribution depends on the congestion level. Network performance is optimized by the algorithm’s effective resource allocation management. Ultimately, it provides a formal SDN energy administration and scheduling method by returning the planned power allocation.
Algorithm 1: Scheduling
Function ScheduleSDNPower ( h a , r 0 , a 0 , v 0 , M , i d , q 0 , r , e t , w c , X , S , D )
Input:
   h a : Service sharing parameter
   r 0 : Initial service rate
   a 0 : Initial allocation parameter
   v 0 , Number of devices
   M : Power management parameter
   i d : Device index
   q 0 : Service allocation queue
   r : Optimized service rate
   e t : Processing time
   w c : Service forwarding parameter
   X : Power allocation parameter
   S : Scheduling parameter
   D : Congestion detection parameter
Output: Scheduled power allocation
Step 1: Calculate scheduling parameters
   h a r w c + M ;
   a 0 r 0 i d + r e t ;
   i d r h a + N ;
   S h a r 0 a 0 + v 0 M + i d + q 0 r e t ;
Step 2: Validate and queue device requests
     for each device in SDN, do
       if device access is valid, then
         Forward service to the requested user on the mentioned time interval;
          q 0 Queue service allocation;
          w c Power allocation;
       end if;
     end
Step 3: Calculate power allocation
     if there is congestion in end-user service, then
       a 0 = s u w c + X q 0 + r 0 D V S
     else if sharing resources when detecting congestion.
       a 0 = S r 0 + v 0 r M X q 0 e t .
     end else if
    end if
  return scheduled power allocation
end function
Queuing is performed to allocate power for the device to work in concurrent processing. Here, the resources (services) are distributed to the end-user. In this computation step, forwarding is conducted by deploying data that avoids congestion in the SDN. The congestion is addressed in this scheduling method, and it is associated with the queuing of resources. Queuing is examined for the end-user at the mentioned time, and it is represented as v 0 × r M X × q 0 e t . Thus, power allocation is performed concerning the scheduling approach; after this process, the proposed work includes two methodologies: probabilistic balancing and recurrent learning. The following section discusses these two methodologies in detail.

3.4. Probabilistic Balancing

This method balances the service with the appropriate end-user by determining whether scheduling is carried out appropriately. This constraint is achieved by examining the tree-based computation that deploys either a probability factor or is used to estimate the detection. Balancing is conducted for the service distributed to the appropriate user, and the validity of the devices is monitored. The access forwarded to the appropriate user is defined in this category and examined in the equation below. Here, queuing is performed to balance the resources.
ρ = q 0 S + r 0 × N M r n + v n
In this case, insertion and deletion are performed to balance the resource allocation method; here, it deploys better service detection in the SDN. Here, probabilistic balancing is examined for reliable computation in the tree-based process, and the root node represents the service from the power management. The left-side node indicates power allocation, whereas the right node represents sharing. Based on this, probabilistics is performed by evaluating the better detection of congestion in the SDN. Power management distinguishes service handling, and balancing is examined in this process. The examination process indicates cumulative congestion in the network and devices and estimates better power management. Figure 3 presents the queuing process of the CS3.
Power management is responsible for transmitting power to the appropriate end-user by deploying power allocation and service distribution. Power allocation is conducted for dynamic user densities and estimates congestion to avoid further failure. Here, scheduling is performed for the power allocation in this probabilistic balancing based on the tree structure, which indicates the point-of-contact-based infrastructure. Thus, probabilistic balancing is conducted for data forwarding in the network from the user to the devices, and it is termed as ρ . The following part discusses recurrent learning for power allocation in the management system.
The probabilistic balancing function uses a tree-based methodology to control service distribution and power allocation in the SDN, as shown in Algorithm 2. First, it creates a tree structure, with the root node representing power management services, the left-side representing power distribution, and the right-side branch representing service sharing. The function iterates through the tree for every service request in the queue, sharing or allocating power according to the nodes it encounters. Throughout this process, it assesses congestion detection to modify sharing and allocations. SDN systems ensure balanced resource use and effective network functioning by appropriately allocating power and other services to devices.
Algorithm 2: Probabilistic Balancing for SDN Data Transmission
Function ProbabilisticBalancing ( q 0 , r 0 , S , N . M , r n , v n )
Input :   q 0 , r 0 , S , N . M , r n , v n
Output: ρ (probability factor)
Step 1: Probabilistic Balancing
    ρ q 0 S + r 0 N M r n + v n
Step 2: Construct a probabilistic balancing tree
      tree     ConstructProbabilisticTree ( q 0 )
Step 3: Perform probabilistic balancing
  for each flow in flow_demands do
       root   tree.root
     while not leaf_node(root)
  end
Step 3: Evaluate congestion on the left and right child nodes
        left _ congestion   EvaluateCongestion(root.left_child, balanced_flows)
        right _ congestion   EvaluateCongestion(root.right_child, balanced_flows)
Step 4: Compute probabilities based on congestion levels
        left _ probability   1 − right_congestion/(left_congestion + right_congestion)
        right _ probability   1 − left_congestion/(left_congestion + right_congestion)
Step 5: Probabilistically choose the next node to traverse
     if RandomChoice(left_probability) do
         root   root.left_child
     else
         root   root.right_child
     endif
 Step 6: Update flow on the leaf node
   UpdateFlow(root, flow, balanced_flows)
  return balanced_flows
end function

3.5. Recurrent Learning

Service distribution and power allocation are estimated by introducing recurrent learning to the user-centric concurrent sharing interval. Here, power management is used to deploy congestion-free transmission and evaluate sharing with the end-user on time. The prediction is achieved by mapping the existing process and deploying resource access for service sharing. Training data determine the input neuron and perform an efficient distribution in this recurrency phase. In this processing, the input is the queued process, and it enters the initial neuron state, from the output of the first neuron forward as the input to the second neuron. Thus, the service delivery is estimated for the power allocation and addresses the deficiency. The equation below is used to analyze the current state of neuron processing.
k 0 = u ( p i ) + r a S + q 0 × o 0
In Equation (5), the current state represents the service distribution by deploying resource access associated with power allocation. Here, the queued process transmits to the next neuron layer, followed by the number of devices and users. Sharing is performed on time for the valid user based on a queued process. Scheduling is performed by deploying network and device cumulative congestion and power management. The first state indicates resource sharing, and from this, it is forwarded to the next neuron layers in recurrent learning. Here, service delivery is conducted to allocate power and deploy distribution. The neuron layer is denoted as o 0 ; the function is used to define the current state, and it is represented as u   a n d   k 0 .
The current state defines the number of neuron layers in recurrent learning that deploys joint scheduling, and power allocation is addressed. Based on the scheduling process, the analysis is performed for the current neuron state; resource access is forwarded to the second neuron layers. The dynamic user densities are due to the concurrency in service sharing and resource access in SDN-based IoT applications. The hidden layer improves training data, concurrent service sharing, and resource access. The following Equation is used to derive the hidden layer, and here, two hidden layers are used to enhance service distribution efficiently.
1 = o 0 = S + r 0 × M + i d q 0 + v 0 + b e t o 1 = S + r 1 × M + i d q 0 + v 1 + b e t o n = S + r n 1 × M + i d q 0 + v n 1 + b e t
2 = o 0 = q 0 × a 0 r 0 + k 0 v 0 × c k ( r 0 ) o 1 = q 0 × a 0 r 1 + k 0 v 1 × c k ( r 1 ) o 0 = q n × a 0 r n 1 + k 0 v n 1 × c k ( r n 1 )
In Equation (6a), the first hidden layers deploy the scheduling for the number of users in an SDN-based IoT application. Here, power management is used to determine the current layers in neuron layers, and balancing is achieved for resource access at the mentioned time. The neuron layer indicates scheduling for the number of users and deploys resource access based on power allocation. Concurrent service sharing and resource access are provided to the appropriate user. Here, the hidden layer enhances the training data for access to resources in the SDN. Scheduling and power allocation are conducted for resource access by performing service sharing. In Figure 4, the recurrent layer process is illustrated.
Figure 4 illustrates an RNN; it should typically show nodes connected in a way that indicates temporal or sequential processing. This might include feedback loops or connections where the output from one time step is used as the input for the next. The first neuron layer holds the power management, and balancing is examined at the mentioned time for the number of neurons, and it is denoted as o n . Scheduling is conducted for the number of resources that determine service distribution on the network. This evaluation step is used to deploy concurrent service sharing and resource access due to the dynamic user densities and point-of-contract-based infrastructure. Equation (6b) represents the second hidden layer that determines access to the end-user. The first layer indicates the queuing of the power allocation process, and balancing is conducted for resource access. The first hidden layer output is given as the input to the second neuron layer and processed to the number of neuron layers in the SDN.
From the first hidden layer, the data are forwarded to the second hidden layer’s first neuron and deploy the current state of the neuron. Here, the neuron’s current and the input state determine resource access and evaluate the better processing. In this computation step, congestion is addressed and decreased concerning the scheduling process, which results in a better training phase. The neuron layers initiate the training in the recurrent learning method, which estimates the concurrent sharing and access provided. The first and second hidden layers are denoted as 1   a n d   2 , respectively. The training data are enhanced from this hidden layer by equating the following equation:
X = r 0 N i d × b + ρ × q 0 + r n + v n M + S e t b = r 0 × w c + c k × M c k = r 0 + V M e t V = c k + w c v n + b
The above equation examines the training data that deploy the better detection of resources, and access is forwarded to the appropriate end-user on time. Here, the access is forwarded to the valid user in the network, which determines the concurrent processing and estimates scheduling and power allocation. Power allocation is performed for the different resources and deploys probabilistic balancing for the access and sharing of services. The derivation indicates the hidden layer process to enhance the training data for balancing resource sharing and estimating service distribution. Power management performs the allocation, which includes training for successive intervals.
The training data are determined for reliable sharing and service distribution to the end-user, and the power allocation is examined. The concurrency in service sharing and resource access is due to the dynamic user densities and point-of-contact based on the heterogeneous platform. Forwarding is performed to derive the balancing factor from the successive intervals of data processing. In this computation step, the prior state defines the current processing state. The evaluation is performed for the training data estimated from the hidden layers, and the queuing for valid access to resources is performed. The following equation distributes the service to the end-user, including power management:
= h a + c k s u v 0 × k 0 + V × c k w c + p i g t + o 0 × 1 + 2
The service distribution analyzes service distribution to the appropriate user better and deploys efficient processing. Here, congestion is addressed for power management, and scheduling for balancing data is estimated. Here, probabilistic balancing is performed for resource sharing, and the current state is used to define the prediction. Queuing is performed on the neuron layer, estimates the sharing of resources, and maintains power. The deficiency is maintained for service forwarding, and the hidden layers are estimated to improve the training data. The determination is achieved by examining the validation process and deploying the prior state.
The training state is represented as g t and the prior state of processing in the neuron layer is denoted as p i . Thus, the hidden layers are responsible for forwarding the data to the appropriate user by determining the resource sharing to estimate better sharing and resource access in IoT applications. The evaluation is represented as and avoids congestion for the scheduled resources in the SDN. The following equation is used to evaluate power management for the training data, and allocation is achieved for the training for the successive interval of resource processing.
M = 1 + 2 × c k g t S + v 0 × i d × b g = 1 + 2 × p i p i = γ × q 0 v n + k 0 γ = p i + w c ( h a + r n )
Power management is performed on the control plane by deploying the SDN in an IoT-based application. Here, power is provided to the required devices concerning the prediction state, γ , which determines the training data for reliable processing. Figure 5 illustrates the power management process of the proposed CS3.
This computation step examines queuing and scheduling to address cumulative congestion and power management. Service sharing is evaluated in the SDN for power allocation and service distribution. Balancing is achieved for the sharing and access provided to the appropriate device at the mentioned time. The power deficiency is addressed and decreased by performing training that deploys successive intervals in the network. Here, resource access is provided to the network and devices, and congestion is estimated. The following equation is used to predict the power usage of the prior device and prevent deficiency:
δ = 1 v n + r n × c k + w c S a 0 + g t × i d M + s u + b × p i e t
The analysis is carried out to predict and deploy access to the appropriate devices and users. Here, balancing is performed for the power management system, and power allocation for the trained data is deployed. Balancing is estimated for the prior and current states of the neuron layer and estimates the forwarding. The hidden layer is responsible for forwarding the data to the other neuron state, deriving the power allocation, and preventing deficiency. The analysis for the prediction is examined in the equation above, and it is denoted as δ . The equation below integrates joint scheduling and power allocation to improve power utilization and decrease latency:
δ z i = r n v n S + a 0 × k 0 + o 0 + + o n × h a + V r 0 × X + s u s u + γ e t
The analysis for utilization is examined in Equation (10); here, scheduling-related queuing is performed by deploying power management and resource allocation. In this computation step, the prediction is performed by mapping with the existing resources and providing relevant information sharing. The service distribution is performed for the trained and valid resources posted to the hidden layer processing. The analysis defines the current and next states of the neuron and estimates the reliable process that integrates scheduling and allocation and addresses the deficiency.
From this processing, latency and outage are addressed and decreased by improving power utilization, whereas deficiency is addressed. Based on the allocation, training is performed on time for successive intervals. This objective is addressed by introducing the CS3, which includes joint scheduling and power allocation. Thus, the SDN-based IoT application resolves the cumulative congestion of networks and devices and manages power. Access is provided on time in this processing concurrency in service sharing and resources. In Table 1, the failed requests in different scheduled intervals are presented.
In the SDN, Algorithm 3 shows the recurrent learning function that improves routing choices and flow conditions. Routing decisions and flow demands are initialized. It performs data transmission simulation, assesses network performance, modifies routing choices depending on metrics for performance, and modifies flow matrices iteratively. This recurrent process iterates until the maximum number of provided iterations is reached. In an SDN setting, the function produces optimal flow and route matrices that facilitate effective data transfer and resource allocation.
Algorithm 3: Recurrent learning
# Function to perform Recurrent Learning and Power Allocation
FunctionRecurrentLearningPowerAllocation ( r 0 , S , i d , q 0 , o 0 , 1 , 2 , v n , r n , g t , p i , c k , w c , γ , h a , X , e t )
Input: r 0 , S , i d , q 0 , o 0 , 1 , 2 , v n , r n , g t , p i , c k , w c , γ , h a , X , e t
Output: Power allocation, Service dissemination, Deficiency, Service distribution
Step 1: Recurrent Learning Phase
     k 0 u ( p i + r a S + q 0 o 0 )
Step 2: First Hidden Layer Calculation
    Initialize gradient array
        for   each   service   rate   r i   in   r do
         calculate   gradient   o i
       o i S + r 1 M + i d q 0 + v 1 + b e t
         add   o i to the gradient array
     return gradient array
Step 3: Second Hidden Layer Calculation
    Initialize gradient array
        for   each   index   i   from   0   t o   n 1 do
      Initialize sum = 0
          for   each   coefficient   c k do
          sum   s u m + v i c k r [ i ]
            calculate   gradient   o i
            o i = q i a 0 r i + s u m
          add   o i to the gradient array
    return gradient array
Step 4: Power Management
       Initialize   M = 0
       Initialize   g  = 0
       Initialize   γ = 0
        for   each   index   i   from   0   t o   n 1 do
         calculate   g
       g = 1 + 2 p i
         calculate   γ
       γ = p i + w c ( h a + r n ) ,
         calculate   M
     M Applsci 14 04951 i001 1 + 2 c k g t S + v 0 i d b + g + γ q 0 v n + k 0
        return   M
Step 5: Prediction and Analysis
       Initialize   δ = 0
     // Calculate the product of the terms inside the sigma
       s i g m a _ p r o d u c t = 1
       for   each   coefficient   c k   i n   c do
      s i g m a _ p r o d u c t = s i g m a _ p r o d u c t ( ( a 0 + g t ) ( ( ( i d   /   M )   /   ( p a r t i a l _ d e r i v a t i v e + s u ) ) + ( b p i ) ) )
       / /   Calculate   the   δ  value
       δ = 1 v n + r n s i g m a p r o d u c t e t
    return delta
Step 6: Service Dissemination Calculation
        Initialize   p a r t i a l _ d e r i v a t i v e = 0
         for   each   coefficient   c k   in   c do
     // Calculate the inner sum
       i n n e r _ s u m = 0
         for   each   coefficient   c k   in   c do
       i n n e r _ s u m = i n n e r _ s u m + ( w c + p i )
    // Calculate the term inside the parenthesis
       t e r m _ i n s i d e _ p a r e n t h e s i s = h a + c k s u v 0 ( ( k 0 + V ) ( i n n e r _ s u m / g t ) )
     // Add the contribution of each term to the partial derivative
       p a r t i a l _ d e r i v a t i v e = p a r t i a l _ d e r i v a t i v e + t e r m _ i n s i d e _ p a r e n t h e s i s
       / /   Add   the   contribution   of   o 0 ( 1 + 2 )  to the partial derivative
       p a r t i a l _ d e r i v a t i v e = p a r t i a l _ d e r i v a t i v e + o 0 ( 1 + 2 )
       return   p a r t i a l _ d e r i v a t i v e
Step 7: Output Results
        Return   ,   δ
End Function
In discussing the performance of the scheduling system, Table 1 provides crucial insights into the system’s behavior across various scheduling intervals. The table presents data on queued requests, service dissemination success rates, and the number of failed requests, which are vital metrics for evaluating the effectiveness of the scheduling algorithm. The sharing interval is determined for the varying scheduling intervals for the queued request and increases if the shared interval increases. Suppose the queued request decreases and the service dissemination increases for varying intervals. In another case, if the queued request decreases, the failure request also decreases (Table 1). The iterates utilized at different sharing intervals are tabulated in Table 2.
The goal of the suggested method is to minimize latency while ensuring adequate power distribution to every user. Allocated power is computed based on an optimization model that distributes the total available power and total among users to reduce the overall latency. Required power signifies the theoretical power necessary to achieve a target latency for every user or sharing interval. The sharing interval is estimated for the power requirement and allocation and shows less processing than the power requirement. If the power allocation decreases, the failed request from the user is enhanced. The training iterates define the sharing intervals and increase if the failed request increases (Table 2). In Table 3, the service dissemination factor for different factors is presented.
User density determines power requirement, and it is enhanced appropriately. The prediction is made by mapping the current and prior service and range increases. If the prediction increases, the deficiency decreases, and service dissemination is performed for the different types of users in the network. Service dissemination also increases if the deficiency is higher (Table 3). The article suggests the CS3, an advanced SDN resource management method. By integrating the SDN’s adaptable control and data plane operations with deep learning’s predictive capabilities, the CS3 efficiently handles high-user-request densities while minimizing latency. Network performance and customer satisfaction are enhanced by integrating these technologies, which provide efficient and resilient power allocation and service scheduling.

4. Performance Assessment

This sub-section assesses the proposed CS3’s performance through comparative analysis. The simulation experiments are modeled using the Contiki Cooja simulator, where 200 mobile users are placed. An IoT cloud architecture with seven service providers, each with 1 TB of storage, is used for service distribution. The transmit power of the devices is set as 30 dBm, and it operates in 60 scheduling intervals. With these simulation details, the performance is assessed using service dissemination factor, latency, service failure ratio, outage, and power utilization metrics. In the comparative analysis, the the Iterative Searching-based Algorithm for Task Offloading and Transmit Power Allocation (ISA-TOTPA) [17], Energy and Time-Efficient Computation Offloading and Resource Allocation (ETCORA) [16], and DBWA [18] methods are used. This study extended our simulation experiments using the Contiki Cooja simulator to include a broader range of scenarios. These scenarios now encompass different numbers of mobile users (50, 100, 200, and 300), varying mobility patterns, and diverse traffic conditions. This comprehensive evaluation aims to provide a more detailed understanding of the system’s performance under varying circumstances. Our results indicate that our approach consistently outperforms others regarding latency and power allocation. Moreover, extended-duration tests reveal the long-term stability of our system.

4.1. Service Dissemination

The service dissemination for the proposed work is high for varying users and scheduling intervals. Here, power management determines appropriate user forwarding at the mentioned time. Power allocation is estimated for probabilistic balancing and examines the training data for the successive interval, and it is represented as s u + i d × N + r n X a 0 . Service distribution is performed by evaluating IoT sharing and resource access. The computation is conducted on the SDN-based IoT application, ensuring valid access. Sharing is derived from the number of users in the network and better deploys power allocation. The current state estimates scheduling based on the queuing process and deployed probabilistic learning. Here, service utilization is performed for service sharing, and resource access is achieved for the training data. In this case, prediction is performed for the prior and current states and provides access to the required resources. Sharing is conducted by determining the point-of-contact-based infrastructure that performs concurrent service sharing and resource access. Scheduling is performed to initialize devices and power allocation is carried out during this process (Figure 6). The different fluctuations in the service dissemination comparison curve in Figure 6 likely result from a combination of factors related to network dynamics, resource management, algorithm efficiency, user behavior, and environmental conditions.

4.2. Latency

Latency decreases concerning user and scheduling intervals, which deploys the service distribution. Here, scheduling and queuing are derived for the control data plane in the network, indicating power management. Based on the scheduling process, queuing is performed for the resources and valid access is forwarded on time. The power allocation is performed for probabilistic balancing, and it is represented as X × q 0 + r 0 D V . Concurrent service sharing and resource access determine the neuron state and deploy the hidden layers. The analysis is estimated for service distribution for the numbers of users and deploys the sharing efficiently. The training devices are used to deploy the scheduling for power management and determine the current state. The predictions are made for resource access, estimated power allocation, and derived scheduling. Joint scheduling is performed for service distribution to determine the SDN-based IoT application. Here, service sharing is conducted by determining the neuron’s current state and evaluating the probabilistic balancing. Integration is achieved for joint scheduling and power allocation and determines better processing. The hidden layer derives reliable service sharing and addresses cumulative congestion and power management (Figure 7). Figure 7 shows latency over different scheduling intervals, and a smaller curve for the CS3 indicates consistently lower latency across all intervals tested.

4.3. Service Failure

Service failure is achieved by determining probabilistic balancing for the service distribution and deriving utilization. Utilization is improved for the valid user and examines the service forwarding, and it is represented as ρ × q 0 + r n + v n M + S . The current state matches the prior state and estimates the service distribution for the number of users. Balancing is performed for resource access, and utilization is determined for efficient processing. Recurrent learning derives training data for efficient service sharing and determines congestion. Here, the analysis is performed for probabilistic learning, and access is derived. Here, concurrent processing is conducted to access the appropriate user network for service sharing and resource access. The analysis estimates power management by balancing the user’s service and deploying sharing. Here, power allocation is achieved for reliable processing and deploys better utilization. Utilization is improved for reliable processing between the user and the devices, and it is evaluated as c k w c + p i g t . Latency is addressed by decreasing the failure transmission, which estimates the service distribution to the valid user. Access is provided to valid resources by determining the resource allocation for power management (Figure 8).

4.4. Outage

In Figure 9, the outage is decreased for the varying users and scheduling intervals, and the balancing of resources is determined. Sharing is derived from reliable resources and provides access to valid devices. The current state is used to define better processing and derive the mapping with the prior state, and it is represented as c k g t S + v 0 × i d . The prediction is performed for the prior state and estimates the joint scheduling and power allocation integration. Concurrent service sharing and resource access determine the probabilistic balancing. The evaluation is performed for power management and determines service distribution. The distribution is achieved to share the service and deploy the SDN. Balancing is performed to determine the current state and provides reliable processing. Training is conducted for the utilization process, and scheduling is estimated. Power management is used to derive reliable processing and determine the balancing. Power allocation is performed to address the outage and estimate congestion. Congestion is addressed by evaluating the prediction for the prior state with the current state. The service distribution is performed by deriving scheduling, and that is performed by performing queuing. Thus, the outage is reduced for the number of users and the service distribution.

4.5. Power Utilization

In Figure 10, power utilization is decreased by estimating the power allocation for the number of users and resources. Scheduling is performed for the queued devices, which deploy the current state and derive the neuron layers. The hidden layer is used to derive access to resources without congestion, and it is represented as i d M + s u + b × p i . The analysis is performed to determine probabilistic balancing. Here, the network and device are used concurrently, and service sharing, and resource access are determined. Probabilistic balancing is conducted for service sharing and examines the power allocation, and it is computed as X + s u s u . Here, the prior state matches the current state and provides the power management for the allocated resources. In this case, service sharing and resource access are provided concurrently for efficient processing. In this computation step, the current state determines the prediction with the prior state and provides reliable sharing. Sharing is achieved for the number of resources and determines the training service by performing hidden layers. If the utilization increases, the failure and outage of this proposed work will decrease. This utilization concerns users and scheduling intervals and shows less processing than the existing three methods.
The power allocation and latency observed for different users and sharing intervals are tabulated in Table 4.
In the article, Table 4 serves a crucial role in presenting empirical data related to the performance of the proposed system under varying conditions. Specifically, it displays the power allocation and latency metrics for different numbers of users and sharing intervals. This table is likely used to illustrate the efficiency and scalability of the system in terms of power distribution and response time as the number of users and the frequency of sharing intervals change. Service sharing is performed for power allocation and latency; latency decreases, and allocation is improved. Compared to user 50, power allocation for user 200 shows less service sharing. In another case, latency is estimated and decreases for user 50 compared to user 200. Both the power allocation and latency increase for the sharing of services (Table 4). Table 5 and Table 6 present the comparative analysis summaries for users and scheduling intervals.

5. Conclusions

This article presents a CS3 for power and service management in an SDN-based IoT. This system uses resource scheduling to meet user request density and minimize latency. The SDN monitors user-resource queuing to detect response times or availability disruptions. Recurrent learning is used to probabilistically examine the service requirement and its queuing, which is the foundation for additional service distributions. The SDN’s control and data plane operations handle scheduling and power management among various resources. The deep learning model in this method guaranteed the requisite power allocation and failure-free training of neurons for queuing. Schedules and subsequent service distributions are scheduled based on the residual and insufficient power. Consequently, the suggested technique controls response failures independent of scheduling intervals and user density. The proposed plan delivers a 14.9% high-service distribution for various users, 18.29% less delay, 13.34% less failure, 5.54% less downtime, and 18.68% less power consumption. However, the limitation of this proposed method is that it is not appropriate for large environments and increases in computing time. Future studies will explore large-scale and resource-constraint IoT scenarios with real data to meet user requests.

Author Contributions

Conceptualization, M.A. and H.M.; methodology, M.A. and H.M.; software, M.A. and H.M.; validation, Z.A. and H.M.; formal analysis, M.A. and Z.A.; resources, Z.A. and H.M.; data curation, Z.A.; writing—original draft preparation, M.A. and H.M.; writing—review and editing, M.A. and Z.A.; visualization, Z.A. and H.M.; funding acquisition, Z.A. and H.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education (No. 2021R1F1A1055408).

Data Availability Statement

Data is contained within the article.

Acknowledgments

The authors express their sincere appreciation to the Researcher Supporting Project Number (RSPD2024R1113) King Saud University, Riyadh, Saudi Arabia.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Rahman, A.; Islam, M.J.; Rahman, Z.; Reza, M.M.; Anwar, A.; Mahmud, M.P.; Noor, R.M. Distb-condo: Distributed blockchain-based IoT-SDN model for bright condominium. IEEE Access 2020, 8, 209594–209609. [Google Scholar] [CrossRef]
  2. Wu, T.; Zhou, P.; Wang, B.; Li, A.; Tang, X.; Xu, Z.; Chen, K.; Ding, X. Joint Traffic Control and Multi-Channel Reassignment for Core Backbone Network in SDN-IoT: A Multi-Agent Deep Reinforcement Learning Approach. IEEE Trans. Netw. Sci. Eng. 2020, 8, 231–245. [Google Scholar] [CrossRef]
  3. Kim, Y.; Song, C.; Han, H.; Jung, H.; Kang, S. Collaborative Task Scheduling for IoT-Assisted Edge Computing. IEEE Access 2020, 8, 216593–216606. [Google Scholar] [CrossRef]
  4. El-Hosseini, M.; ZainEldin, H.; Arafat, H.; Badawy, M. A fire detection model based on power-aware scheduling for IoT-sensors in smart cities with partial coverage. J. Ambient Intell. Humaniz. Comput. 2021, 12, 2629–2648. [Google Scholar] [CrossRef]
  5. Shabbir, G.; Akram, A.; Iqbal, M.M.; Jabbar, S.; Alfawair, M.; Chaudhry, J. Network performance enhancement of multi-sink enabled low power lossy networks in SDN based Internet of Things. Int. J. Parallel Program. 2020, 48, 367–398. [Google Scholar] [CrossRef]
  6. Kharwar, P.; Yadav, S.; Purohit, N. Optimal power allocation for the multiple-sensor single-DF relay-based IoT networks under transceiver hardware impairments. Phys. Commun. 2020, 43, 101183. [Google Scholar] [CrossRef]
  7. Shukla, M.K.; Nguyen, H.H.; Pandey, O.J. Multiuser Full-Duplex IoT Networks with Wireless-Powered Relaying: Performance Analysis and Energy Efficiency Optimization. IEEE Trans. Green Commun. Netw. 2020, 4, 982–997. [Google Scholar] [CrossRef]
  8. Singh, G. Research and implement solutions to tackle security threats in Mobile Cloud Gaming. Master’s Thesis, University of Alberta, Edmonton, AB, Canada, 2023. [Google Scholar]
  9. Zia, K.; Chiumento, A.; Havinga, P.J. AI-enabled reliable QoS in multi-RAT wireless IoT networks: Prospects, challenges, and future directions. IEEE Open J. Commun. Soc. 2022, 3, 1906–1929. [Google Scholar] [CrossRef]
  10. Chakraborty, C.; Rajendran, S.R.; Rehman, M.H. Security of Internet of Things Nodes; Taylor & Francis Ltd: London, UK, 2021. [Google Scholar]
  11. Bekri, W.; Jmal, R.; Chaari Fourati, L. Internet of things management based on software defined networking: A survey. Int. J. Wirel. Inf. Netw. 2020, 27, 385–410. [Google Scholar] [CrossRef]
  12. Ai, Z.; Zhang, M.; Zhang, W.; Kang, J.; Tong, L.; Duan, Y. Survey on the scheme evaluation, opportunities and challenges of software defined-information centric network. IET Commun. 2023, 17, 2237–2274. [Google Scholar] [CrossRef]
  13. Mutruc, V. User Centric Quality of Service Improvement in IoT Networks: A Closed Loop Approach. Bachelor’s Thesis, University of Twente, Enschede, The Netherlands, 2023. [Google Scholar]
  14. Imran; Ghaffar, Z.; Alshahrani, A.; Fayaz, M.; Alghamdi, A.M.; Gwak, J. A topical review on machine learning, software defined networking, internet of things applications: Research limitations and challenges. Electronics 2021, 10, 880. [Google Scholar] [CrossRef]
  15. Cui, X.; Gao, X.; Ma, Y.; Wang, W. A controller deployment scheme in 5G-IoT network based on SDN. EURASIP J. Wirel. Commun. Netw. 2020, 2020, 248. [Google Scholar] [CrossRef]
  16. Sun, H.; Yu, H.; Fan, G.; Chen, L. Energy and time-efficient task offloading and resource allocation on the generic IoT-fog-cloud architecture. Peer-Peer Netw. Appl. 2020, 13, 548–563. [Google Scholar] [CrossRef]
  17. Guo, H.; Zhang, J.; Liu, J.; Zhang, H. Energy-aware computation offloading and transmit power allocation in ultradense IoT networks. IEEE Internet Things J. 2018, 6, 4317–4329. [Google Scholar] [CrossRef]
  18. Guo, M.; Li, L.; Guan, Q. Energy-efficient and delay-guaranteed workload allocation in IoT-edge-cloud computing systems. IEEE Access 2019, 7, 78685–78697. [Google Scholar] [CrossRef]
  19. Samanta, A.; Tang, J. Dyme: Dynamic microservice scheduling in edge computing enabled IoT. IEEE Internet Things J. 2020, 7, 6164–6174. [Google Scholar] [CrossRef]
  20. Wang, Q.; Wu, Z. Beamforming Optimization and Power Allocation for User-Centric MIMO-NOMA IoT Networks. IEEE Access. 2020, 9, 339–348. [Google Scholar] [CrossRef]
  21. Wang, K. Energy-efficient resource allocation optimization algorithm in industrial IoTs scenarios based on energy harvesting. Sustain. Energy Technol. Assess. 2021, 45, 101201. [Google Scholar] [CrossRef]
  22. Aburukba, R.O.; Landolsi, T.; Omer, D. A heuristic scheduling approach for fog-cloud computing environment with stationary IoT devices. J. Netw. Comput. Appl. 2021, 180, 102994. [Google Scholar] [CrossRef]
  23. Umer, A.; Ali, M.; Jehangiri, A.I.; Bilal, M.; Shuja, J. Multi-Objective Task-Aware Offloading and Scheduling Framework for Internet of Things Logistics. Sensors 2024, 24, 2381. [Google Scholar] [CrossRef]
  24. Singh, A.P.; Luhach, A.K.; Gao, X.Z.; Kumar, S.; Roy, D.S. Evolution of wireless sensor network design from technology centric to user centric: An architectural perspective. Int. J. Distrib. Sens. Netw. 2020, 16, 1550147720949138. [Google Scholar] [CrossRef]
  25. Paulson, E.N.; Yusof, K.M.; Marsono MN, B.; Dauda, U.S.; Kofoworola, F. Quality of Service Evaluation of Software Defined Internet of Things Network. ELEKTRIKA-J. Electr. Eng. 2021, 20, 65–75. [Google Scholar] [CrossRef]
  26. Huang, X.; Wu, W.; Hu, S.; Li, M.; Zhou, C.; Shen, X.S. Digital twin-based user-centric resource management for multicast short video streaming. IEEE J. Sel. Top. Signal Process. 2023, 18, 50–65. [Google Scholar] [CrossRef]
  27. Pervej, M.F.; Jin, R.; Lin, S.C.; Dai, H. Efficient content delivery in user-centric and cache-enabled vehicular edge networks with deadline-constrained heterogeneous demands. IEEE Trans. Veh. Technol. 2023, 73, 1129–1145. [Google Scholar] [CrossRef]
  28. Abir, M.A.B.S.; Chowdhury, M.Z.; Jang, Y.M. Software-Defined UAV Networks for 6G Systems: Requirements, Opportunities, Emerging Techniques, Challenges, and Research Directions. IEEE Open J. Commun. Soc. 2023, 4, 2487–2547. [Google Scholar] [CrossRef]
  29. Raeisi, M.; Sesay, A.B. User-Centric Channel Allocation Scheme for 5G High-Speed Users by Utilizing Machine Learning Algorithm to Reduce Handover Rate. IEEE Access 2023, 11, 84113–84133. [Google Scholar] [CrossRef]
  30. Murshed, M. Reinforcement Learning-Based User-Centric Handover Decision-Making in 5G Vehicular Networks. Master’s Thesis, Brock University, St. Catharines, ON, Canada, 2024. [Google Scholar]
  31. Wang, W.; Zhang, Y.; Gu, J.; Wang, J. A proactive manufacturing resources assignment method based on production performance prediction for the smart factory. IEEE Trans. Ind. Inform. 2021, 18, 46–55. [Google Scholar] [CrossRef]
Figure 1. Representation of the proposed CS3.
Figure 1. Representation of the proposed CS3.
Applsci 14 04951 g001
Figure 2. Scheduling process illustration.
Figure 2. Scheduling process illustration.
Applsci 14 04951 g002
Figure 3. CS3 queuing process.
Figure 3. CS3 queuing process.
Applsci 14 04951 g003
Figure 4. Recurrent layer process.
Figure 4. Recurrent layer process.
Applsci 14 04951 g004
Figure 5. CS3 power management process.
Figure 5. CS3 power management process.
Applsci 14 04951 g005
Figure 6. Service dissemination comparisons.
Figure 6. Service dissemination comparisons.
Applsci 14 04951 g006
Figure 7. Latency comparisons.
Figure 7. Latency comparisons.
Applsci 14 04951 g007
Figure 8. Failure ratio comparisons.
Figure 8. Failure ratio comparisons.
Applsci 14 04951 g008
Figure 9. Outage comparisons.
Figure 9. Outage comparisons.
Applsci 14 04951 g009
Figure 10. Power utilization comparisons.
Figure 10. Power utilization comparisons.
Applsci 14 04951 g010
Table 1. Failed requests in scheduling intervals.
Table 1. Failed requests in scheduling intervals.
Scheduling
Intervals
Sharing
Intervals
Queued
Requests
Service
Dissemination
Failed
Requests
1011230.960
2019250.9219
3015200.9412
4025750.9320
5038980.9125
60491030.8736
Table 2. Training iterates for sharing intervals.
Table 2. Training iterates for sharing intervals.
Sharing
Intervals
Power
Requirement (J)
Power
Allocation (J)
Failed
Requests
Training
Iterates
1080.3730.139107
2080.5440.583498
3078.1446.343275
4098.3654.252968
50102.8760.482238
60107.9875.372041
70111.2580.691839
80118.891.331223
Table 3. Service dissemination for different factors.
Table 3. Service dissemination for different factors.
UsersPower
Requirement (J)
Predicted (J)Deficiency (J)Service
Dissemination
4080.3786.5−6.130.87
8082.3693.58−11.220.89
12092.4786.745.730.96
160102.58119.47−16.890.91
200118.8126.3−7.50.94
Table 4. Power allocation and latency for different users and sharing intervals.
Table 4. Power allocation and latency for different users and sharing intervals.
Sharing
Intervals
Users = 50Users = 100Users = 150Users = 200
Power
Allocation
Latency
(ms)
Power
Allocation
Latency
(ms)
Power
Allocation
Latency
(ms)
Power
Allocation
Latency (ms)
1033.47219.738.25360.2140.12420.1270.15541.89
2035.48220.3642.36450.2345.14480.3682.36596.37
3042.15360.2351.25521.3651.36512.36102.54698.47
4045.69390.3658.69584.1255.47520.98135.14745.21
5049.69412.0163.25612.3669.36674.14145.21847.26
6052.14436.1471.25630.1285.21695.48181.36921.47
7063.14470.2373.69624.12121.14714.00196.32963.47
8065.47490.2378.69710.25151.32894.12192.36971.05
Table 5. Comparative analysis for users.
Table 5. Comparative analysis for users.
MetricsISA-TOTPAETCORADBWACS3Findings
Service Dissemination0.8840.8930.9140.946614.9% High
Latency (ms)971.04802.36629.13361.38818.29% Less
Failure (%)12.3210.347.565.62713.34% Less
Outage0.0530.0450.0330.02525.54% Less
Power Utilization (J)238.91185.24124.480.38418.68% Less
Table 6. Comparative analysis for scheduling intervals.
Table 6. Comparative analysis for scheduling intervals.
MetricsISA-TOTPAETCORADBWACS3Findings
Service Dissemination0.8990.9250.9430.95559.95% High
Latency (ms)968.3778.83642.79363.0115.14% Less
Failure (%)12.5110.418.875.9813.85% Less
Outage0.0520.0430.0350.02515.47% Less
Power Utilization (J)248.19178.02125.8385.72617.8% Less
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Anjum, M.; Min, H.; Ahmed, Z. User-Centric Internet of Things and Controlled Service Scheduling Scheme for a Software-Defined Network. Appl. Sci. 2024, 14, 4951. https://0-doi-org.brum.beds.ac.uk/10.3390/app14114951

AMA Style

Anjum M, Min H, Ahmed Z. User-Centric Internet of Things and Controlled Service Scheduling Scheme for a Software-Defined Network. Applied Sciences. 2024; 14(11):4951. https://0-doi-org.brum.beds.ac.uk/10.3390/app14114951

Chicago/Turabian Style

Anjum, Mohd, Hong Min, and Zubair Ahmed. 2024. "User-Centric Internet of Things and Controlled Service Scheduling Scheme for a Software-Defined Network" Applied Sciences 14, no. 11: 4951. https://0-doi-org.brum.beds.ac.uk/10.3390/app14114951

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop