A Hierarchical Optimized Resource Utilization based Content Placement (HORCP) model for cloud Content Delivery Networks (CDNs)

Content Delivery Networks (CDNs) have grown in popularity as a result of the ongoing development of the Internet and its applications. The workload on streaming media service systems can be significantly decreased with the help of the cooperative edge-cloud computing architecture. In the traditional works, a different types of content placement and routing algorithms are developed for improving the content delivery of cloud systems with reduced delay and cost. But, the majority of existing algorithms facing complexities in terms of increased resource usage, ineffective delivery, and high system designing complexity. Therefore, the proposed work aims to develop a new framework, named as, Hierarchical Optimized Resource Utilization based Content Placement (HORCP) model for cloud CDNs. Here, the Chaotic Krill Herd Optimization (CKHO) method is used to optimize the resource usage for content placement. Then, a Hierarchical Probability Routing (HPR) model is employed to enable a dependable end-to-end data transmission with an optimized routing path. The performance of the proposed HORCP model is validated and compared by using several performance metrics. The obtained results are also compared with current state-of-the-art methodologies in order to show the superiority of the proposed HORCP model. By using the HORCP mechanism, the overall memory usage of the network is reduced to 80%, CPU usage is reduced to 20%, response is minimized to 2 s, and total congestion cost with respect to the network load level is reduced to 100.


Introduction
Over the past few decades, the urban population growth has been unprecedented throughout the world [1].According to studies, the number of people living in cities worldwide has increased by roughly 60 million annually, and by 2050, it's predicted that 70% of people intend to do so.The quality of experience (QoE) [2,3] of consumers as well as the data needs of cities may both be met by content delivery networks (CDNs).A network of millions of internet-connected devices is known as a content delivery network [4].Here, a service provider's limited supply of servers are linked together all over the world.Due to the employment of numerous servers, the load distribution should be preserved, significantly assisting the user in receiving high-quality material faster.A CDN [5] is made up of an origin server and a set quantity of proxy servers.In a CDN, duplicate copies of the content are created on the origin server and kept on the proxies of those servers [6,7].Proxy servers are built in a number of different places throughout the world.If a client simply requests a specific piece of online content, the request information is first transmitted to the geographically local proxy server, which then checks the content's availability.Typically, they have emerged as a result of the growth in Internet users and their desire for low-latency delivery of content.The goal of CDN [8][9][10] is to put the physical responsive server as close to the user as possible.
This concept is implemented utilizing an arrangement of proxy servers and cache providers offered by the content provider, with the last level often situated at the edge of the user network.One of the most important user demands is for quick content delivery, and CDN is often referred to as an improved kind of web caching that satisfies this requirement.However, due to their high deployment costs, traditional CDN systems [11,12] are difficult to scale.The emergence of cloud computing has made it possible to create CDNs in the cloud, or cloud CDNs, by leasing resources (such memory and bandwidth).It can make use of the cloud's versatility to disseminate content over the Internet swiftly and easily.While reducing the cost of data storage and their delivery, cloud CDNs can increase capacity, adaptability, and elasticity [13].One of the fundamental technologies of cloud CDNs is content placement.By creating multicast delivery models, the content providers typically decrease the number of replicas in order to cut down the content placement costs.However, the majority of content delivery methodologies are focused on static networks and, are unable to migrate the dynamic features of cloud proxy servers in CDNs [14,15].Furthermore, one of the primary causes of the conventional content placement methods areincreased network congestion and high cost.Determining how to lower the cost of content placement is still a difficult topic in light of these dynamic cloud CDNs.Furthermore, the conventional content placement methods do not take into account the global dynamics of the congestion in cloud CDNs and only give delivery paths based on local decision-making [16][17][18].Because of these dynamic properties, it remains a difficult task to figure out how to lower the cost of content placement.Therefore, the proposed work aims to use an effective content placement model for cloud CDNs [19].Also, the proposed content placement methodology can effectively distribute material and balance the load on dynamic networks better than the classic routing techniques.The major objectives of the proposed work are as follows: In the proposed work, the HORCP mechanism is developed with the CKHO technique for enabling a successful and reliable content delivery in cloud systems.The performance and effectiveness of the proposed HORCP model is validated and assessed by using a variety of parameters including congestion cost, response time, and hit ratio.Moreover, the findings demonstrate that the proposed HORCP provides the reduced congestion cost and time while enabling a successful content delivery in cloud networks.
The remaining sections of the paper are divided into the following categories: The thorough literature overview on the current content delivery techniques used in traditional cloud systems, along with their positive and negative aspects, is provided in Sect. 2. The suggested HORCP paradigm is clearly explained in Sect. 3 along with an overall flow diagram and stage-by-stage justifications.Additionally, various parameters are used to validate the proposed HORCP model's performance and comparative outcomes.

Related works
This section presents an overview of existing content placement and routing algorithms utilized in cloud CDNs.Also, it analyses the advantages and disadvantages of each model based on how each model performs routing activities.
Zolfaghari, et al. [20] presented a comprehensive survey to examine the recent trends and challenges in the CDN.Here, the clear lifecycle of CDN is provided that includes the stages of system design, implementation, deployment, operation, and evaluation.Moreover, the authors investigated the major operations of content distribution, requisition-based routing and performance management.In addition, the authors intended to improve the parameters of cost, scalability, security, and privacy while designing an effective CDNs [21].Implemented a Lyapunov optimization algorithm for ensuring a reliable information sharing in cloud based CDNs.In this framework, the goal of each cloud provider is to increase its expected payout, which is calculated as the reciprocal of the weighted average of the anticipated data transfer cost and latency for its customers [22].The computerized infrastructure of the CDN is managed by an individual content provider, which tracks the subscriber content requests and, assigns wireless channels/ nodes to send the required content for minimizing the overall content transfer costs and delays.As a result, the distribution of material within a CDN can be described as an optimization challenge [23].Formulated a new CDN architecture for enhancing the Quality of Experience (QoE) of smart city networks.The authors suggested a brand-new cellular network infrastructure for a smart city that includes content delivery and resource management.Due to the restricted processing capacity of mobile edge computing, the resource allocation gets an adequate computing equilibrium for an array of services and application.The content delivery portion is more concerned with network cache constraints and transmission delay.The greedy algorithm and the suggested framework can adapt to a variety of user requests and communication situations as demonstrated in the numerical outcomes [24].Additionally, a comprehensive system for multiple content delivery has been developed for different service models with guaranteed user satisfaction.
Sadeghi, et al. [25] utilized a deep reinforcement learning algorithm for enhancing the content delivery of hierarchical networks.Caching is expected to be a key component of internet topologies, mobile networks, and next-generation CDNs.Caching can help the network infrastructure as well as end users by strategically storing the majority of popular contents at the storage-enabled network entities during off-peak demand instances.The authors of this paper indicated that the caching techniques are more useful for the next-generation networks, hence it is essential to take into account for solving the problems network caching, complex dynamics, and state space modeling.The major drawbacks of this work are increased complexity in system design, and low efficiency.Sinky, et al. [26] formulated a content centric framework for enhancing caching and placement operations in the smart city networks.To be more precise, the authors make use of the collaborative filtering theory for providing precise and effective content popularity estimations that support proactive in-network caching of web data.Here, the popularity driven content caching model is used to improve the QoE, where the following functions are computed content popularity, availability, population density, storage capability of node, and delay [27].Implemented a Q-Learning based content placement algorithm for cloud systems.Here, an adaptive delivery tree is formulated to take the better routing decisions in the cloud CDNs.The learning packets in this approach can deduce the local and nonlocal congestion information.The regional traffic data shows how long it takes a packet to go from one cloud proxy server to another.The worldwide perspective of the latency issues from the present cloud proxy server to the terminating proxy server is provided by the non-local traffic data.Every cloud proxy server keeps the data in a Q-table and updates it using learning packets.
Alghamdi, et al. [28] Designed a new fog-cloud architecture model to improve the performance of CDNs.Here, a popularity-based caching strategy is used along with an Optimized Link State Routing (OLSR) protocol for CDNs.This study indicated that the factors such as security, node mobility, node failure rate, and scalability should be addressed for assuring a high quality of content delivery in the cloud systems.The primary advantages of this work are increased scalability and efficiency.Zhao, et al. [29] investigated about some of the recent developments in the field of CDNs.When designing a CDN, the operation costs are a crucial element that should be taken into account.Although some tactics, like expanding server cache capacity, can enhance CDN performance.But, they undoubtedly raise the service providers' operating costs, adding to their burden.Therefore, it is crucial for CDN providers to lower costs and delay.Moreover, the different types of caching mechanisms are reviews in this study for an effective server cache memory management [30].Introduced a new Information Centric Networking (ICN) paradigm for improving the storage performance of cloud CDNs.The suggested content popularity mechanism increases the content availability at the proximity devices with reduced transfer time and packet loss ratio.Moreover, the clustering based efficient caching mechanism offers a suitable solution to the issue of the existing hash and on-path caching mechanisms.
The literature research [31][32][33] revealed that the current content placement and routing algorithms mostly focused on lowering CDN costs and enhancing content delivery.The bulk of techniques are having issues with excessive resource use, dependability issues, and accelerated delay.Therefore, the proposed research work aims to develop a new framework for cloud content placement in CDNs.Table 1 discusses about the list of existing methodologies used in the field of CDN with their findings.

Proposed methodology
This section provides the clear explanation for the proposed cloud content delivery system with the workflow and appropriate illustrations.The original contribution of this paper is to perform an effective content placement and routing with the low time and cost complexities.with the use of hierarchical optimization based routing model.In order to optimally the resource usage, a CKHO technique is implemented, which provides the best optimum solution to simplify the process of resource allocation.By using these processes, the optimized content placement is done in the proposed system with reduced cost and latency parameters.

Content placement
In the proposed HORCP framework, an effective content placement is performed with the reduced cost and latency factors.Typically, the content placement in cloud is one of the most prominent issues need to be addressed for enabling a reliable and effective data transmission.The system for streaming media services has faced enormous hurdles as a result of the increased demand from customers for video streaming services [34,35].The workload on streaming media service systems can be significantly reduced by the cooperative edge-cloud computing design, which combines edge computing and cloud computing.The cooperative edge-cloud computing architecture's streaming content caching is crucial for enhancing user service quality.This led the authors of this research to put forward a HORCP strategy based on user mobility and popular content.In this model, the mobile users submit enormous amounts of data to cloud data centers and edge servers as the source data generators.To decrease data transmission delay and traffic load, the edge servers use a portion of the storage space to cache specific information, particularly for video stream files with strict latency requirements.In this architecture, the user is able to access the contents that are cached in the edge server.The user will send the content request to the edge servers whenever they need to access the data.The content requests made by mobile users can be handled by the edge servers, which have constrained computation and storage capabilities, or they can be forwarded to the cloud data center.Users can access robust but delayed compute and storage capabilities from the cloud data center by sending content requests to it from edge servers.The cooperating cache domain, which is used to exchange the contents of each other's caches, is formed by pooling the cache spaces of all edge servers.In this technique, the number of edge servers E s , steam- ing video S M , and available set of users U are considered as the input, and the predictive value for content placement is delivered as the output.At first, the input parameters are initialized E s = {1, 2, . . ., S r } , S M = {1, 2, . . ., M} , and u = {1, 2, . . ., U} , where each file is split into multiple encoding segments with the fixed length.After that, the request rate of encoding bits of the edge server are estimated, where the total number of times R u s that each user accesses the edge server's encoding segments S M as shown in the following model: where, ϕ denotes each user, f denotes each content, φ is the number of contents, R u s represents the encoding segments of the file at number of times, and S M denotes the encoding segments.Consequently, the frequency for user u in the accessing file S r from the edge server s is esti- mated by using the following equation: where, N ρ represents the time duration and F u j is the frequency of user.Furthermore, the average popularity of file f at the edge server is estimated based on the follow- ing model: After computing the average population of file, the probability of user located on the edge server at the time period (t + 1) is also estimated as shown in below: where, c represents the count of probability for each user U .Finally, the predictive value for the content placement is obtained according to the number of users moving to the edge server at the time period (t + 1) as shown in below: Based on the F u j (t + 1) value, the content is placed in the cloud systems for an efficient and reliable routing.After computing the time cost, the resource usage local on the edge server is optimally computed by using the KHO algorithm, which helps to obtain the low latency and cost factors while placing contents on the cloud. (1) (2)

Chaotic Krill-Herd Optimization (CKHO)
The resource utilization load on the edge is optimized during content placement using the CKHO algorithm's solution.Different optimization algorithms are used in conventional works to solve complex problems.The CKHO technique has the major benefits of increased convergence rate, reduced local optimum, simplicity, adapts for parallel computing and high searching speed.In this technique, each krill's distance from the density of the krill swarm and food determines the objective function for krill movement.Moreover, each krill's position is made up of three components: natural dispersion, foraging motion, and motion induced by other krills.The three motions mentioned above in CKHO can be simplified according to the following Lagrangian model.Here, the target effect, local effect, and repulsive effect are roughly calculated as the three factors that will determine the trajectory of the first motion.The flow of the CKHO algorithm is shown in Fig. 3.In this technique, the parameters such as generalization counter G c , and krill population ρ k are considered as the inputs, and the optimal value ρ k is produced as the output.According to the initial position, the fitness function is estimated based on the following model: where, { t u -frequency of user u and a t u -Average popularity of content.Until reaching the maximum number of iterations, the population is sorted as shown in below: Then, the motion updated is performed by using the following equation: (7) ,

Fig. 3 Flow of CKHO algorithm
Where, M max indicates the maximum speed, w n rep- resents the inertia weight in [0, 1], M old j is the previous motion, ε total j and ε target j are the local and target effects respectively.Moreover, the motion induced by other individuals according to the foraging behavior is shown in below: Where, S f is the foraging speed, ω j is the inertia weight between 0 and 1, F old j is the previous foraging motion, δ food j is the food attraction and δ best j is the effect of the best fitness.Then, the physical diffusion is performed, where the model of motion is represented based on the maximum diffusion speed and random vector as shown in the following models: where, is the diffusion speed, and γ is the random vector in [1,1] .Furthermore, the krill position is updated in the searching space, and the fitness is estimated according the new position of krill as represented in the following model: Based on the M best j , the best optimal value is obtained as the output of CKHO algorithm, which helps to optimize the resource load based on their usage.

Hierarchical Probability Routing (HRP)
After placing the contents in cloud with optimized cost and latency, the routing is performed for enabling the reliable and successful data transmission.For this purpose, the Hierarchical Probability based Routing (HPR) mechanism is implemented, which provides the optimized route to enable the data transmission.A unique router architecture called as, Service-oriented Routing (SoR) is put forth to enable a router to offer content-based reliable networking services.It uses the deep packet inspection to examine the packet's header and payload, and also it records the essential data in a database.In the proposed framework, the CDN is constructed with the routing based system, where the mathematical modeling of routing request is computed for enabling the quality improved data delivery.For improving QoE in CDNs, some of the essential parameters need to be concentrated such as total response time, packet arrival rate, load on server, and latency.By using the HPR mechanism, the aforementioned parameters are improved in the proposed CDN framework.
Here, the number of users N u and content file ∁ f are taken as the inputs for processing, and the optimized route R r is produced as the output.At the beginning, the total link delay is estimated as shown in below: Where, D SoR indicates the surrogate server, ϑ ij repre- sents the optimal route request probability, ρ 1j denotes the end to end delay between user1 st and j th user , ρ 2j is the end to end delay between user2 nd and j th user , ρ ij defines the end to end delay between useri th and j th user , ω 2 j represents the bandwidth between user2 nd and j th user , and ω 1 j is the bandwidth between user1 st and j th user .Consequently, the total waiting time is estimated according to the end-to-end packet transfer rate ( T SoR ) as shown in below: where, G SoR is the end-to-end packet transfer in surrogate server, G R represents end-to-end packet transfer in route nodes, and G RR denotes the end-to-end packet transfer in redirection probability.Moreover, the end-to-end packet transfer rate is estimated with respect to the surrogate server utilization rate G SoR as shown in below: (16) where, τ 1 and τ 11 are the surrogate server utilization, β 11 represents the higher surrogate server loads, and µ SoR 1 indicates the service rate of a router.In addition, the endto-end packet transfer rate in route nodes is estimated by using the following model: where, τ 1 ,τ 11 and τ 12 are the surrogate server utilization.
Then, the end-to-end packet transfer rate according to the redirection probability is computed by using the following model: where, τ DNS represents the domain name server uti- lization, and µ DNS is the service rate of a router for the domain name server.At last, the reduced total link delay is estimated for the server centers by using the following model: where, ρ CS 11 indicates the end to end delay between user1 st and server center, and ρ CS 12 represents the end to end delay between user2 nd and server center.The final optimized route is computed as follows: This optimal route is used for enabling the routing operation with reduced delay and ensured data quality.

Results and discussion
This section validates the simulation and comparison results of the existing and proposed content placement methodologies used in the cloud CDNs.For this (19 assessment, the different types of parameters such as total congestion cost, memory usage, response time, and etc. are considered in this study.The Internet or a transmission network are examples of the real-world systems that are conceptually represented as communication networks.It is made up of links connecting a set of identical nodes.The communication network architecture utilized by Boyan and Littman, which includes an irregular 66 network and a 116 node LATA communication network, is employed in this research to test our approach.An intangible representation of a real-world system, like the Internet or a transmission network, is a communication network.It consists of a collection of isomorphic nodes connected by links.We put the experiment to the test in various network topologies.As shown in Figs. 4 and 6, the traffic congestion cost is computed for the conventional and proposed content placement methods with respect to different network load level.Table 2 compares the total congestion cost of the conventional and proposed content delivery models.
Here, both an irregular network architecture and LATA communication networking structures have been considered for analysis.The network comprises a pair of segments that are closely connected to one another as well as a bridge link.Network congestion is more likely to occur on the middle bridge link.The average number of injected packets per time unit serves as the load in this study and is a parameter value of the Poisson arrival process.Based on the analysis, it is observed that the total congestion cost for the proposed HORCP model is effectively reduced under varying network load level.Figure 5 and Table 3 shows the total congestion cost for LATA communication network and proposed content placement algorithms under varying time step values and iterations respectively.
Figures 6 and 7 shows the total congestion cost of the existing and proposed content placement algorithms under varying time step values and iterations respectively.We contrast the differences in the congestion cost over time between the suggested and present approaches using the LATA network design.We carry out tests with heavy as well as light loads.Low load indicates that there is little network congestion and fewer data packets are injected per unit of time.When there is a high load, there is severe network node congestion and more data packets are injected per unit of time.As seen in Fig. 6 and Table 4, when the time step reaches 2000, the HORCP congestion cost converges to a particular value with low load.The outcomes show that the HORCP approach can reduce the cost of congestion for content delivery along the route request model.Figure 9 validates Cumulative Distributive Function (CDF) of existing and proposed content placement algorithms used for the CDNs with respect to changing response time.Typically, the response time is the most essential parameter used to determine that how effectively the proposed HORCP model could provide response to the given requests.It is clear that the algorithm's initial reaction time is practically identical to that of the greedy method.The performance does start to improve, though, as the response time goes down as the number of queries increases.The partitioning of the server's cache and consequent improvement in response quality can be employed to clarify this drop in response time.Similarly, the cache hit ratio and byte hit ratio are estimated for the conventional [8] and proposed models as shown in Figs. 10 and 11 respectively.In this instance, the cache will take 20% of the entire storage capacity.As can be observed, the hit rate is similar in the beginning when the number of requests is low, but as traffic and the number of requests rise, the HORCP method performs better than the standard models.As can be observed, the hit rate has increased as a result of the great popularity of the items.In the proposed approach, the hit rate for the cache increases as more queries are made.This implies that the HORCP model has improved CDN performance.The segregation of the cache space allows for the storage of both dynamic and static data in the cache region, which is the major reason for the performance improvement.The hit rate will initially be low since the cache space has been used up, but as the number of requests rises, the space is filled by previously processed requests, increasing the hit rate as well as efficiency.Moreover, Figs. 12 and 13 validates the request and response time of the proposed HORCP model under varying unit time respectively.For this analysis, the functions executed with the available local data, remotely available data, and remotely executed data have been considered.Overall, the response time is effectively improved in the proposed model with the use of HPR mechanism, since which provides the optimized routing path for content delivery.In addition, the CPU and memory usage rates are validated for the aforementioned functions as shown in Figs. 14 and 15.In order to assure the reliable data transmission in CDNs, the cloud resource consumption such as CPU and memory should be effectively reduced.Because, an increased usage may degrade overall performance of the content delivery system.Therefore, the CKHO algorithm is implemented in the proposed to optimize the resource usage load.
Figures 16 and 17 validates the traffic congestion cost of proposed HORCP model under varying network load level and time step respectively.In this analysis, the congestion cost before applying optimization and after applying optimization technique is estimated.In order to

Discussion
In order to ensure the reliable content delivery and enable data transmission in the cloud systems, the following factors must be satisfied: • Self-service on demand • Access to a wide network • Resource aggregation • Fast elasticity • Measurable service Content delivery networks have had a big influence on how content is provided to end users over the Internet.In the past, content providers have used independent CDNs to distribute their material to end users.It is difficult for content providers who either completely rely on third-party providers to understand and monitor the performance of their service due to the constantly changing landscape of content types, such as the transition from standard definition video to high definition to full high definition.Additionally, the geographic accessibility of the third-party infrastructure affects the CDN's speed.It offers a versatile approach that enables content producers to dynamically match and arrange material on one or more cloud storage servers based on preferences for coverage, cost, and quality of service.The pay-as-yougo model's advantages and economies of scale are the main implications.By utilizing clouds, content providers may manage circumstances like flash crowds more quickly without having to make infrastructure development investments.In addition to the CCDN-specific difficulties listed above, there are a number of critical

Conclusion
This paper presents a new framework, termed as, HORCP for an effective content placement and enabling optimized routing in cloud CDNs.It also aims to construct a better and reliable cloud delivery environment with minimized congestion.The content placement in the cloud tends to be one of the most important challenges that must be resolved in order to enable a dependable and efficient data transmission.Due to the rising demand from users for video streaming services, the system for streaming media services has encountered significant obstacles.The cooperative edge-cloud computing concept, which integrates edge computing and cloud computing, can greatly reduce the stress on streaming media service systems.Streaming content caching in the cooperative edge-cloud computing architecture is essential for improving user service quality.In this framework, the optimized content is placed in the environment according to the number of edge servers, steaming video, and set of users.Here, the CKHO algorithm's optimal result is used to assess the resource utilization load on the edge server.This method makes it easier and cheaper to upload content to cloud storage services.When the route has been optimized, the HPR mechanism is put into place to allow data transmission.This framework's For this purpose, a novel framework, named as, Hierarchical Optimized Resource Utilization based Content Placement (HORCP) model is developed in the proposed system.It also aims to construct a better and reliable cloud delivery environment with minimized congestion.The overall flow of the proposed HORCP model is shown in Fig. 1, which encompasses the following operations: • System modelling • Content placement • Cloud resource usage analysis using Chaotic Krill Herd Optimization (CKHO) algorithm • Hierarchical Probability based Routing (HPR) • Performance evaluation

Figure 2
Figure2shows the Workflow model of the proposed content delivery system.After system modeling, an effective content placement in cloud is performed with the use of the HORCP model.During this process, the optimized content is placed in the environment according to the number of edge servers, steaming video, and

Fig. 1
Fig. 1 Architecture model information flow in CDN

Fig. 2
Fig. 2 Workflow model of the proposed content delivery system

Figure 8 and
Figure 8 and Table5validates the total congestion cost of the proposed HORCP model with respect to changing optimization based iterations such as 30, 50 and 90.In this analysis, the congestion cost is estimated according to the time step value.The obtained performance outcomes show that the proposed HORCP model effectively reduces the congestion cost with the use of CKHO and HPR mechanisms.Since, the content placement and routing processes are carried out in the proposed framework based on the optimized resource usage and reduced latency.Therefore, the proposed HORCP model outperforms the other existing content placement algorithms.Figure9validates Cumulative Distributive Function (CDF) of existing and proposed content placement algorithms used for the CDNs with respect to changing response time.Typically, the response time is the most essential parameter used to determine that how effectively the proposed HORCP model could provide response to the given requests.It is clear that the algorithm's initial reaction time is practically identical to that of the greedy method.The performance does start to improve, though, as the response time goes down as the number of queries increases.The partitioning of the server's cache and consequent improvement in response quality can be employed to clarify this drop in response time.Similarly, the cache hit ratio and byte hit ratio are estimated for the conventional[8] and proposed models as shown in Figs. 10 and 11 respectively.In this instance, the cache will take 20% of the entire storage capacity.As can be observed, the hit rate is similar in the beginning when the number of requests is low, but as traffic and the number of requests rise, the HORCP method performs better than the standard models.

Fig. 4
Fig. 4 Total congestion cost for irregular network architecture

Fig. 5
Fig. 5 Total congestion cost for LATA communication network

Fig. 18
Fig. 18 Overall congestion cost analysis of HORCP model with and without optimization techniques

Table 1
Literature review of the previous works

Table 2
Comparison based on total congestion cost

Table 3
Comparison based on total congestion cost for LATA communication network

Table 4
Congestion cost analysis for the proposed model

Table 5
Comparison based on total congestion cost of HORCP