Edge intelligence applications are a steadily growing field of interest. The IoT has largely contributed to the amount of data being produced and processed as consumers are replacing traditional home appliances with smart appliances, which not only serve their original purpose but can also collect data. Decentralizing the computing infrastructure allows for more data to be collected, as sensors can be placed anywhere. Edge systems leverage their large number of sensors by using them to feed into different applications. Often these systems utilize machine learning models and large amounts of data in order to train those models. Endpoints on the edge largely serve two major functions, user interfaces and data collection. A purely data collecting endpoint may be something like a security camera or thermometer, whereas a smart refrigerator may be an endpoint that both collects data and serves as user interface into an application. Edge intelligence aims to leverage this massive data collection to better train machine learning models for a variety of purposes.
This section contains a critical discussion of intelligent edge computing systems. We begin by describing the potential use cases for edge intelligence systems in detail, discussing what the problem presented is and how edge intelligence can solve the problem. We also discuss situations in which edge intelligence systems have been used and the lessons learned from those implementations. The second section describes situations in which mathematical approaches to edge intelligence can be used to enhance or optimize any approach for networking, problem solving, or other similar circumstances. The final section describes problems in the field of edge intelligence and where future researchis needed to best address these problems, potentially with approaches described in the second section.
3.1 Potential Edge Intelligence Applications
Explorations into the use of edge intelligence as a solution to problems previewed thus far have yielded several different results. This portion of the survey presents an in-depth review of such problems and how edge intelligence has worked to resolve them. This section discusses where edge intelligence has been explored as a solution to a problem and looks further into practical applications or other systems where edge intelligence has been used.
The intelligence aspect of Edge Intelligence focuses on the implementations and algorithms that edge systems currently use and possible future avenues to explore. Currently, edge intelligence is being used for multiple types of systems, including brain tumor diagnoses [
22], Smart City applications [
23], temperature prediction for agriculture [
24], text recognition from images [
25], and unmanned vehicle operations [
26].
Zhou et al. describe the current state of affairs in regard to edge intelligence [
27]. The authors describe developments in the history of edge intelligence and focus on areas in which growth of edge computing may provide a path forward in artificial intelligence domains.
Neto et al. [
23] provide a theoretical framework for a way to create a “Smart City.” The smart city as a concept refers to an edge network with sensors to automate things such as traffic, traffic enforcement, parking, and facial recognition, as shown in Figure
5. They first identify that a smart city will need to take in massive amounts of data. This poses the first problem they aim to address with their Multilevel Information Distributed Processing Architecture (MELINDA), which theoretically breaks down huge data streams into objects of interest before processing. Data are sent to the processing facility where MELINDA filters raw video data (in their proposed application) into events of interest, frames of video that may be useful for the model. A second software component within this MELINDA architecture then identifies the object of interest within that frame of video. The data are given environmental context and sent to another processing node for decision-making. By sending the massive data streams through layers that cut them down into objects of interest, the data become manageable [
23]. Figure
5 presents a potential application of smart cities/autonomous vehicles: optimizing traffic flow based on predicted behavior of vehicles.
Ahmed et al. use a federated learning system to perform deep learning on mobile devices using smaller local models to make predictions for the user, feeding the information from those tasks to a server. The server then takes the data and updates a larger model with the information, creating a more robust and precise network. The server is where the larger deep learning model is stored, whereas the endpoints have smaller, similar versions. This server then pushes out the updated weights for the models onto endpoints so that they may more accurately predict future tasks [
28].
Li et al. discuss the growth of machine learning as of late, with focus on deep neural networks and their growth in software applications [
29]. Mobile applications, with limited storage and computing power, find it difficult to leverage these capabilities. As such, the authors suggest the implementation of an accessible edge intelligence interface with partitioning features to reduce latency. A prototype developed on lightweight hardware indicates a degree of success in such an approach.
Christensen et al. describe a system built on edge intelligence to create and deploy network resources automatically as needed by the network [
30]. The system, called OpenEdge, involves three components that provide for network abstraction, authentication, and broadband termination. The system enables participants to decide which resources specifically are exploited and the time frame for doing so. The article suggests that edge intelligence can be incorporated into networks with minimal impact to traditional networking architectures, and can even serve to reduce costs through monitoring usage.
Maier et al. [
31] discuss the potential for edge intelligence to assist in the growth of fiber-wireless networks. Present approaches to increasing the potential of optical networks focus more on increasing capacity of the network. Maier et al. suggest using edge computing networks to route requests through the network in a more efficient, orderly manner with the goal of reducing latency across the entire network.
Al-Rakhami et al. suggest using Docker, a virtualization software, to enable a more manageable, widely distributed set of tools to be distributed among different hardware devices and software applications in the edge computing environment [
32]. They focus on the use of human activity identification to highlight the effectiveness of this approach, which uses support vector machines (SVM) to identify the behavior exhibited in these situations.
The capacity for edge networks to operate in a manner capable of running highly responsive and demanding processes is described by Dai et al. [
33] Their approach focuses on exploring the potential for edge computing systems to enhance these capabilities in the context of multi-user wireless networks. In these systems, processes or users can be remotely engaged from a multitude of devices or situations. The approach highlighted in this article reportedly provides a great deal of improvement over its contemporaries.
Ren et al. propose a generative-coding group evolution algorithm that shows potential for edge devices to become more prominent in the industrial environment [
34]. A simple grouping strategy is proposed, which provides an almost-optimized solution in a small number of situations. Enhancing this approach or adapting it to generalized situations could provide significant help in addressing similar edge intelligence problems.
Combining the growth of both Blockchain and edge intelligence technologies, Doku and Rawat propose a solution to identify relevant datapoints when operating in a distributed machine learning environment. [
35]. A “Proof of Common Interest” metric is described that can be applied to a machine learning model through which decisions can be made on novel datapoints. The model can then be distributed among members of the network to enhance the rate at which decisions can be made among other components.
Meloni et al. discuss the potential for smart grids and their monitoring of several different smart devices to maintain an accurate representation of a network’s current state [
36]. Through the use of an edge computing architecture, the authors suggest that the flexibility provided by edge computing allows for a more effective and efficient approach to be instantiated. Case studies are analyzed that provide a more thorough analysis of the situation.
Modern personal assistant software (such as Apple’s Siri) are operated largely on a cloud-based infrastructure. Kang et al. [
37] highlight the potential performance enhancements that could be achieved by migration of similar services to an edge computing environment. The study put forth describes 8 intelligent applications, with results including an increased latency, reduced energy consumption, and higher data center throughput.
Chen et al. propose “green edge intelligence,” an enhanced approach to operating artificial intelligence algorithms in a mobile edge computing environment [
38]. The approach highlights the effectiveness of using a cache, which could potentially be shared among edge nodes in an edge computing environment to reduce the workload across the network. The goal of the authors is for researchers to focus more efforts on developing potential applications of green edge intelligence.
The identification and understanding of smart-home activities is the focus of an article by Zhang et al. [
39] on the potential use of open-source software in edge computing. The discussion looks for low-cost solutions to identify activities in smart homes equipped with IoT technology. Promising when compared with other approaches, preliminary results suggest that these approaches can be explored in more detail and applied in other contexts.
A text discussing the potential of mobile solutions devotes an entire chapter to analyzing the potential for IoT-enabled houses to provide a SmartLiving lifestyle. Rahman et al. discuss their primary concern, teaching automated systems the process of reasoning [
40]. The authors strongly suggest that an implementation based on edge computing can significantly enable reasoning among the many multifaceted utilities in a SmartLiving environment.
A review of edge computing by Sittón-Candanedo et al. also proposes a modular, tiered method for handling edge-driven IoT devices [
41]. Using Blockchain technology, their approach provides greater deal of flexibility and security than other contemporary methods. In its implementation, the proposed solution lowers the costs necessary for data throughput.
Unmanned aerial vehicles (UAVs), more commonly known as
drones, are another new development that could hinge on advancements in the edge computing industry. Allowing UAVs to communicate directly with one another or with a remote base station could enable a larger and safer variety of uses. Chung et al. describe an approach to using UAVs to monitor offshore wind farms [
42]. Using UAVs to monitor wind conditions, the positioning of turbines can be optimized to increase power output by substantial margins.
In a manner similar to that of UAVs, autonomous, self-driving vehicles face a similar need to communicate amongst each other. Dai et al. suggest that enhanced artificial intelligence algorithms operating in these vehicles may be capable of addressing some of the current deficits [
43]. By caching certain information at various points in the network, autonomous vehicles are able to make more effective decisions in a shorter amount of time.
The topics discussed in this section briefly highlight the great variety of topics and domains in which edge intelligence implementation can be used. The simple fact of its widespread potential use cases and improvements upon existing architectures suggest that further enhancements in edge intelligence can be explored. Making the technology more efficient or otherwise more useful should allow for more robust and effective implementation in a greater number of domains. Thus, improvements on the theoretical level must also be considered.
3.2 Theoretical Systems and Optimizations
In this portion of the survey, we focus on proposals that are more theoretical in nature, such as algorithms, system design, and other optimization practices. This section can be used as a springboard for developments into the subsequent section in which issues are presented. The goal is to promote ideas that can be used to solve the problems occurring in edge intelligence applications. Current edge systems suffer from large energy usage and inefficient or bulky algorithms. Intelligent networking aims to solve these problems to make edge systems usable and fast. Edge systems leverage many microservices, spinning up machines as needed and connecting to various different processing centers to use the data that endpoints have collected. Problems arise when these processing centers are busy or have a long queue from other endpoints. Zhou et al. propose an algorithm for an efficient way to send microservice requests to endpoints in order to optimize edge performance [
44]. Before sending a request, this algorithm polls processing centers and receives information regarding response time, current system load, and more in order to more strategically decide which processing center to send data to [
44]. Processing nodes may lie dormant when not being used or endpoints may be requesting too much of single processing nodes overloading them.
Al-Rakhami et al. explore the advantages of using a containerization technology such as Docker to create on-demand processing for the data produced by the edge system. Creating processing only as needed saves on resources and provides for edge systems to remain lightweight when not in use [
32].
Wang et al. look into the possibility of using edge computing to optimize social virtual reality (VR) applications on all devices. They propose an optimization problem and model that encompasses the most pressing problems related to using edge technology for social VR. ITEM, the algorithm they use to solve this optimization problem, iteratively constructs graphs and uses well-known max-flow algorithms to solve smaller graph cut problems. ITEM outperforms other algorithms in most preliminary test scenarios [
45].
Li and Lan discuss how to determine execution cost of edge computing and how to effectively assign tasks in order to minimize total execution cost. The authors model their problem as a multichoice game and implement the Shapley value, which is commonly used in similar cooperative game theory problems. The proposed solution shows promise by outperforming other similar solutions to this problem. The analysis of various simulations shows that their algorithm works well in heterogeneous edge networks but might fall short in other networks [
46].
In an attempt to find a solution to the AI service placement problem in a multi-user mobile edge computing system, Lin et al. [
47] proposed a mixed-integer linear programming problem that looks to minimize computation time and energy consumption of users. By deriving various expressions that represent the optimal allocation of resources while still maintaining low complexity, the authors were able to find a solution that uses various search-based algorithms to efficiently solve the linear program stated earlier. The algorithm can even be scaled up to be used on larger networks. The results of many simulations of running the algorithm leads them to believe that this solution performs exceptionally and outperforms many other algorithms.
Zhou et al. [
48] takes a look at a UAV-enabled wireless powered mobile edge computing system and a related power minimization problem. As this type of problem is difficult to solve outright, an alternative solution is proposed. An optimization algorithm based on sequential convex optimization is suggested to solve the power minimization problem. This approach is efficient and preferable to similar algorithms for this problem according to the results of the simulations.
Wang et al. propose a power allocation algorithm, to be used with edge intelligence, that prioritizes maximizing learning instead of communication throughput. The proposed learning-centric power allocation (LPMA) algorithm involves allocating radio resources and makes use of an error classification model. This algorithm is different from many well-known algorithms and is shown to outperform other power allocation algorithms [
49].
Huang et al. propose an edge intelligence framework for creating IoT applications. The main goal of this new framework is to streamline data analytics in a more reliable fashion. The authors accomplished this by creating annotation based programming primitives that developers can use to build learning capabilities and implement a user activity recognition system on edge devices. Using their proposed edge intelligence framework will increase performance without sacrificing accuracy of activity recognition [
50].
Mobile edge computing is another area for which many are looking to optimize performance. Bouet et al. look into optimizing resource space and capacity of MEC servers [
51]. They formulate a mixed-integer linear program problem that takes into account MEC server size, operation area, and number and still meets all MEC goals. Their proposed solution to this problem is a graph-based algorithm that involves forming MEC clusters.
In the field of health care, Sodhro et al. wish to improve quality of service (QoS) using edge computing. The Window-based Rate Control Algorithm (w-RCA) is their proposed algorithm to optimize QoS [
52]. When compared to conventional battery smoothing algorithm (BSA) and a baseline using MPEG-4 encoder, w-RCA is shown to outperform both algorithms when optimizing QoS in remote health care applications. It is also important to note that the w-RCA has better results while using smaller buffer sizes than the other two algorithms.
Edge computing is a possibility for improving mobile vehicles. Luo et al. investigated how the 5G-enabled vehicular ad hoc network (5G-VANET) and edge computing can be optimized for many vehicle applications. They propose a prefetching scheme to deal with rapidly changing data and employ common graph theory algorithms to efficiently handle the data. Overall, the proposed scheme efficiently optimizes the 5G-VANET [
53].
Liu et al. study the problem of large-volume data dissemination that would be used for automated driving assisted by cellular networks. First, models are examined that take into account the high variability in vehicle mobility. The data dissemination problem is then formulated into an NP-hard optimization problem. The authors do not stop there; they then propose a low-complexity dynamic programming solution and run many simulations displaying the effectiveness of this solution [
54].
Liu et al. proposes a blockchain-based framework for video streaming on mobile edge computing devices [
55]. The proposed framework uses adaptive block-size while also using two different offloading models. The authors also formulate an optimization problem that takes into account the adaptive block size, resource allocation, and offload scheduling. Simulations show that the proposed method solves the problem quite efficiently.
An et al. propose the use of edge intelligence in order to disperse Hypertext Transfer Protocol (HTTP) anomaly detection across multiple nodes rather than being centralized on a single server [
56]. The framework proposed efficiently and reliably identifies anomalies in HTTP of the IoT. in addition to the framework, a data processing algorithm is proposed to reduce redundant HTTP data and to more effectively classify the anomalies. This framework and proposed algorithms show an increase in accuracy and speed when detecting unknown anomalies.
Lodhi et al. look at the current state of edge intelligence and note major flaws that need to be addressed in the development of these systems. They propose that edge systems need to address the computational gap, cost, latency, scalability, and security [
57]. The computational gap refers to finding ways to make low-power edge endpoints capable of doing high-level machine learning tasks. Cost refers to the massive amount of infrastructure that must be created for an edge system to function. There must be many endpoints deployed, data centers and processing facilities created, algorithms to process data, databases, and communication systems between all of these things. The cost of creating this infrastructure is reflected in the complexity of the system and is not always a worthwhile investment for many implementations. Some applications require real-time responses from data collected. Latency refers to the time it takes for data to be collected, processed, and sent back to an endpoint for use; the goal is to minimize that time by keeping as much processing on endpoints as possible. Travel time between processing centers, databases, and endpoints can cause serious delays. Scalability aims to make edge systems that are comprehensive but not bloated and taking up network resources unnecessarily. The ideal edge system is infinitely scalable as you add endpoints, processing facilities, and databases as needed. Security is a massive concern when discussing edge computing. The nature of the technology requires some data to be sent over networks. Ideally, endpoints would be able to scrap unnecessary data, clean useful data, and encrypt it in order to send it to be processed. Data being transmitted must be secured as it can be potentially sensitive information. However, edge systems are especially vulnerable due to the massive amounts of endpoints on the system. Every endpoint becomes an attack surface; therefore, every endpoint must be meticulously designed for security and constantly updated to prevent attacks. This is both expensive and resource intensive. Figure
6 provides a summary of practical and theoretical benefits of edge intelligence to exploit big data.
Federated edge learning, first proposed by Konečný et al. in 2015, defines a system for edge computing in which edge endpoints can share a model between them without sharing local data, allowing for training to occur on nodes simultaneously and securely [
58,
59]. Zhou et al. propose a system for IoT machine learning model sharing between endpoints [
60]. While not explicitly federated edge learning, their proposed system mirrors it closely. They propose to alleviate security concerns by using encryption to hide model parameters between endpoints and granular access control systems in order to regulate who may access the edge system. Federated edge learning systems come with data quality challenges when model sharing [
61]: one endpoint’s bad data can poison the performance of the model. Zhaohang et al. discuss the problem in greater detail, introducing the concept of device heterogeneity and statistical heterogeneity [
62]. Device heterogeneity recognizes that not every device performs in the same way. Some edge endpoints may be running different hardware, some may have less energy available, some have poor network connections, and so on. All of these factors can affect their overall performance on the network. Zhaohang et al. propose an optimization algorithm in which a “staleness function” is implemented to identify nodes that are underperforming and have them process smaller jobs or take them off the network [
62]. Statistical heterogeneity comes from a single system taking in data from multiple edge sensors, but the context of that data may differ entirely. The example they use of statistical heterogeneity used is an edge network deployed both in a school and in an office using video sensors, taking in children and adult’s data and running them through the same models when the features of these groups should be separated [
62].
When building an edge network, physical hardware is an important design decision made for the edge endpoints and processing nodes. Li et al. analyze multiple edge endpoint hardware configurations and calculate their performances [
63]. They identify that physical distance is an outlying factor that will affect network latency. Data have to travel from point A to point B; longer distances from edge to processing node causes latency. Model performance is directly tied to the actual physical hardware that it is being run on. Edge endpoints can run some ML models effectively but it comes at the cost of energy and time. The highest-performing endpoints have cloud-based processing systems, which comes with both security risks and introduces network latency.
This section shows that many areas can be studied and optimized by using edge computing. Even within in systems that already use edge computing, these proposed algorithms and solutions can drastically improve the efficiency and reliability of edge networks and systems. Formulating new problems and optimizing different aspects of a given scenario allows for steady improvement in edge computing and the implementation of edge intelligence. One such optimization is the implementation of machine learning models into edge systems to leverage their strengths and give them more power. Lightweight Machine Learning is a technique that is especially useful in this area.
3.6 Concerns and Issues with Edge Intelligence
This final subsection describes the many problems and challenges that have been encountered within edge intelligent systems. In this section, we discuss the challenges that have limited the growth of edge intelligence in regard to domain, implementation specification, and other areas. The objective here is to address these issues using the information for improvement presented in the previous subsection.
Highlighting the increasing use of IoT-enabled devices, Plastiras et al. provide an exceptional overview of some of the growing challenges in edge computing [
75]. Some specific items they point to include the high computational requirements of decision-making tasks and algorithms in these devices. The efficiency and enhancement approaches described in the previous section may be able to improve upon these areas. Additionally, the large number of devices involved in edge networks, and potential for a large number of tasks to be performed at once, require a serious amount of consideration for optimization. Efficiently handling numerous disparate devices is likely a critical component to enhance in any edge computing system.
One particular instance that could be analyzed in more depth is the success of edge intelligence systems implemented with user-driven recommendation systems. Su et al. [
76] focus on information to provide tourist experiences to users from other locations. Improvements in this area could focus on applying more efficient algorithms in systems with sparse datasets or finding more unique information to use with users who have otherwise dissimilar data profiles.
Efficiency of tasks in edge intelligence systems is a large concern in many areas. Jia et al. [
77] describe some of the challenges in working with extracting unique user-derived vocal profiles from auditory inputs (such as those provided to automatic phone systems) to more accurately and effectively respond to the client. The approach used in this study seems to have provided a great deal of success in making these enhancements operational in an edge-based system. The results of their experiment are highly accurate and generated in extremely short times. Thus, any study working to improve upon this would need to yield vastly higher accuracy in a comparable or better time frame.
Roman et al. provide a unique outlook on the potential security and privacy implications produced by the growth of edge intelligence systems [
78]. As edge intelligence grows and has the potential to move into more domains, so too does the potential for its malicious or unintended uses. As such, it is important to evaluate IoT sytems in a similar manner as that of Roman et al. in an effort to reduce the potential for indirect harm to users or clients.
Huh and Seo have produced a survey with focus on the challenges 5G networks present to edge intelligence [
79]. While many systems can and do achieve efficiency in these regards, the nature of mobile networks and mobile edge computing needs to be addressed more fully in order to achieve maximum efficiency. In some cases, it might prove useful for mobile edge computing to have some buffer system in place to store data temporarily in instances when data are being produced while the network is unavailable.
Yee et al. focus on another prominent privacy risk that edge intelligence implementations will likely have to address, that of facial recognition and its potential implications [
80]. Facial recognition enables a company or application to accurately and securely identify users but with the potential for serious impact on the privacy of these users. Beyond privacy concerns, there are also technical and efficiency concerns in implementing facial recognition algorithms. While Yee et al. touch on the latter, the former is still a concern that would need expanded upon.
The necessity of responsive, effective solutions in emergency situations is described by Wang et al. [
81]. Their analysis of Mission Cognitive Wireless Emergency Networks (MCWENs) highlights how edge networks can effectively and efficiently support these critical applications. While this article provides a detailed analysis and some application examples, the assured efficiency of these applications and their ability to support and enable these systems should be further researched for the benefit of the general public in emergency situations.
As described in Section
3.1, the automotive industry, with specific focus on autonomous vehicles, is a major field in which intelligent edge systems may operate. Pan et al. describe an approach to making these systems more efficient [
82]. Their approach and the simulations conducted indicate a comparable error rate with traditional methods, which is a great achievement. Practical applications, however, need to account for other unpredictable hazards in real-world environments, such as weather, obstructions, pedestrians, or other external interference that can be difficult to capture in a simulation.
Gill et al. provide a thorough analysis of the many instances in which current cloud computing paradigms operate [
83]. The authors focus on how Blockchain, the IoT, and AI could potentially operate in the cloud environment. While edge intelligence is not the primary focus, the information presented and concerns expressed are all still valid when viewed from the intelligent edge perspective.
The general sentiment expressed suggests that edge computing architectures can be and often are successful at replicating the efforts of traditional cloud systems. By focusing on applying more efficient algorithms to these systems, we can more accurately identify situations in which a greater operational efficiency can work towards a more productive, safe, and effective computing system. As such, applying algorithms and concepts detailed in the previous subsection to the problems and challenges detailed in this subsection can serve as a greater and more effective solution for organizations and businesses alike. The discussion of this notion continues in the following section.
Table
1 summarizes the sets of potential edge intelligence applications through the literature and outlines the period of publications. The table is divided into sections corresponding with those of the survey itself, Potential Edge Intelligence Applications, Theoretical Systems and Optimizations, and Concerns and Issues with Edge Intelligence.
3.7 Future Edge Intelligence Applications
Several other practical functions for edge intelligence are currently being implemented across several of the industries we mentioned previously. Some of these implementations are described in this section.
ADEPOS [
84] is a machine learning algorithm for edge computing. The algorithm combines the Extreme Learning Machine (ELM) framework with boundary methods to detect anomalies in hardware at the edge layer. Their application was geared towards machine health to reduce downtime and maintenance costs by using IoT sensors to monitor machinery in its new state and be able to detect an anomaly that indicates the machinery requires maintenance.
The authors of [
85] discuss the use of machine learning models to detect security anomalies in an IoT network. Their model efficiently reconstructs inputs that resemble normal network traffic and poorly reconstructs input attacks or anomalous inputs. They use this reconstruction error as a classifier to identify normal versus abnormal network traffic.
The authors of [
86] present a series of ML techniques for mapping and adapting machine learning workloads to Field Programable Gate Arrays (FPGAs). FPGAs are limited in their capabilities due to memory and energy constraints. They are relatively inexpensive devices that are a good match for IoT usage. The authors present techniques for utilizing Deep Neural Networks with these resource-limited FPGAs. They use pruning and quantization to reduce the size of the DNN to improve resource utilization.
The authors of [
87] discuss the heavy burden on processing power required for data preprocessing and training in machine learning. The authors rightly note that IoT devices are less likely to have the computing resources needed to perform machine learning tasks. In order to meet this need, the authors introduce Transparent Learning (TL). TL is a framework that moves training tasks away from the IoT devices out to an edge device or edge servers. Since data are being sent to an edge node or server for computing, an efficient cache system must be utilized to optimize performance. TL does just this, and the results show that it decreases the time cost and accuracy rate as compared with the TensorFlow framework.
The author of [
88] elaborates on some of the methodologies behind edge computing when working with data streams – in which information is pushed constantly from the bottom rather than pulled selectively by the top. A practical system is highlighted, which takes information from the pipeline and sends it both to real-time view and develops it into training data simultaneously.
The authors of [
89] propose a framework for deploying DNNs on resource-limited hardware called Edgent. Edgent utilizes adaptive partitioning of DNN resources between the device and the edge server to cover the shortcomings of either individually. The other specialized approach is right-sizing via early exit, which allows for a reduction in latency by allowing output to be pulled when certain intermediate nodes are reached.
The emerging federated learning and its capacity to support large-scale model learning without the need to move the data into a centralized data center is now the driving force for future applications of edge intelligence in transportation, smart cities, and other IoT-based applications and services [
90].