Base Paper Edge Computing
Base Paper Edge Computing
Abstract—The fifth-generation mobile network is characterized an important milestone in the 5G era. Furthermore, live video
as the edge of wireless connectivity for all intelligent automation. analytics can be a common application in the MEC system
Technically, the services’ requirements for Quality of Service because of low-latency demand. For example, surveillance
(QoS) have become more strict on latency and throughput. As
a result, the concept of Mobile Edge Computing (MEC) has cameras play an essential role in maintaining public safety. In
become promising. By placing servers close to the user-equipment addition, the widespread development of traffic monitors, self-
(UE), the paradigm enables much lower data transmission time driving, Virtual Reality (VR), and Augmented Reality (AR)
compared to the cloud-based scenario. With this advantage, MEC related to video analytics. Object detection and recognition
reaches the requirements of low-latency. Moreover, recognition apply in many areas of computer vision, including image
and detection technology can be thus implemented in several live
video analytics scenarios. However, due to the limited physical retrieval, security, surveillance, automated vehicle systems,
size on the edge server, resource allocation becomes a crucial and machine inspection.
issue. In this paper, we proposed a Resource Management method However, limited computing resources of edge servers may
with Multiple Applications in Edge architecture (RMMAE) to not afford multiple complexity and huge DL models. In brief,
intelligently reallocate computing tasks in the heterogeneous it is a trade-off problem between low latency and high quality.
network. We design an algorithm to allocate computing resources
to applications such as facial detection, object detection and Delivering all data to the cloud server and performing high-
pose estimation in our Edge testbed, and we prove impressive quality video analytics come with exceptionally high trans-
improvement and performance on our testbed with multiple mission costs. If we only process the data to the edge server,
applications. the server near users, the computing power may constrain the
Index Terms—Edge Computing, Resource Allocation, Live result. So, we choose edge computing of cloud computing
Video Analytics
on demand. Some works proposed the idea that determines
whether the service is quality-oriented or speed-oriented by
I. I NTRODUCTION
scoring users’ needs, such as [2] and [3]. The distributed
With the rapid rise of the fifth-generation mobile commu- computing for the edge server and the cloud server is another
nications era, artificial intelligence and Deep Learning (DL), choice. The concept is similar to federated learning [4]. The
the efficiency and speed of network transmission and powerful authors applied a decentralized approach to make the system
computing capabilities are becoming increasingly important. more efficient. Both [5] and [6] deployed the work through
Therefore, cloud computing is also popular because of its distributed deep learning models deployed on the MEC nodes
sufficient data storage and excellent computing power without for better performance. We can divide one task into sub-tasks
direct active management by the user. Nowadays, several and distribute them to an edge server and a cloud server
providers have paid service of the cloud server, such as for computing in an Edge-Cloud system. On the other hand,
Amazon Web Services (AWS), Google Cloud Platform (GCP), resource allocation is also an essential issue for the edge-cloud
Microsoft Azure, etc. Users may choose suitable services system. Hence the restriction of the computing resource on the
to solve the problems. Nevertheless, the weakness of cloud edge server, we need to allocate limited resources effectively
servers that the transmission delay between end-users and and use them where it is most need. [7] proposed a block-
cloud servers may keep the cost increasing considerably. Video chain-based video streaming system, and they formulate the
analytics services in real-time are also harsh constraints on problem of optimizing the offload scheduling to optimize. [8]
network bandwidth and congestion. Therefore, the purpose provided a promoted QoS method to offload tasks.
of our works is to provide a high-quality and low-latency The DICE-IoT system in [9] is our previous work. It
computing system. Base on the ETSI Mobile Edge Computing provides a facial recognition app for video analytics and for-
(MEC) in 5G white paper [1], we realize the edge computing is mulates the management to incentivize cooperative computing
provision between the Edge and the Cloud. Inspired by this
The authors are grateful for the funding support from Ministry of Science
and Technology (MOST) of Taiwan under grant 109-2221-E-002-148-MY2 work, we will continue the concept of resource allocation
and 109-2218-E-002-018-. and implement more applications on our testbed to deal with
Authorized licensed use limited to: VTU Consortium. Downloaded on October 22,2024 at 05:07:11 UTC from IEEE Xplore. Restrictions apply.
978-1-6654-2772-2/21/$31.00 ©2021 IEEE 80
2021 30th Wireless and Optical Communications Conference (WOCC)
TABLE I
PARAMETERS OF CONFIGURATION
Notation Definition
f rφ Frame sampling rate (in fps)
wφ , hφ Frame {width, height} (in pixels)
rsφ Video resolution rsφ = wφ × hφ
bφ
edge Data size uploaded
from the device to the edge node per frame (in MB)
bφ
cloud Data size output
from the edge node to the cloud per frame (in MB)
mφ
edge GPU memory usage at the edge node (in MB)
mφcloud GPU memory usage at the cloud (in MB)
Fig. 1. Hierarchical architecture of two-tier Edge-Cloud in our testbed uφ
edge GPU utilization at the edge node (in [0, 1])
uφ
cloud GPU utilization at the cloud (in [0, 1])
cφ
edge Energy consumption at the edge node (in Watts)
more complex problems. Furthermore, we consider a model- cφ Energy consumption at the cloud (in Watts)
cloud
level pipeline composition that can decompose the model on pφ Per frame processing time at the edge node (in ms)
edge
the different servers to cooperate for one application. The pφ Per frame processing time at the cloud (in ms)
cloud
architecture of our MEC system is a two-tiered edge cloud
computing with real-time video analytics. On top of that,
we provide an improved method RMMAE for resource load while the cloud server is assumed to have unlimited resource
balancing and implement the complete MEC system in the budget.
actual environment. We considered three types of end devices: mobile phones
The main contribution of our work is as follows: connected to a 4G/5G base station, those connected to a WiFi
• We reduce the transmitted latency by allocating resources access point, and PCs connected with the Ethernet; each of
appropriately according to the system’s current situation. them has a camera. The solid line is the data flow, and the
• This is the Edge system with implied reasonable algo- dashed line means is the control flow. The camera records
rithms to discuss and realize the situation where multiple live video and streaming it to the MEC server for real-time
applications are in the system at the same time. analytics. The control nodes and the orchestrator in the MEC
• We distribute the resources by the detection accuracy server will then allocate computing resources and forward the
request, current video transmission quality and GPU streaming request to either an appropriate MEC node or the
status of edge server. cloud server, based on the end users’ demand and the GPU
status of the edge server. Generally, the edge server aims to
II. P ROPOSED S CHEME D ESIGN provide the service with high quality and low latency.
In this section, we introduce the three-layered architecture of
the system, including the Cloud Layer, the MEC Layer and the B. System Model and Configurations
User Application Layer. The interaction between these layers Our video programs run in both Cloud and MEC Layers,
allows better allocation and improves the system performance. but the resource management is only considered in MEC due
Also, we design the data flow and the control flow with our to the infinite resource assumption in Cloud. The devices in
management mechanism. the User Application Layer capture video and streaming it to
the MEC node for real-time video analytics, such as object
A. Edge-Cloud Cooperative Computing Architecture detection, facial detection, and pose estimation. Suppose that
Edge and Fog Management and Orchestration is essential each device has a specific latency requirement Lreq and a
in the MEC Layer, which plays a critical role in coordination quality requirement, Qreq . We also define configuration φ as
between components. According to hierarchical management, a specific decomposable pipeline and Φ as the collection of
we can master and integrate all the MEC nodes easily and all configurations.
orderly. Figure 1 shows the architecture of the Edge system We also consider the resolution and the video frame sam-
in our testbed. First, the Edge server includes several MEC pling rate for all devices and the GPU resources as the
nodes and the Edge Orchestrator; then, the Edge Orchestrator computing power on the cloud server and the edge server for
will manage and coordinate MEC nodes to work in the MEC each configuration φ. The partial offloading indicator ϕφ is a
server. As for a single MEC node, it consists of a control binary indicator and is defined as follows: If ϕφ = 1 means
node and a compute node. The control node may receive the the video is uploaded to Cloud Layer for serving. Otherwise,
control signals sent by the orchestrator. It also collects and MEC Layer processes the task.
sends the information about GPU status to the orchestrator To sum up, we denote the complete configuration set
periodically. The compute node in the MEC node provides Φφ (rsφ , f rφ , ϕφ ), which consists of all available configura-
low latency computing, but with limited resource constraint tions for the proposed DICE scheme. On the other hand, we
Authorized licensed use limited to: VTU Consortium. Downloaded on October 22,2024 at 05:07:11 UTC from IEEE Xplore. Restrictions apply.
81
2021 30th Wireless and Optical Communications Conference (WOCC)
Authorized licensed use limited to: VTU Consortium. Downloaded on October 22,2024 at 05:07:11 UTC from IEEE Xplore. Restrictions apply.
82
2021 30th Wireless and Optical Communications Conference (WOCC)
utility function of the cloud may like: A. Deploy MEC system in our Testbed
1) Testbed Environment: Chunghwa Telecom (CHT) has set
Ucld = Rcld − Ccld (2) up a 5G campus network and MEC server. The edge server
of our experimental testbed is point-to-point connected with
Ccld means the operation cost of the cloud per second: Chunghwa Telecom’s MEC server to develop various scenarios
X X X φ for the main architecture of the system.
Ccld = γcld xij priceφcld tφcld (3) MEC server: The server consists of a control node and
i∈N j∈M φ∈Φ a compute node. The compute node is equipped with an
individual GeForce RTX 3080 GPU instance. As for external
The γcld is a constant that denote the cost of converting
connections, there is a CHT SDN switch link between the
power consumption at the cloud server. We regard the total
MEC server and three kinds of UEs.
GPU utilization as the effort per second of the cloud.
Cloud server: Elastic Compute Cloud (EC2) of Amazon
X X X φ φ φ φ Web Service (AWS) is a cloud-server rental service of Ama-
ecld = xij tcld ucld f r (4)
zon. We rent Amazon EC2 G4 Instances and its type is
i∈N j∈M φ∈Φ
G4dn.xlarge equipped with NVIDIA T4 GPUs and custom
And the revenue of the cloud present as: Intel Cascade Lake CPUs.
User Devices: We have three scenarios to simulate the
Rcld = πecld (5) user devices: PCs connected to the Ethernet to surf the
C. Utility of MEC layer Internet, laptops to Wi-Fi, and mobile phones to 4G/5G mobile
networks. PCs and laptops connect to the MEC server via the
The expected utility function of the MEC stratum is: campus internet, while phones connecting via the CHT fiber
Umec = Rmec − Cmec − Pmec (6) optic link from the base station to the MEC server.
Applications: Our test platform has three types of ap-
plications: facial detection, object detection, and pose es-
And Pmec = Rcld represents the total service fee. timation. In our previous work [9], we provide the facial
The operation cost per second of the MEC stratum is similar detection application. This paper implements object detection
to the Cloud stratum: and pose estimation with YOLOv4, YOLOv4-tiny AlphaPose,
X X X φ φ φ
Cmec = γmec xij pj f r (7) and OpenPose. Also, we combine different kinds of video
i∈N j∈M φ∈Φ
analytics models above for a decomposable inference pipeline
and deploy with three offloading types: Edge only, Cloud only,
The γmec is a constant that denote the cost of converting power and Edge-Cloud.
consumption at the MEC node. We assume the service fee is 2) Parameter Measurement in Testbed: Taking the actual
proportional to the user satisfaction, we have the revenue as situation like the scene, we use the mobile phone with CHT
follows: X X X φ 5G network to capture the video stream as input and adjust
xij P q φ
Rmec = (8) the relevant camera and transmission information such as
i∈N j∈M φ∈Φ frame rate and video resolutions to measure various situations.
Both the edge server and cloud server have individual GPU
instances as in the previous description. We list 70 kinds of
And according to equation (3) and (7), the social welfare configurations with the different model schemes, and three
can be derived as: offloading types: Edge Only means the inference pipeline ran
X X X φ at the MEC layer entirely, Cloud Only means the inference
xij P q φ − Ωφj
W (a, x, n) = W (x) =
i∈N j∈M φ∈Φ
pipeline ran at the Cloud layer entirely, and Edge-Cloud, the
(9) inference pipeline is decomposed to run at the MEC layer and
Cloud layer.
The total operation cost 3) Testing Scheme: We simulate the user requests and
implement the RMMAE system in the actual testbed based
Ωφj = Υmec pφj + Υcld pφcld f rφ , ∀j, φ (10) on the configurations. The simulation parameters are as fol-
lows: the frame rate as 30 fps, the video resolutions are
uniform distribution in 1080p, 720p, and 480p, the accuracy
requirement is the uniform distribution between 0.5 to 0.8, the
IV. E XPERIMENTAL R ESULTS latency requirement is the uniform distribution between 200
This chapter set up an implementation platform and applied to 1200(ms) and the application selection also random in three
the RMMAE algorithm. There are multiple video analytic choices. Further, we compare six schemes for experiment:
applications on a real testbed to select and prove our MEC (1) Proposed-C: Completely framework with enable pipeline-
system’s better performance by the experimental results. sharing.
Authorized licensed use limited to: VTU Consortium. Downloaded on October 22,2024 at 05:07:11 UTC from IEEE Xplore. Restrictions apply.
83
2021 30th Wireless and Optical Communications Conference (WOCC)
Fig. 3. Performance versus number of devices. (a) Social Welfare. (b) Serving Ratio. (c) Utility of MEC layer. (d) Utility of Cloud layer. (e) Total Processing
Time(ms)
(2) Proposed- Edge: Edge-only system with enable pipeline- accuracy and latency with our management method and apply
sharing. it in several applications.
(3) Dedicated-C: Completely framework with a dedicated
V. C ONCLUSIONS
pipeline.
(4) Dedicated-Edge: Edge-only system with a dedicated We provided multi-application services on the system, such
pipeline. as facial recognition, object detection and pose estimation,
The proposed approaches enable one pipeline to serve and also proposed the RAMME method to deal with more
several task types. The dedicated serve by dedicated task scenarios and deploy with the intelligent MEC orchestrator.
type. We implemented the system with the CHT campus 5G system
(5) Greedy for Latency: Greedy for the minimum latency and run the complete service on our actual testbed. Finally,
configuration. we proved that the proposed scheme of the RAMME has a
(6) Greedy for Accuracy: Greedy for the highest accuracy better performance than other schemes and with higher social
configuration. welfare as well.
Authorized licensed use limited to: VTU Consortium. Downloaded on October 22,2024 at 05:07:11 UTC from IEEE Xplore. Restrictions apply.
84