while optimizing runtime energy savings. This
provides tangible benefits such as orders of magnitude
lower power consumption for resource-intensive
applications, improved refresh rates for latency-
sensitive games, and overcoming smartphone
limitations in voice-based language translation
applications by remotely triggering unsupported
components.
While cloud servers may be farther from the end-
users, potentially increasing latency, their abundant
computational resources can still effectively reduce
the overall processing time if resource allocation is
done optimally. An example illustrating this concept
is Distributed Deep Neural Networks (DDNN).
DDNNs (Teerapittayanon et al 2017) can scale neural
network size and geographic coverage, improving
sensor fusion, fault tolerance, and data protection. By
assigning DNN sections to this hierarchy and training
them together, DDNNs minimize communication and
resource consumption and support automatic sensor
coupling and fault tolerance. As a proof of concept,
DDNNs leverage the geographic diversity of sensors,
improving object detection accuracy and reducing
communication costs by more than twenty times
compared to traditional processing of raw sensor data
in the cloud.
The final approach involves distributing
computations across different edge servers, and a
representative study demonstrating this concept is
DeepThings (Zhao et al 2018). it optimizes memory
utilization through scalable Fused Tile Partitioning
convolutional layers, provides dynamic load
balancing through distributed work stealing, and
improves data reuse and latency reduction through
innovative work scheduling and achieves a scalable
CNN inference speed of 1.7 to 3.5 times on 2 to 6 edge
devices with less than 23 MB of memory each, which
outperforms existing methods.
3.4 Private Inference
When performing computations on servers with user
data, privacy concerns become paramount.
GAZELLE (Chiraag et al 2018) and DeepSecure
(Darvish et al 2018) are two effective methods for
encrypting user privacy data without impeding the
inference process of deep learning networks.
4 CONCLUSION
In summary, this research has underscored the pivotal
role played by deep learning in augmenting the
potential of edge computing networks. It has
responded to the pressing demand for streamlined,
real-time processing capabilities at the edge and
examined the viability of incorporating deep learning
models into established edge computing frameworks.
The research has introduced novel approaches,
including fine-tuned neural network designs and
resource-efficient training methodologies, which have
yielded substantial enhancements in performance
across various edge applications.
These findings have profound implications for
areas such as autonomous systems, the Internet of
Things and healthcare, where low-latency decision-
making is paramount.
REFERENCES
J. Manyika, M. Chui, P. Bisson, J. Woetzel, R. Dobbs, J.
Bughin, and D. Aharon. The Internet of Things:
Mapping the Value Behind the Hype. Technical Report.
McKinsey and Company. pages 1-3, (2015)
R. Badnakhe, IoT Is Not a Buzzword But Necessity, Feb.
(2021) [online] Available:
https://www.iotcentral.io/blog/iot-is-not-a-buzzword-
but-necessity. Retrieved September 10, 2023.
X. Wang, Y. Han, V. C. M. Leung, D. Niyato, X. Yan and
X. Chen, Convergence of Edge Computing and Deep
Learning: A Comprehensive Survey, in IEEE
Communications Surveys & Tutorials, vol. 22, no. 2,
pp. 869-904, Secondquarter (2020)
M. Satyanarayanan, The emergence of edge
computing, Computer, vol. 50, no. 1, pp. 30-39, (2017)
J. Chen and X. Ran, Deep Learning With Edge Computing:
A Review, in Proceedings of the IEEE, vol. 107, no. 8,
pp. 1655-1674, Aug. (2019)
H. Pihkola, M. Hongisto, O. Apilo, M. Lasanen. Evaluating
the Energy Consumption of Mobile Data Transfer—
From Technology Development to Consumer
Behaviour and Life Cycle
Thinking. Sustainability (2018)
M. G. Murshed et al. Machine learning at the network edge:
A survey. ACM Computing Surveys (CSUR) 54, no. 8
(2021): 1-37
M. Satyanarayanan, P. Bahl, R. Caceres and N. Davies, The
Case for VM-Based Cloudlets in Mobile Computing, in
IEEE Pervasive Computing, vol. 8, no. 4, pp. 14-23,
Oct.-Dec. (2009)
M. Jang, K. Schwan, K. Bhardwaj, A. Gavrilovska and A.
Avasthi, Personal clouds: Sharing and integrating
networked resources to enhance end user experiences,
IEEE INFOCOM 2014 - IEEE Conference on
Computer Communications, Toronto, ON, Canada,
(2014) pp. 2220-2228
P. Liu, D. Willis and S. Banerjee, ParaDrop: Enabling
Lightweight Multi-tenancy at the Network’s Extreme
Edge, 2016 IEEE/ACM Symposium on Edge
Computing (SEC), Washington, DC, USA, (2016) pp.
1-13
DAML 2023 - International Conference on Data Analysis and Machine Learning
330