DNN Partitioning for Inference Throughput Acceleration at the Edge

Deep neural network (DNN) inference on streaming data requires computing resources to satisfy inference throughput requirements. However, latency and privacy sensitive deep learning applications cannot afford to offload computation to remote clouds because of the implied transmission cost and lack o...

Full description

Saved in:
Bibliographic Details
Published in:IEEE access Vol. 11; pp. 52236 - 52249
Main Authors: Feltin, Thomas, Marcho, Leo, Cordero-Fuertes, Juan-Antonio, Brockners, Frank, Clausen, Thomas H.
Format: Journal Article
Language:English
Published: Piscataway IEEE 01.01.2023
The Institute of Electrical and Electronics Engineers, Inc. (IEEE)
Subjects:
ISSN:2169-3536, 2169-3536
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Be the first to leave a comment!
You must be logged in first