Compressing and Fine-tuning DNNs for Efficient Inference in Mobile Device-Edge Continuum

Pruning deep neural networks (DNN) is a well-known technique that allows for a sensible reduction in inference cost. However, this may severely degrade the accuracy achieved by the model unless the latter is properly fine-tuned, which may, in turn, result in increased computational cost and latency....

Full description

Saved in:
Bibliographic Details
Published in:2024 IEEE International Mediterranean Conference on Communications and Networking (MeditCom) pp. 305 - 310
Main Authors: Singh, Gurtaj, Chukhno, Olga, Campolo, Claudia, Molinaro, Antonella, Chiasserini, Carla Fabiana
Format: Conference Proceeding
Language:English
Published: IEEE 08.07.2024
Subjects:
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Be the first to leave a comment!
You must be logged in first