Compressing and Fine-tuning DNNs for Efficient Inference in Mobile Device-Edge Continuum
Pruning deep neural networks (DNN) is a well-known technique that allows for a sensible reduction in inference cost. However, this may severely degrade the accuracy achieved by the model unless the latter is properly fine-tuned, which may, in turn, result in increased computational cost and latency....
Saved in:
| Published in: | 2024 IEEE International Mediterranean Conference on Communications and Networking (MeditCom) pp. 305 - 310 |
|---|---|
| Main Authors: | , , , , |
| Format: | Conference Proceeding |
| Language: | English |
| Published: |
IEEE
08.07.2024
|
| Subjects: | |
| Online Access: | Get full text |
| Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Be the first to leave a comment!