High-Speed Data Communication With Advanced Networks in Large Language Model Training

Large language models (LLMs) like Generative Pre-trained Transformer, Bidirectional Encoder Representations from Transformers, and T5 are pivotal in natural language processing. Their distributed training is influenced by high-speed interconnects. This article characterizes their training performanc...

Full description

Saved in:
Bibliographic Details
Published in:IEEE MICRO Vol. 44; no. 2; pp. 31 - 40
Main Authors: Dai, Liuyao, Qi, Hao, Chen, Weicong, Lu, Xiaoyi
Format: Journal Article
Language:English
Published: Los Alamitos IEEE 01.03.2024
The Institute of Electrical and Electronics Engineers, Inc. (IEEE)
Subjects:
ISSN:0272-1732, 1937-4143
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Be the first to leave a comment!
You must be logged in first