Li, Z., Lyu, D., Wang, G., Chen, Y., Chen, L., Li, W., . . . He, G. (2025, June 22). KVO-LLM: Boosting Long-Context Generation Throughput for Batched LLM Inference. 2025 62nd ACM/IEEE Design Automation Conference (DAC), 1-7. https://doi.org/10.1109/DAC63849.2025.11132542
Citace podle Chicago (17th ed.)Li, Zhenyu, Dongxu Lyu, Gang Wang, Yuzhou Chen, Liyan Chen, Wenjie Li, Jianfei Jiang, Yanan Sun, a Guanghui He. "KVO-LLM: Boosting Long-Context Generation Throughput for Batched LLM Inference." 2025 62nd ACM/IEEE Design Automation Conference (DAC) 22 Jun. 2025: 1-7. https://doi.org/10.1109/DAC63849.2025.11132542.
Citace podle MLA (9th ed.)Li, Zhenyu, et al. "KVO-LLM: Boosting Long-Context Generation Throughput for Batched LLM Inference." 2025 62nd ACM/IEEE Design Automation Conference (DAC), 22 Jun. 2025, pp. 1-7, https://doi.org/10.1109/DAC63849.2025.11132542.