Search Results - "Information systems Data management systems Data structures Data layout Data compression"

Refine Results
  1. 1

    Data Reduction Techniques for Simulation, Visualization and Data Analysis by Li, S., Marsaglia, N., Garth, C., Woodring, J., Clyne, J., Childs, H.

    ISSN: 0167-7055, 1467-8659
    Published: Oxford Blackwell Publishing Ltd 01.09.2018
    Published in Computer graphics forum (01.09.2018)
    “…Data reduction is increasingly being applied to scientific data for numerical simulations, scientific visualizations and data analyses. It is most often used…”
    Get full text
    Journal Article
  2. 2

    Base-delta-immediate compression: Practical data compression for on-chip caches by Pekhimenko, Gennady, Seshadri, Vivek, Mutlu, Onur, Kozuch, Michael A., Gibbons, Phillip B., Mowry, Todd C.

    Published: ACM 01.09.2012
    “…Cache compression is a promising technique to increase on-chip cache capacity and to decrease on-chip and off-chip bandwidth usage. Unfortunately, directly…”
    Get full text
    Conference Proceeding
  3. 3

    Quantum Neural Network Compression by Hu, Zhirui, Dong, Peiyan, Wang, Zhepeng, Lin, Youzuo, Wang, Yanzhi, Jiang, Weiwen

    ISSN: 1558-2434
    Published: ACM 29.10.2022
    “…Model compression, such as pruning and quantization, has been widely applied to optimize neural networks on resource-limited classical devices. Recently, there…”
    Get full text
    Conference Proceeding
  4. 4

    Faster and Stronger Lossless Compression with Optimized Autoregressive Framework by Mao, Yu, Li, Jingzong, Cui, Yufei, Xue, Jason Chun

    Published: IEEE 09.07.2023
    “…Neural AutoRegressive (AR) framework has been applied in general-purpose lossless compression recently to improve compression performance. However, this paper…”
    Get full text
    Conference Proceeding
  5. 5

    NN-AdderNet: Nonnegative and Sparse Weight Optimization Towards Ultra-Low Bitwidth AdderNet Quantization and Compression by Zhang, Yunxiang, Sun, Gengchen, Fang, Lizhi, Sun, Biao, Zhao, Wenfeng

    Published: IEEE 22.06.2025
    “…Emerging efficient deep neural network (DNN) models, such as AdderNet, have shown great promise in significantly improving hardware efficiency compared to…”
    Get full text
    Conference Proceeding
  6. 6

    PacTrain: Pruning and Adaptive Sparse Gradient Compression for Efficient Collective Communication in Distributed Deep Learning by Wang, Yisu, Wu, Ruilong, Li, Xinjiao, Kutscher, Dirk

    Published: IEEE 22.06.2025
    “…Large-scale deep neural networks (DNN) exhibit excellent performance for various tasks. As DNNs and datasets grow, distributed training becomes extremely…”
    Get full text
    Conference Proceeding
  7. 7

    BLOOM: Bit-Slice Framework for DNN Acceleration with Mixed-Precision by Liu, Fangxin, Yang, Ning, Wang, Zongwu, Zhu, Xuanpeng, Yao, Haidong, Xiong, Xiankui, Jiang, Li, Guan, Haibing

    Published: IEEE 22.06.2025
    “…Deep neural networks (DNNs) have revolutionized numerous AI applications, but their vast model sizes and limited hardware resources present significant…”
    Get full text
    Conference Proceeding
  8. 8

    BirdMoE: Reducing Communication Costs for Mixture-of-Experts Training Using Load-Aware Bi-random Quantization by Wu, Donglei, Yang, Weihao, Zou, Xiangyu, Jia, Jinda, Tao, Dingwen, Xia, Wen, Tian, Zhihong

    Published: IEEE 22.06.2025
    “…Mixture-of-Experts (MoE) model parallelism is prevalent in training Large Language Models (e.g., ChatGPT). However, the intensive all-to-all collective…”
    Get full text
    Conference Proceeding
  9. 9

    MemSens: Significantly Reducing Memory Overhead in Adjoint Sensitivity Analysis Using Novel Error-Bounded Lossy Compression by Li, Chenxi, Feng, Yihang, Deng, Fuxing, Tao, Dingwen, Liu, Weifeng, Jin, Zhou

    Published: IEEE 22.06.2025
    “…Adjoint sensitivity analysis is an exceptionally efficient method for computing the gradient of an objective function with respect to given parameters, playing…”
    Get full text
    Conference Proceeding
  10. 10

    PISA: Efficient Precision-Slice Framework for LLMs with Adaptive Numerical Type by Yang, Ning, Wang, Zongwu, Sun, Qingxiao, Lu, Liqiang, Liu, Fangxin

    Published: IEEE 22.06.2025
    “…Large language models (LLMs) have transformed numerous AI applications, with on-device deployment becoming increasingly important for reducing cloud computing…”
    Get full text
    Conference Proceeding
  11. 11

    SNAPPIX: Efficient-Coding-Inspired In-Sensor Compression for Edge Vision by Lin, Weikai, Ma, Tianrui, Boloor, Adith, Feng, Yu, Xing, Ruofan, Zhang, Xuan, Zhu, Yuhao

    Published: IEEE 22.06.2025
    “…Energy-efficient image acquisition on the edge is crucial for enabling remote sensing applications where the sensor node has weak compute capabilities and must…”
    Get full text
    Conference Proceeding
  12. 12

    ClusterKV: Manipulating LLM KV Cache in Semantic Space for Recallable Compression by Liu, Guangda, Li, Chengwei, Zhao, Jieru, Zhang, Chenqi, Guo, Minyi

    Published: IEEE 22.06.2025
    “…Large Language Models (LLMs) have been widely deployed in a variety of applications, and the context length is rapidly increasing to handle tasks such as…”
    Get full text
    Conference Proceeding
  13. 13

    DCDiff: Enhancing JPEG Compression via Diffusion-based DC Coefficients Estimation by Zhang, Ziyuan, Qiu, Han, Zhang, Tianwei, Chen, Bin, Zhang, Chao

    Published: IEEE 22.06.2025
    “…JPEG is the most widely-used image compression method on low-cost cameras which cannot support learning-based compressors. One promising approach to enhance…”
    Get full text
    Conference Proceeding
  14. 14

    Late Breaking Results: Less Sense Makes More Sense: In-Sensor Compressive Learning for Efficient Machine Vision by Liang, Yiwen, Cao, Weidong

    Published: IEEE 22.06.2025
    “…Integrating deep learning and image sensors has significantly transformed machine vision applications. Yet, conventional highresolution image acquisition…”
    Get full text
    Conference Proceeding
  15. 15

    CognitiveArm: Enabling Real-Time EEG-Controlled Prosthetic Arm Using Embodied Machine Learning by Basit, Abdul, Nawaz, Maha, Rehman, Saim, Shafique, Muhammad

    Published: IEEE 22.06.2025
    “…Efficient control of prosthetic limbs via non-invasive brain-computer interfaces (BCIs) requires advanced EEG processing capabilities-including pre-filtering,…”
    Get full text
    Conference Proceeding
  16. 16

    APSQ: Additive Partial Sum Quantization with Algorithm-Hardware Co-Design by Tan, Yonghao, Dong, Pingcheng, Wu, Yongkun, Liu, Yu, Liu, Xuejiao, Luo, Peng, Liu, Shih-Yang, Huang, Xijie, Zhang, Dong, Liang, Luhong, Cheng, Kwang-Ting

    Published: IEEE 22.06.2025
    “…DNN accelerators, significantly advanced by model compression and specialized dataflow techniques, have marked considerable progress. However, the frequent…”
    Get full text
    Conference Proceeding
  17. 17

    Easz: An Agile Transformer-based Image Compression Framework for Resource-constrained IoTs by Mao, Yu, Li, Jingzong, Wang, Jun, Xu, Hong, Kuo, Tei-Wei, Guan, Nan, Xue, Chun Jason

    Published: IEEE 22.06.2025
    “…Neural image compression, necessary in various machine-to-machine communication scenarios, suffers from its heavy encode-decode structures and inflexibility in…”
    Get full text
    Conference Proceeding
  18. 18

    MILLION: MasterIng Long-Context LLM Inference Via Outlier-Immunized KV Product QuaNtization by Wang, Zongwu, Xu, Peng, Liu, Fangxin, Hu, Yiwei, Sun, Qingxiao, Li, Gezi, Li, Cheng, Wang, Xuan, Jiang, Li, Guan, Haibing

    Published: IEEE 22.06.2025
    “…Large language models (LLMs) are increasingly utilized for complex tasks requiring longer context lengths, with some models supporting up to 128 K or 1 M…”
    Get full text
    Conference Proceeding
  19. 19

    KVO-LLM: Boosting Long-Context Generation Throughput for Batched LLM Inference by Li, Zhenyu, Lyu, Dongxu, Wang, Gang, Chen, Yuzhou, Chen, Liyan, Li, Wenjie, Jiang, Jianfei, Sun, Yanan, He, Guanghui

    Published: IEEE 22.06.2025
    “…With the widespread deployment of long-context large language models (LLMs), efficient and high-quality generation is becoming increasingly important. Modern…”
    Get full text
    Conference Proceeding
  20. 20

    DuQTTA: Dual Quantized Tensor-Train Adaptation with Decoupling Magnitude-Direction for Efficient Fine-Tuning of LLMs by Dong, Haoyan, Chen, Hai-Bao, Chang, Jingjing, Yang, Yixin, Gao, Ziyang, Ji, Zhigang, Wang, Runsheng, Huang, Ru

    Published: IEEE 22.06.2025
    “…Recent parameter-efficient fine-tuning (PEFT) techniques have enabled large language models (LLMs) to be efficiently fine-tuned for specific tasks, while…”
    Get full text
    Conference Proceeding