AI hardware

The future of AI depends on whether we can design next generation hardware that better supports the scaling laws. At some point of time, AI model architecture will even be influenced by the design decision on AI hardware. The codesign of AI and hardware will become norm in the future.

Here are some considerations on AI hardware.

References

2025

  1. LUT Tensor Core: A Software-Hardware Co-Design for LUT-Based Low-Bit LLM Inference
    Zhiwen Mo, and 10 more authors
    In Proceedings of the 52nd Annual International Symposium on Computer Architecture (ISCA), 2025
  2. LUT-DLA: Lookup Table as Efficient Extreme Low-Bit Deep Learning Accelerator
    Guoyu Li, and 8 more authors
    In 31st International Symposium on High-Performance Computer Architecture, HPCA, 2025
  3. WaferLLM: A Wafer-Scale LLM Inference System
    Congjie He, and 7 more authors
    In 19th USENIX Symposium on Operating Systems Design and Implementation, OSDI, 2025

2023

  1. OliVe: Accelerating Large Language Models via Hardware-friendly Outlier-Victim Pair Quantization
    Cong Guo, and 8 more authors
    In Proceedings of the 50th Annual International Symposium on Computer Architecture, ISCA, 2023

2022

  1. ANT: Exploiting Adaptive Numerical Data Type for Low-bit Deep Neural Network Quantization
    Cong Guo, and 7 more authors
    In 55th IEEE/ACM International Symposium on Microarchitecture, MICRO, 2022