Skip to content

terarachang/QError

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

3 Commits
 
 
 
 
 
 

Repository files navigation

Why Do Some Inputs Break Low-Bit LLM Quantization? (EMNLP 2025)

awq gptq NF

Ting-Yun Chang, Muru Zhang, Jesse Thomason, and Robin Jia

Paper: https://arxiv.org/abs/2506.12044

Data

  • We release $D_\text{large}$ and $D_\text{ctrl}$ under data/
  • The data are tokenized, where each split has shape [1000, 512], and each row corresponds to a FineWeb example with sequence length = 512.
  • To convert them back into texts, run: python read_data.py --split large --quant_type awq3 --model_name Qwen/Qwen2.5-7B

About

Why Do Some Inputs Break Low-Bit LLM Quantization?

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages