arXiv2
Search
Dark
/ Date
/ Name
Aa
W
/ Date
/ Name
"au:"Seungwoo Son"" — arXiv2 Search
Showing 1–5 of 5 results
/ Date
/ Name
Jun 17, 2024
Prefixing Attention Sinks can Mitigate Activation Outliers for Large Language Model Quantization
Feb 21, 2023
The Role of Masking for Efficient Supervised Knowledge Distillation of Vision Transformers
Feb 7, 2026
On the Importance of a Multi-Scale Calibration for Quantization
Feb 4, 2026
TurboBoA: Faster and Exact Attention-aware Quantization without Backpropagation
Feb 2, 2026
Two-Stage Grid Optimization for Group-wise Quantization of LLMs