Attn-QAT: 4-Bit Attention With Quantization-Aware Training — arXiv2