The document presents a method for improving model quantization in vision transformers by introducing a power-of-two factor for layer normalization and a log-int-softmax approach for softmax layers. It addresses issues of accuracy degradation associated with prior quantization methods while preserving high inference speed. Experiments demonstrate that the proposed quantization leads to better accuracy in image classification and object detection tasks compared to existing techniques.