This document summarizes a research paper on efficient Transformer-based speech enhancement using long frames and STFT magnitudes. It introduces the task of speech enhancement and issues with learned-domain approaches. The proposed method uses STFT magnitudes as input to an encoder-decoder architecture with a Transformer masker. Experiments show the method achieves equivalent quality and intelligibility scores compared to learned features, while reducing computations by 8x for 10-second utterances. The conclusions are that using STFT magnitudes enables efficient, high-quality speech enhancement on embedded devices.