This document summarizes a research paper that proposes a new method called TTT-MAE (Test-Time Training with Masked Autoencoders) to address the problem of domain shift in visual recognition tasks. TTT-MAE uses masked autoencoders as the self-supervised pretext task in test-time training, instead of rotation prediction as used in previous work. Experimental results on datasets like ImageNet-C and ImageNet-R show that TTT-MAE achieves higher performance gains than prior methods under different types of distribution shifts. However, TTT-MAE is slower at test time than directly applying a fixed model. Future work could focus on improving efficiency and generalizing the approach to other tasks