The document discusses BERT, a new linguistic model that uses transformers to represent text. BERT performs very well on natural language understanding tasks and there are different methodologies for adapting BERT to specific tasks, including further pre-training, retraining strategies, and multitasking learning. The basic BERT model has 12 transformer blocks and 512 token input, and outputs vector representations. Research is ongoing to improve subject-specific text classification using BERT.