With the release of BERT (Bidirectional Encoder Representations from Transformers https://arxiv.org/abs/1810.04805), Google has shown that large Transformer models can lead to state-of-the-art results for various common NLP (Natural Language Processing) tasks. Through transfer learning, it’s possible to take a pre-trained language model and re-purpose (fine-tune) it for a second related…