
Fine-tuning BERT model for arbitrarily long texts, Part 2
Author: Michał Brzozowski This is part 2 of our series about fine-tuning BERT: if you want to read the first part, go to this link, and if you want to use the code, go to our GitHub. Fine-tuning the pre-trained BERT on longer texts Now, this is the time to address the elephant in the room of the previous approach. We were lucky to find the already fine-tuned model for our IMDB dataset. However, more often, we are in a more unfortunate situation when we have the labelled dataset and we need to fine-tune the classifier from