kaggle competition : https://www.kaggle.com/c/contradictory-my-dear-watson
These are the notebooks that:
- fine tune a Bert model,
- fine tune a RoBERTa model,
- add a Global Average Pooling to an xlm-roberta-large model output hidden layer and fine tune.
Datasets used for fine tuning are glue/mnli, and xnli.