Skip to content

Latest commit

 

History

History
 
 

trn1-bert-fine-tuning-on-sagemaker

Folders and files

NameName
Last commit message
Last commit date

parent directory

..
 
 
 
 
 
 

Fine tuning BERT base model from HuggingFace on Amazon SageMaker

Overview

This tutotial uses the Hugging Face transformers and datasets libraries with Amazon SageMaker to fine-tune a pre-trained BERT base model on binary text classification with using Hugging Face Trainer API.

A pre-trained BERT base model is available in the transformers library from Hugging Face. You’ll be fine-tuning this pre-trained model using the Amazon Polarity dataset which classify the content into either positive or negative feedback.

You will see how Neuron Persistent Cache can be used for SageMaker Training jobs. At the end of the training job, Neuron Persistent Cache is uploaded to S3 so the cache can be reused for the another training jobs.

Getting started

Run bert-base-uncased-amazon-polarity.ipynb either on SageMaker notebook instance or SageMaker Studio Notebook.

You can set up your SageMaker Notebook instance by following the Get Started with Amazon SageMaker Notebook Instances or SageMaker Studio Notebook by following the Use Amazon SageMaker Studio Notebooks

The notebook was tested on a ml.t3.medium SageMaker Studio Notebook with Data Science image.