Welcome to this repository dedicated to fine-tuning Llama2 and Llama3 models on Amazon SageMaker using Transformer Reinforcement Learning (TRL) methods and custom configurations for generating recommendation systems.
This repository provides resources and guidelines for training Llama2 and Llama3 models with advanced TRL fine-tuning techniques. Our goal is to enhance the capabilities of these models for specific use-cases, for examle recommendation systems, leveraging the latest in machine learning to provide highly relevant and accurate recommendations.
Before you begin, ensure you have the following prerequisites installed:
- Python 3.8 or later
- PyTorch 1.8 or later
- Transformers library
- Datasets library
- A suitable CUDA-enabled GPU
Clone this repository to your local machine using:
git clone https://github.com/mccartni-aws/llama-training.git
- Run the
llama3_orpo_sft.ipynb
notebook, accompanied by the Medium blog titled Fine-tune Llama3 with ORPO in Amazon SageMaker Studio.