Providing context for fine-tuning LLMs and walking through why I chose to fine tune an instance of GPT-3.5-Turbo, my specific use case exploration, and all of the steps I took to get there!
The Code: github.com/ALucek/ft-openai-video
Output Comparison: docs.google.com/spreadsheets/d/1-8SzT4tZBS0Nti5oEz…
GoEmotions Dataset: huggingface.co/datasets/go_emotions?row=0
Fine Tuning Graphic: miro.medium.com/v2/resize:fit:1400/1*JSJBBnslBE9S5…
OpenAI Fine Tuning Documentation: platform.openai.com/docs/guides/fine-tuning
OpenAI Cookbook - Fine Tuning Data Prep : cookbook.openai.com/examples/chat_finetuning_data_…
Chapters:
00:00 - Intro
00:21 - What is Fine Tuning?
01:40 - OpenAI Fine Tuning Documentation
03:51 - Why I Fine Tune Models
05:04 - Dataset of Interest
06:44 - Training Data Format
07:18 - Loading and Examining the Dataset
10:24 - Generating a GPT-4-T Baseline
13:00 - Formatting Dataset into JSONL
15:47 - Validating Training Data & Price Estimation
19:13 - Starting the Fine Tuning Job
19:47 - Fine Tuned Model Metrics Overview
20:31 - Talking About Training Loss
23:29 - Running Inference With the Fine Tuned Model
24:02 - Output Comparisons & Discussion
27:01 - Outro