EASIEST Way to Fine-Tune a LLM and Use It With Ollama

EASIEST Way to Fine-Tune a LLM and Use It With Ollama

306.745 Lượt nghe
EASIEST Way to Fine-Tune a LLM and Use It With Ollama
In this video, we go over how you can fine-tune Llama 3.1 and run it locally on your machine using Ollama! We use the open source repository "Unsloth" to do all of the fine tuning on an SQL dataset! Throughout this video, we discuss the ins and outs of what fine-tuning an LLM is, how you can work with your data so that the LLM can process it and then importing it into Ollama for you to run locally on your machine! Ollama is available on all platforms! Dataset used: https://huggingface.co/datasets/gretelai/synthetic_text_to_sql Ollama: https://github.com/ollama/ollama Unsloth: https://github.com/unslothai/unsloth Make sure you follow for more content! ___________________________ Try Warp now for FREE 👉 bit.ly/warpdotdev Twitter 🐦 https://twitter.com/warpdotdev TikTok 📱 https://www.tiktok.com/@warp.dev TIMESTAMPS 0:00 Intro 0:10 Getting the dataset 0:45 The Tech Stack 1:18 Installing Dependencies 1:48 Fast Language Model Explained 2:35 LORA Adaptes Explained 3:02 Converting your data to fine-tune 3:36 Training the Model.... 4:01 Converting to Ollama compatibility 4:11 Creating a Modelfile for Ollama 4:50 Final Output! 5:01 Check out Ollama in 2 minutes!