MENU

Fun & Interesting

5. Comparing Quantizations of the Same Model - Ollama Course

Matt Williams 14,084 6 months ago
Video Not Working? Fix It Now

Welcome back to the Ollama course! In this lesson, we dive into the fascinating world of AI model quantization. Using variations of the llama3.1 model, we explore how different quantization levels affect performance and output quality. Through this video, you'll gain a deeper understanding of how to choose the right quantization for how you use AI models, ensuring you get the best performance and results for your specific needs. Don't forget to subscribe for more lessons in this free Ollama course! Thanks for watching! My Links 🔗 👉🏻 Subscribe (free): https://www.youtube.com/technovangelist 👉🏻 Join and Support: https://www.youtube.com/channel/UCHaF9kM2wn8C3CLRwLkC2GQ/join 👉🏻 Newsletter: https://technovangelist.substack.com/subscribe 👉🏻 Twitter: https://www.twitter.com/technovangelist 👉🏻 Discord: https://discord.gg/uS4gJMCRH2 👉🏻 Patreon: https://patreon.com/technovangelist 👉🏻 Instagram: https://www.instagram.com/technovangelist/ 👉🏻 Threads: https://www.threads.net/@technovangelist?xmt=AQGzoMzVWwEq8qrkEGV8xEpbZ1FIcTl8Dhx9VpF1bkSBQp4 👉🏻 LinkedIn: https://www.linkedin.com/in/technovangelist/ 👉🏻 All Source Code: https://github.com/technovangelist/videoprojects Want to sponsor this channel? Let me know what your plans are here: https://www.technovangelist.com/sponsor 00:00 - Start with an example 00:24 - Introduction 00:56 - Lots of claims on the Discord 01:26 - Intro to the app 01:57 - Where to find the code 02:20 - Grab a few quantizations 02:57 - You should regularly pull the models again 03:30 - Back to the Black Hole answers 04:39 - The classic logic problem 05:35 - How about function calling 08:31 - How about for prompts with more reasoning 09:01 - Are those questions stupid? 09:30 - Which quant to use?

Comment