In this video, we'll explore the performance differences when running Large Language Models (LLMs) in Ollama using both the CPU and GPU. Watch as I demonstrate a live sample in C# using Microsoft.Extensions.AI to run Ollama inside a Docker container. Curious to see how these models perform locally? Let's dive in and compare the results!
Useful links:
.NET Video Analyzer repository: aka.ms/netaivi...
Ollama in Docker: ollama.com/blo...
.NET & AI Show: • .NET AI Community Stan...
Негізгі бет GPU vs CPU: Running Small Language Models with Ollama & C#
Пікірлер: 6