[VLM101] Building a Multimodal Chatbot with Fine-tuning (feat.MCP / RunPod)
dreamingbumblebee
$59.40
Basic / Vision Transformer, transformer, Llama, Model Context Protocol
4.6
(28)
This is an introductory course for understanding the concept and application methods of Vision-Language Models (VLM), and practicing running the LLaVA model in an Ollama-based environment while integrating it with MCP (Model Context Protocol). This course covers the principles of multimodal models, quantization, service development, and integrated demo development, providing a balanced mix of theory and hands-on practice.
Basic
Vision Transformer, transformer, Llama










![Just 1 hour! Creating 'My Own AI Senior Developer' to install on my computer (Antigravity Vibe Coding) [Source code provided]Course Thumbnail](https://cdn.inflearn.com/public/files/courses/340332/cover/ai/3/e87ee52b-1099-42db-a384-64ab8c725470.png?w=420)