강의

멘토링

커뮤니티

AI Technology

/

AI Agent Development

[VLM101] Building a Multimodal Chatbot with Fine-tuning (feat.MCP / RunPod)

This is an introductory course for understanding the concept and application methods of Vision-Language Models (VLM), and practicing running the LLaVA model in an Ollama-based environment while integrating it with MCP (Model Context Protocol). This course covers the principles of multimodal models, quantization, service development, and integrated demo development, providing a balanced mix of theory and hands-on practice.

(4.6) 18 reviews

119 learners

Level Basic

Course period Unlimited

  • dreamingbumblebee
Vision Transformer
Vision Transformer
transformer
transformer
Llama
Llama
Model Context Protocol
Model Context Protocol
Vision Transformer
Vision Transformer
transformer
transformer
Llama
Llama
Model Context Protocol
Model Context Protocol