Deploying and Optimizing LLMs with Ollama Training Course
Ollama offers an efficient way to deploy and run large language models (LLMs) locally or in production environments, providing control over performance, cost, and security.
This instructor-led, live training (online or on-site) is designed for intermediate-level professionals who wish to deploy, optimise, and integrate LLMs using Ollama.
By the end of this training, participants will be able to:
- Set up and deploy LLMs using Ollama.
- Optimise AI models for performance and efficiency.
- Leverage GPU acceleration to improve inference speeds.
- Integrate Ollama into workflows and applications.
- Monitor and maintain AI model performance over time.
Course Format
- Interactive lectures and discussions.
- Extensive exercises and practice opportunities.
- Hands-on implementation in a live-lab environment.
Course Customisation Options
- To request a customised training session for this course, please contact us to arrange.
Course Outline
Introduction to Ollama for LLM Deployment
- Overview of Ollama's capabilities
- Advantages of local AI model deployment
- Comparison with cloud-based AI hosting solutions
Setting Up the Deployment Environment
- Installing Ollama and required dependencies
- Configuring hardware and GPU acceleration
- Dockerising Ollama for scalable deployments
Deploying LLMs with Ollama
- Loading and managing AI models
- Deploying Llama 3, DeepSeek, Mistral, and other models
- Creating APIs and endpoints for AI model access
Optimising LLM Performance
- Fine-tuning models for efficiency
- Reducing latency and improving response times
- Managing memory and resource allocation
Integrating Ollama into AI Workflows
- Connecting Ollama to applications and services
- Automating AI-driven processes
- Using Ollama in edge computing environments
Monitoring and Maintenance
- Tracking performance and debugging issues
- Updating and managing AI models
- Ensuring security and compliance in AI deployments
Scaling AI Model Deployments
- Best practices for handling high workloads
- Scaling Ollama for enterprise use cases
- Future advancements in local AI model deployment
Summary and Next Steps
Requirements
- Basic experience with machine learning and AI models
- Familiarity with command-line interfaces and scripting
- Understanding of deployment environments (local, edge, cloud)
Audience
- AI engineers optimising local and cloud-based AI deployments
- ML practitioners deploying and fine-tuning LLMs
- DevOps specialists managing AI model integration
Open Training Courses require 5+ participants.
Deploying and Optimizing LLMs with Ollama Training Course - Booking
Deploying and Optimizing LLMs with Ollama Training Course - Enquiry
Deploying and Optimizing LLMs with Ollama - Consultancy Enquiry
Provisional Upcoming Courses (Require 5+ participants)
Related Courses
Advanced Ollama Model Debugging & Evaluation
35 HoursAdvanced Ollama Model Debugging & Evaluation is an in-depth course focused on diagnosing, testing, and measuring model behaviour when running local or private Ollama deployments.
This instructor-led, live training (online or on-site) is aimed at advanced-level AI engineers, ML Ops professionals, and QA practitioners who wish to ensure reliability, fidelity, and operational readiness of Ollama-based models in production.
By the end of this training, participants will be able to:
- Perform systematic debugging of Ollama-hosted models and reliably reproduce failure modes.
- Design and execute robust evaluation pipelines using both quantitative and qualitative metrics.
- Implement observability (logs, traces, metrics) to monitor model health and detect drift.
- Automate testing, validation, and regression checks integrated into CI/CD pipelines.
Course Format
- Interactive lectures and discussion.
- Hands-on labs and debugging exercises using Ollama deployments.
- Case studies, group troubleshooting sessions, and automation workshops.
Course Customisation Options
- To request a customised training for this course, please contact us to arrange.
Building Private AI Workflows with Ollama
14 HoursThis instructor-led, live training in New Zealand (online or on-site) is aimed at advanced-level professionals who wish to implement secure and efficient AI-driven workflows using Ollama.
By the end of this training, participants will be able to:
- Deploy and configure Ollama for private AI processing.
- Integrate AI models into secure enterprise workflows.
- Optimise AI performance while maintaining data privacy.
- Automate business processes with on-premises AI capabilities.
- Ensure compliance with enterprise security and governance policies.
Fine-Tuning and Customizing AI Models on Ollama
14 HoursThis instructor-led, live training in New Zealand (online or on-site) is designed for advanced-level professionals who wish to fine-tune and customise AI models on Ollama to achieve enhanced performance and support domain-specific applications.
By the end of this training, participants will be able to:
- Set up an efficient environment for fine-tuning AI models on Ollama.
- Prepare datasets for supervised fine-tuning and reinforcement learning.
- Optimise AI models for performance, accuracy, and efficiency.
- Deploy customised models in production environments.
- Evaluate model improvements and ensure robustness.
Multimodal Applications with Ollama
21 HoursOllama is a platform that enables running and fine-tuning large language and multimodal models locally.
This instructor-led, live training (online or onsite) is aimed at advanced-level ML engineers, AI researchers, and product developers who wish to build and deploy multimodal applications with Ollama.
By the end of this training, participants will be able to:
- Set up and run multimodal models with Ollama.
- Integrate text, image, and audio inputs for real-world applications.
- Build document understanding and visual QA systems.
- Develop multimodal agents capable of reasoning across modalities.
Course Format
- Interactive lectures and discussion.
- Hands-on practice using real multimodal datasets.
- Live-lab implementation of multimodal pipelines with Ollama.
Course Customisation Options
- To request a customised training session for this course, please contact us to make arrangements.
Getting Started with Ollama: Running Local AI Models
7 HoursThis instructor-led, live training in New Zealand (delivered online or on-site) is designed for beginner-level professionals who wish to install, configure, and use Ollama to run AI models on their local machines.
By the end of this training, participants will be able to:
- Understand the fundamentals of Ollama and its capabilities.
- Set up Ollama to run local AI models.
- Deploy and interact with LLMs using Ollama.
- Optimise performance and resource usage for AI workloads.
- Explore use cases for local AI deployment across various industries.
Ollama & Data Privacy: Secure Deployment Patterns
14 HoursOllama is a platform that enables the local execution of large language and multimodal models while supporting secure deployment strategies.
This instructor-led, live training (available online or on-site) is designed for intermediate-level professionals seeking to deploy Ollama with robust data privacy and regulatory compliance measures.
By the end of this training, participants will be able to:
- Deploy Ollama securely in containerised and on-premises environments.
- Apply differential privacy techniques to protect sensitive data.
- Implement secure logging, monitoring, and auditing practices.
- Enforce data access controls aligned with compliance requirements.
Course Format
- Interactive lecture and discussion.
- Hands-on labs featuring secure deployment patterns.
- Compliance-focused case studies and practical exercises.
Course Customisation Options
- To request a customised training session for this course, please contact us to arrange.
Ollama Applications in Finance
14 HoursOllama is a lightweight platform for running large language models locally.
This instructor-led, live training (online or on-site) is designed for intermediate-level finance practitioners and IT personnel who wish to implement, customise, and operationalise Ollama-based AI solutions within financial environments.
By completing this training, participants will gain the skills needed to:
- Deploy and configure Ollama for secure use in financial operations.
- Integrate local LLMs into analytical and reporting workflows.
- Adapt models to finance-specific terminology and tasks.
- Apply security, privacy, and compliance best practices.
Course Format
- Interactive lectures and discussions.
- Hands-on exercises using real financial data.
- Live-lab implementation of finance-focused scenarios.
Course Customisation Options
- To request a customised training session for this course, please contact us to arrange.
Ollama Applications in Healthcare
14 HoursOllama is a lightweight platform for running large language models locally.
This instructor-led, live training (online or on-site) is designed for intermediate-level healthcare practitioners and IT teams who wish to deploy, customise, and operationalise Ollama-based AI solutions within clinical and administrative settings.
Upon completing this training, participants will be able to:
- Install and configure Ollama for secure use in healthcare environments.
- Integrate local LLMs into clinical workflows and administrative processes.
- Customise models for healthcare-specific terminology and tasks.
- Apply best practices for privacy, security, and regulatory compliance.
Course Format
- Interactive lectures and discussions.
- Hands-on demonstrations and guided exercises.
- Practical implementation within a sandboxed healthcare simulation environment.
Course Customisation Options
- To request a customised version of this training, please contact us to arrange.
Ollama: Self-Hosted Large Language Models Replacing OpenAI and Claude APIs
14 HoursOllama is an open-source utility designed to run large language models locally on both consumer-grade and enterprise hardware. It simplifies complex processes such as model quantization, GPU allocation, and API servicing into a single command-line interface. This allows organisations to self-host LLMs like Llama, Mistral, and Qwen without needing to send prompts or data to providers like OpenAI, Anthropic, or Google.
Ollama for Responsible AI and Governance
14 HoursOllama is a platform for running large language and multimodal models locally, supporting governance and responsible AI practices.
This instructor-led, live training (online or on-site) is aimed at intermediate to advanced professionals seeking to embed fairness, transparency, and accountability into Ollama-powered applications.
By the end of this training, participants will be able to:
- Apply responsible AI principles in Ollama deployments.
- Implement content filtering and bias mitigation strategies.
- Design governance workflows for AI alignment and auditability.
- Establish monitoring and reporting frameworks for compliance.
Course Format
- Interactive lectures and discussions.
- Hands-on labs for designing governance workflows.
- Case studies and compliance-focused exercises.
Course Customisation Options
- To request a customised training session for this course, please contact us to arrange.
Ollama Scaling & Infrastructure Optimization
21 HoursOllama is a platform for running large language and multimodal models locally and at scale.
This instructor-led, live training (online or onsite) is aimed at intermediate-level to advanced-level engineers who wish to scale Ollama deployments for multi-user, high-throughput, and cost-efficient environments.
By the end of this training, participants will be able to:
- Configure Ollama for multi-user and distributed workloads.
- Optimise GPU and CPU resource allocation.
- Implement autoscaling, batching, and latency reduction strategies.
- Monitor and optimise infrastructure for performance and cost efficiency.
Course Format
- Interactive lectures and discussions.
- Hands-on deployment and scaling labs.
- Practical optimisation exercises in live environments.
Course Customisation Options
- To request a customised training session for this course, please contact us to arrange.
Prompt Engineering Mastery with Ollama
14 HoursOllama is a platform that enables running large language and multimodal models locally.
This instructor-led, live training (online or onsite) is aimed at intermediate-level practitioners who wish to master prompt engineering techniques to optimise Ollama outputs.
By the end of this training, participants will be able to:
- Design effective prompts for diverse use cases.
- Apply techniques such as priming and chain-of-thought structuring.
- Implement prompt templates and context management strategies.
- Build multi-stage prompting pipelines for complex workflows.
Course Format
- Interactive lecture and discussion.
- Hands-on exercises in prompt design.
- Practical implementation within a live-lab environment.
Course Customisation Options
- To request a customised training session for this course, please contact us to arrange.