Get in Touch

Course Outline

Introduction to Optimising Large Models

  • Overview of large model architectures
  • Challenges in fine-tuning large models
  • Importance of cost-effective optimisation

Distributed Training Techniques

  • Introduction to data and model parallelism
  • Frameworks for distributed training: PyTorch and TensorFlow
  • Scaling across multiple GPUs and nodes

Model Quantisation and Pruning

  • Understanding quantisation techniques
  • Applying pruning to reduce model size
  • Trade-offs between accuracy and efficiency

Hardware Optimisation

  • Choosing the right hardware for fine-tuning tasks
  • Optimising GPU and TPU utilisation
  • Using specialised accelerators for large models

Efficient Data Management

  • Strategies for managing large datasets
  • Preprocessing and batching for performance
  • Data augmentation techniques

Deploying Optimised Models

  • Techniques for deploying fine-tuned models
  • Monitoring and maintaining model performance
  • Real-world examples of optimised model deployment

Advanced Optimisation Techniques

  • Exploring low-rank adaptation (LoRA)
  • Using adapters for modular fine-tuning
  • Future trends in model optimisation

Summary and Next Steps

Requirements

  • Experience with deep learning frameworks like PyTorch or TensorFlow
  • Familiarity with large language models and their applications
  • Understanding of distributed computing concepts

Audience

  • Machine learning engineers
  • Cloud AI specialists
 21 Hours

Number of participants


Price per participant

Provisional Upcoming Courses (Require 5+ participants)

Related Categories