Get in Touch

Course Outline

Introduction to Huawei CloudMatrix

  • The CloudMatrix ecosystem and deployment workflow
  • Supported models, formats, and deployment modes
  • Typical use cases and supported chipsets

Preparing Models for Deployment

  • Exporting models from training tools (MindSpore, TensorFlow, PyTorch)
  • Using ATC (Ascend Tensor Compiler) for format conversion
  • Static versus dynamic shape models

Deploying to CloudMatrix

  • Service creation and model registration
  • Deploying inference services via the UI or CLI
  • Routing, authentication, and access control

Serving Inference Requests

  • Batch versus real-time inference flows
  • Data preprocessing and postprocessing pipelines
  • Invoking CloudMatrix services from external applications

Monitoring and Performance Tuning

  • Deployment logs and request tracking
  • Resource scaling and load balancing
  • Latency tuning and throughput optimisation

Integration with Enterprise Tools

  • Connecting CloudMatrix with OBS and ModelArts
  • Leveraging workflows and model versioning
  • CI/CD for model deployment and rollback

End-to-End Inference Pipeline

  • Deploying a complete image classification pipeline
  • Benchmarking and validating accuracy
  • Simulating failover scenarios and system alerts

Summary and Next Steps

Requirements

  • A solid understanding of AI model training workflows
  • Hands-on experience with Python-based machine learning frameworks
  • Basic familiarity with cloud deployment concepts

Audience

  • AI operations teams
  • Machine learning engineers
  • Cloud deployment specialists working with Huawei infrastructure
 21 Hours

Number of participants


Price per participant

Testimonials (2)

Provisional Upcoming Courses (Require 5+ participants)

Related Categories