Get in Touch

Course Outline

Performance Concepts and Metrics

  • Latency, throughput, power consumption, and resource utilization
  • Distinguishing between system-level and model-level bottlenecks
  • Differentiating profiling needs for inference versus training

Profiling on Huawei Ascend

  • Leveraging CANN Profiler and MindInsight
  • Analyzing kernel and operator diagnostics
  • Understanding offload patterns and memory mapping

Profiling on Biren GPU

  • Utilizing Biren SDK performance monitoring features
  • Focusing on kernel fusion, memory alignment, and execution queues
  • Conducting power and temperature-aware profiling

Profiling on Cambricon MLU

  • Employing BANGPy and Neuware performance tools
  • Gaining kernel-level visibility and interpreting logs
  • Integrating the MLU profiler with deployment frameworks

Graph and Model-Level Optimization

  • Strategies for graph pruning and quantization
  • Operator fusion and restructuring the computational graph
  • Standardizing input sizes and tuning batch parameters

Memory and Kernel Optimization

  • Optimizing memory layout and reuse patterns
  • Managing buffers efficiently across different chipsets
  • Applying platform-specific kernel tuning techniques

Cross-Platform Best Practices

  • Achieving performance portability through abstraction strategies
  • Developing shared tuning pipelines for multi-chip setups
  • Case study: Tuning an object detection model across Ascend, Biren, and MLU

Summary and Next Steps

Requirements

  • Hands-on experience with AI model training or deployment workflows
  • Solid understanding of GPU/MLU computing principles and model optimization techniques
  • Familiarity with basic performance profiling tools and key metrics

Target Audience

  • Performance engineers
  • Machine learning infrastructure teams
  • AI system architects
 21 Hours

Upcoming Courses

Related Categories