logo


your one source for IT & AV

Training Presentation Systems Services & Consulting Cloud Services Purchase Client Center Computer Museum
Arrow Course Schedule | Classroom Rentals | Student Information | Free Seminars | Client Feedback | Partners | Survey | Standby Discounts

DCLLM - Implementing and Operating LLM Inferencing Systems with Cisco and NVIDIA Data Center Technologies

SS Course: GK860052

Course Overview

TOP

This comprehensive training equips participants with the knowledge and skills required to design, deploy, and optimize Large Language Models (LLMs) using NVIDIA GPUs and Cisco infrastructure. Through in-depth modules, hands-on labs, and real-world case studies, participants will learn how to manage data preparation, build scalable pipelines, optimize performance, ensure security, and migrate from cloud to on-premises deployments. The course provides a holistic approach to mastering the technical complexities of LLM systems while leveraging cutting-edge NVIDIA and Cisco technologies for scalability, efficiency, and security.

                                                                  

Scheduled Classes

TOP
12/08/25 - GVT - Virtual Classroom - Virtual Instructor-Led
02/09/26 - GVT - Virtual Classroom - Virtual Instructor-Led
04/13/26 - GVT - Virtual Classroom - Virtual Instructor-Led

Outline

TOP

Module 1: Large Language Model (LLM) Foundations

Objectives:

  • Understand the architecture and mathematical principles of LLMs.
  • Learn design trade-offs for scalability and performance.
  • Explore emerging innovations in LLM development.

Topics:

  • Transformer architecture, self-attention mechanism, and positional encoding.
  • Types of LLMs: Encoder-only, decoder-only, and encoder-decoder.
  • Training objectives: Masked language modeling (MLM), causal language modeling (CLM), and sequence-to-sequence modeling.
  • Scaling laws and challenges: Parameter size, dataset size, and compute.
  • Emerging architectures: Reformer, Longformer, and multi-modal LLMs.

Module 2: Data Collection and Preparation for LLM Training

Objectives:

  • Understand data requirements for LLMs and their impact on performance.
  • Learn techniques for sourcing, cleaning, and managing large-scale datasets.
  • Explore NVIDIA and Cisco tools for efficient data handling.

Topics:

  • Data sourcing: Open-source, proprietary, and domain-specific datasets.
  • Preprocessing: Cleaning, deduplication, tokenization, and filtering.
  • Data management: Sharding, scalable storage, and high-speed data transfer.
  • Ethical considerations: Bias detection, privacy compliance, and fairness.

Module 3: Deployment of LLMs for Inferencing

Objectives:

  • Deploy LLMs for production inferencing with high performance and scalability.
  • Use NVIDIA TensorRT and Cisco Nexus Dashboard for optimized deployment.

Topics:

  • Deployment architectures: On-premises, cloud, and hybrid.
  • Optimizing inferencing with NVIDIA TensorRT: Precision calibration, layer fusion, and batching.
  • Traffic management and load balancing with Cisco Nexus Dashboard.
  • Exposing LLM APIs: RESTful and gRPC endpoints with security mechanisms.

Module 4: Optimizing LLM Models for Inferencing

Objectives:

  • Optimize LLM inferencing pipelines for low latency and high throughput.
  • Learn techniques like quantization, pruning, and model compression.

Topics:

  • Quantization: FP16, INT8, and mixed precision.
  • Pruning and knowledge distillation for lightweight models.
  • TensorRT optimization: Dynamic batching and asynchronous execution.
  • Benchmarking tools: NVIDIA Triton Inference Server, TensorRT Profiler.

Module 5: Scalable Pipeline Design for LLM Inferencing

Objectives:

  • Build robust, scalable, and fault-tolerant pipelines for inferencing.
  • Use batching, caching, and dynamic scaling for efficient pipelines.

Topics:

  • Pipeline components: Batching, caching, and queuing.
  • Load balancing with Cisco Nexus Dashboard for traffic optimization.
  • Fault tolerance: Automatic failover and disaster recovery plans.
  • Monitoring pipeline performance with NVIDIA DCGM and Cisco Nexus Dashboard.

Module 6: Monitoring, Logging, and Maintenance for LLM Systems

Objectives:

  • Monitor and maintain LLM deployments using NVIDIA and Cisco tools.

Topics:

  • Key metrics: Latency, throughput, GPU utilization, and memory usage.
  • Monitoring tools: NVIDIA DCGM and Cisco Nexus Dashboard Insights.
  • Maintenance workflows for hardware and software reliability.

Module 7: Security and Privacy Considerations in LLM Training and Inferencing

Objectives:

  • Secure LLM pipelines using Cisco Nexus Dashboard, Cisco XDR, and NVIDIA tools.

Topics:

  • NVIDIA runtime encryption and secure boot.
  • Cisco Robust Intelligence for adversarial defense and vulnerability detection.
  • Cisco XDR for unified threat detection and automated response.
  • Traffic segmentation and endpoint authentication.

Module 8: Migrating from Cloud-Based Training to On-Premises Inferencing

Objectives:

  • Transition LLM models from cloud training to on-premises Cisco infrastructure.

Topics:

  • Migration strategies for exporting and deploying models.
  • Data transfer optimization using Cisco Nexus Dashboard.
  • Integrating models with on-premises inferencing pipelines.

Module 9: On-Premises Data Center Design for LLM Inferencing Systems

Objectives:

  • Design an on-premises data center with Cisco and NVIDIA technologies.

Topics:

  • Cisco UCS and NVIDIA GPUs for high-performance compute.
  • Network design and automation with Cisco Nexus Dashboard.
  • Storage solutions for large-scale data management.

Module 10: On-Premises Data Center Implementation for LLM Inferencing Systems

Objectives:

  • Implement and configure an LLM inferencing data center using NVIDIA and Cisco technologies.

Topics:

  • Physical setup: NVIDIA GPUs on Cisco UCS and Nexus networking configuration.
  • Performance testing and validation of inferencing pipelines

    Prerequisites

    TOP

    Participants should possess basic knowledge of LLM models, server infrastructure, Cloud knowledge, networking concepts and virtualization fundamentals

      Who Should Attend

      TOP

      This course is tailored for professionals involved in designing and managing AI and data infrastructure, including:

      • Systems Architects: To understand the integration of LLM systems into broader IT environments.
      • Network Architects: To optimize network configurations for high-speed LLM training and inferencing.
      • Storage Architects: To manage the storage and retrieval of large-scale datasets used in LLM systems.
      • AI Infrastructure Architects: To build robust and scalable AI platforms optimized for LLM workloads.
      • Data Scientists: To prepare high-quality datasets and fine-tune LLMs for specific use cases.
      • Machine Learning Engineers: To deploy and optimize LLMs for real-world applications with low latency and high throughput.