Featured resource
2025 Tech Upskilling Playbook
Tech Upskilling Playbook

Build future-ready tech teams and hit key business milestones with seven proven plays from industry leaders.

Check it out

GenAI for Developers

Course Summary

This course equips participants with the skills to design, build, and deploy modern Generative AI applications using Large Language Models (LLMs). Moving beyond basic prompt engineering and simple chatbot examples, participants will explore real-world engineering patterns such as Retrieval-Augmented Generation (RAG), tool integration, and orchestration.
Through hands-on labs and activities, participants will be equipped to build intelligent systems that are reliable, scalable, and secure.

Prerequisites

To get the most out of this session, participants should have:

  • Experience with Python
  • Basic understanding of REST APIs
  • Familiarity with JSON and data structures
  • Basic knowledge of machine learing
     
Purpose
Learn the skills needed to design, build and deploy Generative AI applications using Large Language Models (LLMs)
Audience
IT professionals interested in creating Generative AI applications using LLMs
Role
Software Developers | Technical Managers | Data Professionals | DevOps Engineers
Skill level
Intermediate
Style
Lecture | Hands-on Activities | Labs
Duration
2 days
Related technologies
AI/ML | Python | REST APIs

 

Learning objectives
  • Explain how large language models (LLMs) work
  • Design effective prompts and structured outputs
  • Implement modern AI application patterns, including Retrieval-Augmented Generation (RAG) and tool-based orchestration
  • Evaluate, debug, and improve LLM outputs, including handling hallucinations and testing system reliability
  • Design production-ready and responsible AI systems

What you'll learn:

In this GenAI for Developers course, you'll learn:

Introduction to Generative AI

  • What is generative AI?
  • Overview of LLMs
  • Capabilities and limitations
  • Real-world use cases

Prompt Engineering Fundamentals

  • Prompt structure and best practices
  • Zero-shot vs few-shot prompting
  • Controlling tone, format, and constraints
  • Structured outputs

Working with LLM APIs

  • API interaction patterns
  • Token usage and context windows
  • Streaming responses
  • Error handling strategies

Modern LLM Application Patterns

  • Core Patterns
    • Retrieval-Augmented Generation (RAG)
    • Embeddings and vector databases
    • Chunking strategies (basic vs semantic)
  • Orchestration Concepts
    • Chains vs routing workflows
    • Intent-based decision systems
    • Tool/function calling
  • Improving RAG QualityConversation memory and session handling
  • Managing context and avoiding drift
  • Multi-step workflows
  • Streaming UX patterns
    • Hybrid search (keyword + vector)
    • Re-ranking results (conceptual)
    • Why naive RAG fails in production
  • When NOT to use LLMs

Building a Production-Ready AI Assistant

  • Conversation memory and session handling
  • Managing context and avoiding drift
  • Multi-step workflows
  • Streaming UX patterns

Evaluation and Debugging

  • Why LLM outputs are hard to test
  • Prompt evaluation strategies
  • Golden datasets and test cases
  • Introduction to “LLM-as-a-judge”
  • Detecting hallucinations and inconsistencies

Cost, Performance, and Architecture

  • Model selection tradeoffs (small vs large models)
  • Cost optimization strategies
  • Latency vs quality tradeoffs
  • Caching and batching

Responsible AI and Safety

  • Bias and fairness
  • Data privacy and protection
  • Guardrails and moderation

Dive in and learn more

When transforming your workforce, it’s important to have expert advice and tailored solutions. We can help. Tell us your unique needs and we'll explore ways to address them.

Let's chat

By clicking submit, you agree to our Privacy Policy and Terms of Use, and consent to receive marketing emails from Pluralsight.