For a hands-on learning experience to develop LLM applications, join our LLM Bootcamp today.
Early Bird Discount Ending Soon!

4.98/5 - 6,000+ student reviews

Large Language Models Bootcamp

In just one week, we will teach you how to build agentic AI applications. Learn the entire LLM application stack.

Technologies and Tools

#1

Ranked Bootcamp

10,000+

Graduates Worldwide

7000+

hiring companies

For who

Who is this bootcamp for?

The Large Language Models Bootcamp is tailored for a diverse audience, including: 

Data professionals

Looking to enhance their skills with generative AI tools. Learn how to integrate LLMs into your data workflows and boost efficiency with real-world AI applications.

Product leaders

At enterprises or startups aiming to leverage LLMs for improving products, processes, or services .Discover how LLMs can drive innovation and streamline decision-making across your product lifecycle.

Beginners

Seeking a head start in understanding and working with LLMs. Build a solid foundation in generative AI with guided, beginner-friendly lessons and hands-on exercises.

Professionals

Wanting to supercharge their expertise in building and deploying custom LLM-powered applications. Gain advanced skills in fine-tuning, prompting, and integrating LLMs into scalable systems and tools.

Instructors and guest speakers

Learn from though leaders at the forefront of building agentic AI applications

Luis - Agentic AI Panelist

Luis Serrano

Founder, Serrano Academy
Raja Iqbal-Data Science Dojo

Raja Iqbal

Founder, Ejento AI
Sebastian Witalec | Weaviate

Sebastian Witalec

Director of Developer Relations, Weaviate
John Gilhuly | Arize AI | Data Science Dojo

John Gilhuly

Head of Developer Relations, Arize AI
Kartik - Agentic AI Conference

Kartik Talamadupula

Head of AI, Wand AI
Jerry Agentic AI Panelist

Jerry Liu

CEO/Co-founder, LlamaIndex
Zain - Agentic AI Panelist

Zain Hasan

Senior DevRel Engineer, Together AI
1703300423927 (1)

Sage Elliot

AI Engineer, Union AI
Instructor Sophie Daly from Stripe , guiding participants through the LLM Bootcamp.

Sophie Daly

Staff Data Scientist, Stripe
Rehan Jalil Securiti AI

Rehan Jalil

Co-Founder | CEO, Securiti AI
Adam Cowley | Developer Advocate Neo4j

Adam Cowley

Developer Advocate, Neo4j
hamza farooq

Hamza Farooq

Founder, Travesaal AI

Earn A Verified Certificate

Earn a verified certificate from The University of New Mexico Continuing Education:

  • 5 Continuing Education Credit (CEU)
  • Acceptable by employers for reimbursements
  • Valid for professional licensing renewal
  • Verifiable by The University of New Mexico Registrar’s office
  • Add to LinkedIn and share with your network

curriculum

Explore the bootcamp curriculum

Overview of the topics and practical exercises.

Understanding the LLM Ecosystem

In this module we will understand the common use cases of large language models and fundamental building blocks of such applications. Learners will be introduced to the following topics at a very high level without going into the technical details: 

  • Large language models and foundation models  
  • Prompts and prompt engineering  
  • Context window and token limits  
  • Embeddings and vector databases  
  • Build custom LLM applications by:  
  • Training a new model from scratch  
  • Fine-tuning foundation LLMs  
  • In-context learning  
  • Canonical architecture for an end-to-end LLM application 

Adoption Challenges and Risks

In this module, we will explore the primary challenges and risks associated with adopting generative AI technologies. Learners will be introduced to the following topics at a very high level without going into the technical details: 

  • Misaligned behavior of AI systems 
  • Handling complex datasets 
  • Limitations due to context length 
  • Managing cost and latency 
  • Addressing prompt brittleness 
  • Ensuring security in AI applications 
  • Achieving reproducibility 
  • Evaluating AI performance and outcomes 

The Building Blocks of LLM

In this module, we will be reviewing how embeddings have evolved from the simplest one-hot encoding approach to more recent semantic embeddings approaches. The module will go over the following topics: 

Review of classical techniques  

  • Review of binary/one-hot, count-based and TF-IDF techniques for vectorization 
  • Capturing local context with n-grams and challenges

Semantic Encoding Techniques  

  • Overview of Word2Vec and dense word embeddings 
  • Application of Word2Vec in text analytics and NLP tasks

Text Embeddings  

  • Word and sentence embeddings 
  • Text similarity measures 
  • Dot product, Cosine similarity, Inner product

Hands-on Exercise  

  • Creating a TF-IDF embeddings on a document corpus 
  • Calculating similarity between sentences using cosine similarity and dot product

A comprehensive introduction to attention mechanism and transformer architecture

Dive into the world of large language models, discovering the potent mix of text embeddings, attention mechanisms, and the game-changing transformer model architecture. 

  • Attention mechanism and transformer models 
  • Encoder decoder 
  • Tokenization, embedding, positional encoding and transformers block 
  • Attention mechanism 
  • Self-Attention 
  • Multi-head Attention
  • Transformer models

Hands-on Exercise  

  • Understanding attention mechanisms: Self-attention for contextual word analysis

A comprehensive introduction to vector databases

Learn about efficient vector storage and retrieval with vector database, indexing techniques, retrieval methods, and hands-on exercises.  

  • Overview:Rationale for vector databases, importance of vector databases in LLMs, popular vector databases 
  •  Understand how semantic search overcomes the fundamental limitation in lexical search: The lack of semantic understanding. Learn how to use embeddings and similarity to build a semantic search model.  
  • Core Database Operations and Vector Fundamentals:CRUD operations (Create, Read, Update, Delete), vector embeddings and similarity search, query processing and transaction management, distance metrics and data ingestion  
  • Indexing, Search and Performance: HNSW indexing and approximate nearest neighbor, vector search, keyword search, hybrid search.  performance tuning and optimization, filtering and ranking mechanisms 
  • Production and Advanced Features: Horizontal scaling, sharding, and distributed architecture, security, backup, recovery, and data management.  RAG implementation and multi-modal capabilities

Hands-on Exercise  

  • Learn how to performvector similarity searches with embedding models, combine vector and keyword (BM25F) results using hybrid search, implement multi-tenancy for secure data management, and compress vectors using product quantization to reduce memory footprint. 

Unleash your creativity and efficiency with prompt engineering.

Unleash your creativity and efficiency with prompt engineering. Seamlessly prompt models, control outputs, and generate captivating content across various domains and tasks. 

  • Introduction to Prompt Engineering

 ○ Designing inputs to guide LLM behavior  

○ Benefits: Control, efficiency, customization 

  • Anatomy of a Prompt 

○ Elements: Instructions, Context, Input, Output  

○ Formats: Declarative, Interrogative, Structured 

  • Prompt Engineering Techniques

 ○ Few-Shot: Examples guide behavior  

○ Chain-of-Thought: Step-by-step reasoning  

○ Tree-of-Thought: Branching problem-solving  

○ ReAct: Reasoning + external actions ○ Advanced: Maieutic, DSP, Multi-modal, Function Calling 

  • Applications:Translation, Dialogue Systems, Information Retrieval
  • Parameters and Control 

○ Temperature: Randomness control (0-1) 

 ○ Top-k/Top-p: Token selection limits 

  • Risks and Safety:Injection, Leaking, Jailbreaking vulnerabilities

A practical introduction to fine tuning

In-depth discussion on fine-tuning of large language models through theory discussions, exploring rationale, limitations, and Parameter Efficient Fine Tuning. 

  • Introduction to Fine-Tuning 

○ Continuing training on pre-trained model with custom dataset  

○ Benefits: Domain adaptation, tone/style adjustment, task-specific transformation 

  • Alternatives to Fine-Tuning :Prompt Engineering,RAG, Transfer Learning, Knowledge Distillation 
  • Fine-Tuning Methods
    ○ SFT: Labeled input-output pairs  

○ DPO/RLHF: Human preference alignment  

○ PEFT vs Full: <10% vs all parameters 

  • Optimization Strategies:Low-Rank Adaptation (LoRA),Quantization, Quantized Low-Rank Adaptation (QLoRA) 
  • Data and Challenges 

○ Data Quality: Clean, domain-specific, balanced  

○ Challenges: Catastrophic forgetting, overfitting, drift 

  • Hands-on Exercise: 

○ In-Class: Fine-tuning open-source LLMs  

○ Homework: OpenAI/Llama on Azure AI 

LangChain for LLM Application Development

Build LLM Apps using LangChain framework for orchestrating language models, prompts, memory, chains, and retrieval systems. 

Introduction to LangChainOrchestration framework for LLM applications, simplifies AI model integration, and provides modular components for complex workflows 

Model I/O (Interface with any LLM)Prompts (templates, example selectors), Language models (LLM, Chat, Embedding), Output parsers for structured responses 

Retrieval (Connecting external data)Document loaders (public/private/structured/unstructured data), transformers (chunking, metadata), embedding/vector stores, optimized retrieval techniques 

Chains (Complex LLM workflows)Foundational types (LLM, Router, Sequential, Transformation), Document chains (stuff, refine, map-reduce) for large document summarization 

Memory (Context retention)Short-term (thread-scoped) and long-term (cross-thread) memory for past interactions, buffer management, and conversation continuity.

LangGraph & Advanced Features Cyclical workflows, agency vs reliability trade-off management, stateful memory, and cognitive architectures for agentic applications 

Hands-on Exercise  

  • Interface with any LLM using Model I/O
  • Building RAG application with Retrieval 
  • Creating complex LLM workflows with Chains 
  • Adding Memory to LLM-based application 
  • Harnessing dynamic decision-making using Agents 
  • Supplementary Exercises: Many coding exercises onLangChainModel I/O, Memory, Chains, Memory and Agents 
  • LLMOps encompasses the practices, techniques and tools used for the operational management of large language models in production environments. LLMs offer tremendous business value, humans are involved in all stages of the lifecycle of an LLM from acquisition of data to interpretation of insights. In this module we will learn about the following: 

    Principles of Responsible AI • Fairness and Eliminating Bias • Reliability and Safety • Privacy and Data Protection 

    Review techniques for assessing LLM applications, including: • Model fine-tuning • Model inference and serving • Model monitoring with human feedback 

    Data-centric LLM Ops • Guardrails: Define rules to govern prompts and responses for LLM applications • Evaluation: Assess LLM performance using known prompts to identify issues • Observability: Collect telemetry data about the LLM’s internal state for monitoring and issue detection 

    Hands-on Exercise • Using Langkit Evaluate LLM performance on specific prompts  

Challenges in building RAG applications

In this module, we’ll explore the challenges in developing RAG-based enterprise-level Large Language Model (LLM) applications. We will discuss the following: 

Basic RAG pipeline. Limitations of naïve approach 

Indexing • Chunking size optimization • Embedding Models 

Querying – Challenges • Large Document Slices • Query Ambiguity 

Query – Optimizations • Multi-Query Retrieval • Multi-Step Retrieval • Step-Back Prompting • Query Transformations 

Retrieval – Challenges • Inefficient Retrieval of Large Documents • Lack of Conversation Context • Complex Retrieval from Multiple Sources 

Retrieval – Optimizations • Hybrid Search and Meta-data integration • Sentence window retrieval • Parent-child chunk retrieval • Hierarchical Index Retrieval • Hypothetical Document embeddings (HyDE) 

Generation – Challenges • Information Overload • Insufficient Context Window • Chaotic Contexts • Hallucination • Inaccurate Responses 

Generation – Optimizations • Information Compression • Thread of Thought (ThoT) • Generator Fine-tuning • Adapter methods • Chain of Note (CoN) • Expert Prompting • Access control and governance  

Evaluating Large Language Models (LLMs)

Explore LLM evaluation, key metrics like BLEU and ROUGE, with hands-on exercises. 

Introduction to LLM Evaluation 

  • Evaluation importance, common LLM mistakes, and benchmark datasets/metrics overview 
  • Evaluation Metrics 
  • Automatic metrics (BLEU, ROUGE, BERTScore) with strengths/weaknesses analysis 
  • Human evaluation techniques (Likert scale) 

RAGAS 

  • Introduction and workflow overview 
  • Key evaluation metrics: Faithfulness, Context precision, Answer relevancy, Context recall 

Detailed Workflow Stages 

  • Practical Applications 
  • Summarization, Open-domain QA, and Fact-checking applications 

Hands-on Exercise 

  • Evaluating LLMs summarization using metrics like Rouge and Bertscore 
  • Evaluation GPTEval 
  • Evaluation of end-to-end RAG pipeline with RAGAS.

Build A Multi-agent LLM Application

Apply bootcamp concepts to build custom LLM applications with real-world deployment and scaling capabilities. 

Project Options 

Basic Chatbot (general queries), Chatbot Agent (data integration), Chat with Your Data (document upload/interaction), Web Search Assistant, Question Answering systems 

Provided Resources 

Comprehensive industry datasets, step-by-step implementation guides, ready-to-use code templates in sandbox environments, cloud resources with OpenAI key access 

Deployment & Outcome 

Streamlit cloud deployment with CI/CD pipeline, fully operational application, functionality demonstration, and skills for real-world scaling 

Showcase & Impact 

Leverage AI for industry transformation, competitive advantage, social media presentation, and professional development  

Attend the LLM Bootcamp for free

We Accept Tuition Benefits

All of our programs are backed by a certificate from The University of New Mexico, Continuing Education. This means that you may be eligible to attend the bootcamp for FREE.

Not sure? Fill out the form so we can help.

Get a certificate from The University of New Mexico Continuing Education with 5 CEUs

UNM's continuing education | Data Science Dojo

Upcoming sessions

Reserve your spot

Learn to build Large Language Model applications from leading experts in industry. 

Large Language Models Bootcamp

Next session -> September Cohort

Use LLM1500 for USD 1500 discount.

Confused ? Schedule a call with an Advisor 

Pace

Dates

Time

Price

Enroll Now

Seattle

Sept 22 -> 26

Monday to Friday,
9 AM to 5 PM PT

$3499

Online

Sept 22 -> 26

Monday to Friday,
9 AM to 5 PM PT

$3499

A Word From Our Alumni

Japneet Singh | LLM Bootcamp | Data Science Dojo
Play Video
Usman Dankoly | LLM Bootcamp | Data Science Dojo
Play Video
Alejandro W. | LLM Bootcamp | Data Science Dojo
Play Video
Stan Zaporozhets | LLM Bootcamp | Data Science Dojo
Play Video
Omar Smith | LLM Bootcamp | Data Science Dojo
Play Video
Roger Campbell | LLM Bootcamp | Data Science Dojo
Play Video
Tariq Hook | LLM Bootcamp | Data Science Dojo
Play Video
Ali Abuharb | LLM Bootcamp | Data Science Dojo
Play Video
Dave Horton | LLM Bootcamp | Data Science Dojo
Play Video
Francisco Morales | LLM Boorcamp | Data Science Dojo
Play Video
Shakeeb Syed | LLM Bootcamp | Data Science Dojo
Play Video
Yashwant Reddy | LLM Bootcamp | Data Science Dojo
Play Video
Sahar Nesaei | LLM Bootcamp | Data Science Dojo
Play Video
Florian Klonek | LLM Bootcamp | Data Science Dojo
Play Video
Large Language Models Bootcamp - Maryam Bagher
Play Video
Large Language Models Bootcamp - Kshitij Singh
Play Video
Large Language Models Bootcamp - Jared Miller
Play Video
Victor Green | Large Language Models Bootcamp | Data Science Dojo
Play Video
Ken Btler | Large Language Models Bootcamp | Data Science Dojo
Play Video
Abrar Bhuiyan | Large Language Models Bootcamp | Data Science Dojo
Play Video
Erika Davis | Large Language Models Bootcamp | Data Science Dojo
Play Video
Aishwariya Raman | Large Language Models Bootcamp | Data Science Dojo
Play Video
Luis Armando | Large Language Models Bootcamp | Data Science Dojo
Play Video
Amity Fox | Large Language Models Bootcamp | Data Science Dojo
Play Video
David Martins | Large Language Models Bootcamp | Data Science Dojo
Play Video
Ed Wiley | Large Language Models Bootcamp | Data Science Dojo
Play Video

FAQ

Your questions answered

I am from a non-technical background. Is the LLM Bootcamp also for me?

Our LLM Bootcamp has been attended by many individuals from non-technical backgrounds, including those in business consulting and strategy roles. The program is designed to make Generative AI concepts accessible, regardless of your technical expertise.
You’ll gain practical insights into how LLMs are applied across industries, empowering you to advise clients better, lead AI initiatives, or collaborate effectively with technical teams.

Are there any prerequisites?

You need a very basic level of Python programming for our LLM Bootcamp.

I don’t have a background in Python. Can I still join the LLM Bootcamp?

Yes! We offer a short introductory course to help you get comfortable with Python. Plus, all code is provided in Jupyter Notebooks, so you’ll focus on understanding rather than writing code from scratch.

Where will the in-person session be held?

The address for the LLM Bootcamp venue is given below:

Seattle Venue Address: Data Science Dojo, 2331 130th Ave NE, Bellevue, WA 98005, United States. [View on Map]

What is the duration of the Bootcamp?

Our LLM Bootcamp is an immersive five-day, 40-hour learning experience, available both in-person (Seattle) and online.

Are the online sessions live?

Yes, these sessions are live and are designed to be highly interactive.

Is the online instruction held at the same time as the in-person session, with the same instructors?

Yes, the online session will be held at the same time with the same instructors as the in-person session.

What does the LLM Bootcamp include?

By joining the LLM Bootcamp, you will receive:

  • 40 hours of theory and hands-on learning
  • Live sessions with industry experts
  • 1-year access to dedicated learner and coding sandboxes
  • LLM tokens, GPU clusters, and other required subscriptions
  • A verified certificate from the University of New Mexico upon completion
Is the program accredited?

Yes. You will receive a certificate from The University of New Mexico with 5 CEUs.

Will I get a Certificate after the completion of the Bootcamp?

Yes, participants who complete the bootcamp will receive a certificate of completion in association with the University of Mexico. This certificate can be a valuable addition to your professional portfolio and demonstrate your expertise in building large language model applications.

What if I’m unable to attend a live session?

Each live session is recorded and made available for review to both online and in-person participants a few days after the boot camp concludes, allowing them to view it at their convenience.

Is the price for the online and in-person sessions different?

No, the price for the Large Language Models Bootcamp will remain the same, regardless of whether you attend in person or online.

What is the refund policy?

If, for any reason, you decide to cancel, we will gladly refund your registration fee in full if you notify us at least five business days before the start of the training. We would also be happy to transfer your registration to another cohort. Refunds cannot be processed if you have moved to a different cohort after registration.

What is the transfer policy for the LLM Bootcamp?

Transfers are allowed once with no penalty. Transfers requested more than once will incur a $200 processing fee.

I am from a non-technical background. Is the LLM Bootcamp also for me?

Our LLM Bootcamp has been attended by many individuals from non-technical backgrounds, including those in business consulting and strategy roles. The program is designed to make Generative AI concepts accessible, regardless of your technical expertise.
You’ll gain practical insights into how LLMs are applied across industries, empowering you to advise clients better, lead AI initiatives, or collaborate effectively with technical teams.

Are there any prerequisites?

You need a very basic level of Python programming for our LLM Bootcamp.

I don’t have a background in Python. Can I still join the LLM Bootcamp?

Yes! We offer a short introductory course to help you get comfortable with Python. Plus, all code is provided in Jupyter Notebooks, so you’ll focus on understanding rather than writing code from scratch.

Where will the in-person session be held?

The address for the LLM Bootcamp venue is given below:

Seattle Venue Address: Data Science Dojo, 2331 130th Ave NE, Bellevue, WA 98005, United States. [View on Map]

What is the duration of the Bootcamp?

Our LLM Bootcamp is an immersive five-day, 40-hour learning experience, available both in-person (Seattle) and online.

What does the LLM Bootcamp include?

By joining the LLM Bootcamp, you will receive:

  • 40 hours of theory and hands-on learning
  • Live sessions with industry experts
  • 1-year access to dedicated learner and coding sandboxes
  • LLM tokens, GPU clusters, and other required subscriptions
  • A verified certificate from the University of New Mexico upon completion
Are the online sessions live?

Yes, these sessions are live and are designed to be highly interactive.

Is the online instruction held at the same time as the in-person session, with the same instructors?

Yes, the online session will be held at the same time with the same instructors as the in-person session.

Is the program accredited?

Yes. You will receive a certificate from The University of New Mexico with 5 CEUs.

Will I get a Certificate after the completion of the Bootcamp?

Yes, participants who complete the bootcamp will receive a certificate of completion in association with the University of Mexico. This certificate can be a valuable addition to your professional portfolio and demonstrate your expertise in building large language model applications.

What if I’m unable to attend a live session?

Each live session is recorded and made available for review to both online and in-person participants a few days after the boot camp concludes, allowing them to view it at their convenience.

Is the price for the online and in-person sessions different?

No, the price for the Large Language Models Bootcamp will remain the same, regardless of whether you attend in person or online.

What is the refund policy?

If, for any reason, you decide to cancel, we will gladly refund your registration fee in full if you notify us at least five business days before the start of the training. We would also be happy to transfer your registration to another cohort. Refunds cannot be processed if you have moved to a different cohort after registration.

What is the transfer policy for the LLM Bootcamp?

Transfers are allowed once with no penalty. Transfers requested more than once will incur a $200 processing fee.