Sergio Perez & Harshita Seth
Adding knowledge to open-source LLMs
#1about 4 minutes
Understanding the LLM training pipeline and knowledge gaps
LLMs are trained through pre-training and alignment, but require new knowledge to stay current, adapt to specific domains, and acquire new skills.
#2about 5 minutes
Adding domain knowledge with continued pre-training
Continued pre-training adapts a foundation model to a specific domain by training it further on specialized, unlabeled data using self-supervised learning.
#3about 6 minutes
Developing skills and reasoning with supervised fine-tuning
Supervised fine-tuning uses instruction-based datasets to teach models specific tasks, chat capabilities, and complex reasoning through techniques like chain of thought.
#4about 8 minutes
Aligning models with human preferences using reinforcement learning
Preference alignment refines model behavior using reinforcement learning, evolving from complex RLHF with reward models to simpler methods like DPO.
#5about 2 minutes
Using frameworks like NeMo RL to simplify model alignment
Frameworks like the open-source NeMo RL abstract away the complexity of implementing advanced alignment algorithms like reinforcement learning.
Related jobs
Jobs that call for the skills explored in this talk.
Matching moments
02:26 MIN
Understanding the core capabilities of large language models
Data Privacy in LLMs: Challenges and Best Practices
Unlock full access
Log in or set up an account to access this feature and more.
01:12 MIN
Introducing InstructLab for accessible LLM fine-tuning
Unlocking the Power of AI: Accessible Language Model Tuning for All
Unlock full access
Log in or set up an account to access this feature and more.
08:05 MIN
How large language models are trained
Inside the Mind of an LLM
Unlock full access
Log in or set up an account to access this feature and more.
05:18 MIN
Addressing the core challenges of large language models
Accelerating GenAI Development: Harnessing Astra DB Vector Store and Langflow for LLM-Powered Apps
Unlock full access
Log in or set up an account to access this feature and more.
02:55 MIN
Addressing the key challenges of large language models
Large Language Models ❤️ Knowledge Graphs
Unlock full access
Log in or set up an account to access this feature and more.
03:42 MIN
Using large language models as a learning tool
Google Gemini: Open Source and Deep Thinking Models - Sam Witteveen
Unlock full access
Log in or set up an account to access this feature and more.
02:21 MIN
The training process of large language models
Google Gemini: Open Source and Deep Thinking Models - Sam Witteveen
Unlock full access
Log in or set up an account to access this feature and more.
03:17 MIN
Using RAG to enrich LLMs with proprietary data
RAG like a hero with Docling
Unlock full access
Log in or set up an account to access this feature and more.
Featured Partners
Related Videos
Inside the Mind of an LLM
Emanuele Fabbiani
LLMOps-driven fine-tuning, evaluation, and inference with NVIDIA NIM & NeMo Microservices
Anshul Jindal
Unlocking the Power of AI: Accessible Language Model Tuning for All
Cedric Clyburn & Legare Kerrison
Self-Hosted LLMs: From Zero to Inference
Roberto Carratalá & Cedric Clyburn
Exploring LLMs across clouds
Tomislav Tipurić
Give Your LLMs a Left Brain
Stephen Chin
Three years of putting LLMs into Software - Lessons learned
Simon A.T. Jiménez
Unveiling the Magic: Scaling Large Language Models to Serve Millions
Patrick Koss
Related Articles
View all articles.png?w=240&auto=compress,format)
.gif?w=240&auto=compress,format)
.png?w=240&auto=compress,format)

From learning to earning
Jobs that call for the skills explored in this talk.




Warmwind
€240K
Senior
PyTorch
Tensorflow
Machine Learning



Warmwind
€120K
PyTorch
Tensorflow
Machine Learning
