Sergio Perez & Harshita Seth
Adding knowledge to open-source LLMs
#1about 4 minutes
Understanding the LLM training pipeline and knowledge gaps
LLMs are trained through pre-training and alignment, but require new knowledge to stay current, adapt to specific domains, and acquire new skills.
#2about 5 minutes
Adding domain knowledge with continued pre-training
Continued pre-training adapts a foundation model to a specific domain by training it further on specialized, unlabeled data using self-supervised learning.
#3about 6 minutes
Developing skills and reasoning with supervised fine-tuning
Supervised fine-tuning uses instruction-based datasets to teach models specific tasks, chat capabilities, and complex reasoning through techniques like chain of thought.
#4about 8 minutes
Aligning models with human preferences using reinforcement learning
Preference alignment refines model behavior using reinforcement learning, evolving from complex RLHF with reward models to simpler methods like DPO.
#5about 2 minutes
Using frameworks like NeMo RL to simplify model alignment
Frameworks like the open-source NeMo RL abstract away the complexity of implementing advanced alignment algorithms like reinforcement learning.
Related jobs
Jobs that call for the skills explored in this talk.
Matching moments
01:31 MIN
Understanding the core capabilities of large language models
Data Privacy in LLMs: Challenges and Best Practices
00:02 MIN
Introducing InstructLab for accessible LLM fine-tuning
Unlocking the Power of AI: Accessible Language Model Tuning for All
07:35 MIN
How large language models are trained
Inside the Mind of an LLM
00:27 MIN
Addressing the core challenges of large language models
Accelerating GenAI Development: Harnessing Astra DB Vector Store and Langflow for LLM-Powered Apps
01:47 MIN
Addressing the key challenges of large language models
Large Language Models ❤️ Knowledge Graphs
28:51 MIN
Using large language models as a learning tool
Google Gemini: Open Source and Deep Thinking Models - Sam Witteveen
06:29 MIN
The training process of large language models
Google Gemini: Open Source and Deep Thinking Models - Sam Witteveen
00:40 MIN
Using RAG to enrich LLMs with proprietary data
RAG like a hero with Docling
Featured Partners
Related Videos
Inside the Mind of an LLM
Emanuele Fabbiani
LLMOps-driven fine-tuning, evaluation, and inference with NVIDIA NIM & NeMo Microservices
Anshul Jindal
Unlocking the Power of AI: Accessible Language Model Tuning for All
Cedric Clyburn & Legare Kerrison
Self-Hosted LLMs: From Zero to Inference
Roberto Carratalá & Cedric Clyburn
Exploring LLMs across clouds
Tomislav Tipurić
Give Your LLMs a Left Brain
Stephen Chin
Creating Industry ready solutions with LLM Models
Vijay Krishan Gupta & Gauravdeep Singh Lotey
Three years of putting LLMs into Software - Lessons learned
Simon A.T. Jiménez
Related Articles
View all articles.png?w=240&auto=compress,format)
.gif?w=240&auto=compress,format)
.png?w=240&auto=compress,format)

From learning to earning
Jobs that call for the skills explored in this talk.


AI/ML Team Lead - Generative AI (LLMs, AWS)
Provectus
Remote
€96K
Senior
PyTorch
Tensorflow
Computer Vision
+2


Senior Machine Learning Engineer (LLM & GPU Architecture)
European Tech Recruit
Intermediate
Docker
PyTorch
Kubernetes
Computer Vision
Machine Learning


AI Engineer Knowledge Graphs & Large Language Models
digatus it group
Remote
€62-79K
Intermediate
Neo4j
TypeScript
Microservices
+1


