TODAY'S ISSUE
TODAYβS DAILY DOSE OF DATA SCIENCE
βA crash course on RAG systems - Part 2
Last week, we started a crash course on building RAG systems.
Part 2 is now available, where we are building on the foundations laid in Part 1.
Read here: βA Crash Course on Building RAG Systems β Part 2 (With Implementation)β.
Why care?
Over the last few weeks, we have spent plenty of time understanding the key components of real-world NLP systems (like the deep dives on bi-encoders and cross-encoders for context pair similarity scoring).
RAG is another key NLP system that got massive attention due to one of the key challenges it solved around LLMs.
More specifically, if you know how to build a reliable RAG system, you can bypass the challenge and cost of fine-tuning LLMs.
Thatβs a considerable cost saving for enterprises.
And at the end of the day, all businesses care about impact. Thatβs it!
- Can you reduce costs?
- Drive revenue?
- Can you scale ML models?
- Predict trends before they happen?
Thus, the objective of this crash course is to help you implement reliable RAG systems, understand the underlying challenges, and develop expertise in building RAG apps on LLMs, which every industry cares about now.
- Read the βfirst part here ββ
- Read the βsecond part here ββ
Of course, if you have never worked with LLMs, thatβs okay. We cover everything in a practical and beginner-friendly way.
IN CASE YOU MISSED IT
Extend the context length of LLMs
- GPT-3.5-turbo had a context window of 4,096 tokens.
- Later, GPT-4 took that to 8,192-32k tokens.
- Claude 2 reached 100,000 tokens.
- Llama 3.1 β 128,000 tokens.
- Gemini β 1M+ tokens.
We have been making great progress in extending the context window of LLMs.
But how?
We covered techniques that help us unlock larger context windows earlier this week.
βRead the techniques to extend the context length of LLMs here ββ
IN CASE YOU MISSED IT
Building a 100% local multi-agent Internet research assistant with OpenAI Swarm & Llama 3.2
Recently, OpenAI released Swarm.
Itβs an open-source framework designed to manage and coordinate multiple AI agents in a highly customizable way.
AI Agents are autonomous systems that can reason, think, plan, figure out the relevant sources and extract information from them when needed, take actions, and even correct themselves if something goes wrong.
We published a practical and hands-on demo of this in the newsletter. We built an internet research assistant app that:
- Accepts a user query.
- Searches the web about it.
- And turns it into a well-crafted article.
The demo is shown below:
THAT'S A WRAP
No-Fluff Industry ML resources to
Succeed in DS/ML roles
At the end of the day, all businesses care about impact. Thatβs it!
- Can you reduce costs?
- Drive revenue?
- Can you scale ML models?
- Predict trends before they happen?
We have discussed several other topics (with implementations) in the past that align with such topics.
Here are some of them:
- Learn sophisticated graph architectures and how to train them on graph data in this crash course.
- So many real-world NLP systems rely on pairwise context scoring. Learn scalable approaches here.
- Run large models on small devices using Quantization techniques.
- Learn how to generate prediction intervals or sets with strong statistical guarantees for increasing trust using Conformal Predictions.
- Learn how to identify causal relationships and answer business questions using causal inference in this crash course.
- Learn how to scale and implement ML model training in this practical guide.
- Learn 5 techniques with implementation to reliably test ML models in production.
- Learn how to build and implement privacy-first ML systems using Federated Learning.
- Learn 6 techniques with implementation to compress ML models.
All these resources will help you cultivate key skills that businesses and companies care about the most.
SPONSOR US
Advertise to 450k+ data professionals
Our newsletter puts your products and services directly in front of an audience that matters β thousands of leaders, senior data scientists, machine learning engineers, data analysts, etc., around the world.