TODAY'S ISSUE
TOGETHER WITH FIRECRAWL
π₯βTurn ANY website into LLM-ready data [Open-source]
AI systems love neatly formatted dataβMarkdowns, Structured data, HTML etc.
And now it is easier than ever to produce LLM-digestible data!
βFirecrawlβ is an open-source framework that takes a URL, crawls it, and converts it into a clean markdown or structured format. Star the repo below:
Why Firecrawl?
- LLM-ready formats β Markdown, HTML, Structured data, metadata.
- Handles the hard stuff β proxies, anti-bots, dynamic content.
- Customizable β exclude tags, custom headers, max depth.
- Reliable β gets the data you need, no matter what.
- Batching β scrape thousands of URLs at once
- Media parsing β PDFs, DOCX, images
- Actions β click, scroll, input, wait.
If you prefer βFireCrawl's managed serviceβ, you can use the code "DDODS" for a 10% discount code βhere ββ
Thanks to βFirecrawlβ for partnering with us today.
TODAY's DAILY DOSE OF DATA SCIENCE
What is temperature in LLMs?ββ
A low temperate value produces identical responses from the LLM (shown below):
But a high temperate value produces gibberish.
What exactly is temperature
in LLMs?
Letβs understand this today!
Traditional classification models use softmax to generate the final prediction from logits over all classes. In LLMs, the output layer spans the entire vocabulary.
The difference is that a traditional classification model predicts the class with the highest softmax score, which makes it deterministic.
But LLMs sample the prediction from these softmax probabilities:
Thus, even though βToken 1
β has the highest probability of being selected (0.86
), it may not be chosen as the next token since we are sampling.
Temperature introduces the following tweak in the softmax function, which, in turn, influences the sampling process:
1) If the temperature is low, the probabilities look more like a max value instead of a βsoft-maxβ value.
- This means the sampling process will almost certainly choose the token with the highest probability.
- This makes the generation process look greedy and (almost) deterministic.
2) If the temperature is high, the probabilities start to look like a uniform distribution:
- This means the sampling process may select any token.
- This makes the generation process random and heavily stochastic.
A quick note: In practice, the model can generate different outputs even if temperature=0
. This is because there are still several other sources of randomness, such as race conditions in multithreaded code.
Here are some best practices for using temperature:
- Set a low temperature value to generate predictable responses.
- Set a high temperature value to generate more random and creative responses.
- An extremely high temperature value rarely has any real utility, as we saw at the top.
And this explains the objective behind temperature in LLMs.
That said, any AI system will only be as good as the data going in.
βFireCrawlβ helps you ensure that your AI systems always receive neatly formatted dataβMarkdowns, Structured data, HTML, etc.
If you prefer ββFireCrawl's managed serviceββ, you can use the code βDDODSβ for a 10% discount code ββhere ββ
π Over to you: How do you determine an ideal value of temperature?
IN CASE YOU MISSED IT
βRAG crash course
RAG is a key NLP system that got massive attention due to one of the key challenges it solved around LLMs.
More specifically, if you know how to build a reliable RAG system, you can bypass the challenge and cost of fine-tuning LLMs.
Thatβs a considerable cost saving for enterprises.
And at the end of the day, all businesses care about impact. Thatβs it!
- Can you reduce costs?
- Drive revenue?
- Can you scale ML models?
- Predict trends before they happen?
Thus, the objective of this crash course is to help you implement reliable RAG systems, understand the underlying challenges, and develop expertise in building RAG apps on LLMs, which every industry cares about now.
- βIn Part 1β, we explored the foundational components of RAG systems, the typical RAG workflow, and the tool stack, and also learned the implementation.
- βIn Part 2β, we understood how to evaluate RAG systems (with implementation).
- βIn Part 3β, we learned techniques to optimize RAG systems and handle millions/billions of vectors (with implementation).
- βIn Part 4β, we understood multimodality and covered techniques to build RAG systems on complex docsβones that have images, tables, and texts (with implementation):
- βIn Part 5β, we understood the fundamental building blocks of multimodal RAG systems that will help us improve what we built in Part 4.
- In Part 6, we utilized learnings from Part 5 to build a more extensive and capable multimodal RAG system.
Of course, if you have never worked with LLMs, thatβs okay. We cover everything in a practical and beginner-friendly way.
THAT'S A WRAP
No-Fluff Industry ML resources to
Succeed in DS/ML roles
At the end of the day, all businesses care about impact. Thatβs it!
- Can you reduce costs?
- Drive revenue?
- Can you scale ML models?
- Predict trends before they happen?
We have discussed several other topics (with implementations) in the past that align with such topics.
Here are some of them:
- Learn sophisticated graph architectures and how to train them on graph data in this crash course.
- So many real-world NLP systems rely on pairwise context scoring. Learn scalable approaches here.
- Run large models on small devices using Quantization techniques.
- Learn how to generate prediction intervals or sets with strong statistical guarantees for increasing trust using Conformal Predictions.
- Learn how to identify causal relationships and answer business questions using causal inference in this crash course.
- Learn how to scale and implement ML model training in this practical guide.
- Learn 5 techniques with implementation to reliably test ML models in production.
- Learn how to build and implement privacy-first ML systems using Federated Learning.
- Learn 6 techniques with implementation to compress ML models.
All these resources will help you cultivate key skills that businesses and companies care about the most.
SPONSOR US
Advertise to 450k+ data professionals
Our newsletter puts your products and services directly in front of an audience that matters β thousands of leaders, senior data scientists, machine learning engineers, data analysts, etc., around the world.