Subscribe to our Newsletter for Updates and Tips¶
Good LLM Validation is Just Good Validation
What if your validation logic could learn and adapt like a human, but operate at the speed of software? This is the future of validation and it's already here.
Validation is the backbone of reliable software. But traditional methods are static, rule-based, and can't adapt to new challenges. This post looks at how to bring dynamic, machine learning-driven validation into your software stack using Python libraries like Pydantic
and Instructor
. We validate these outputs using a validation function which conforms to the structure seen below.
def validation_function(value):
if condition(value):
raise ValueError("Value is not valid")
return mutation(value)
Enhancing Python Functions with Instructor: A Guide to Fine-Tuning and Distillation
Introduction
Get ready to dive deep into the world of fine-tuning task specific language models with Python functions. We'll explore how the instructor.instructions
streamlines this process, making the task you want to distil more efficient and powerful while preserving its original functionality and backwards compatibility.
If you want to see the full example checkout examples/distillation
RAG is more than just embedding search
With the advent of large language models (LLM), retrieval augmented generation (RAG) has become a hot topic. However throughout the past year of helping startups integrate LLMs into their stack I've noticed that the pattern of taking user queries, embedding them, and directly searching a vector store is effectively demoware.
What is RAG?
Retrieval augmented generation (RAG) is a technique that uses an LLM to generate responses, but uses a search backend to augment the generation. In the past year using text embeddings with a vector databases has been the most popular approach I've seen being socialized.
So let's kick things off by examining what I like to call the 'Dumb' RAG Model—a basic setup that's more common than you'd think.
Generating Structured Output / JSON from LLMs
Language models have seen significant growth. Using them effectively often requires complex frameworks. This post discusses how Instructor simplifies this process using Pydantic.