Skip to content

Commit 06928bf

Browse files
authored
Merge branch 'main' into evaluating-search-engines
2 parents 8499a94 + 8273c06 commit 06928bf

30 files changed

+18683
-8105
lines changed

notebooks/en/_toctree.yml

+18-5
Original file line numberDiff line numberDiff line change
@@ -5,6 +5,12 @@
55
- local: index
66
title: Overview
77

8+
- title: MLOps Recipes
9+
isExpanded: false
10+
sections:
11+
- local: mlflow_ray_serve
12+
title: Signature-Aware Model Serving from MLflow with Ray Serve
13+
814
- title: LLM Recipes
915
isExpanded: false
1016
sections:
@@ -66,7 +72,9 @@
6672
title: Enhancing RAG Reasoning with Knowledge Graphs
6773
- local: phoenix_observability_on_hf_spaces
6874
title: Phoenix Observability Dashboard on HF Spaces
69-
75+
- local: search_and_learn
76+
title: Scaling Test-Time Compute for Longer Thinking in LLMs
77+
7078
- title: Computer Vision Recipes
7179
isExpanded: false
7280
sections:
@@ -100,23 +108,26 @@
100108
title: Fine-tuning SmolVLM with TRL on a consumer GPU
101109
- local: multimodal_rag_using_document_retrieval_and_smol_vlm
102110
title: Smol Multimodal RAG, Building with ColSmolVLM and SmolVLM on Colab's Free-Tier GPU
103-
111+
- local: fine_tuning_vlm_dpo_smolvlm_instruct
112+
title: Fine-tuning SmolVLM using direct preference optimization (DPO) with TRL on a consumer GPU
113+
- local: structured_generation_vision_language_models
114+
title: Structured Generation from Images or Documents Using Vision Language Models
104115

105116
- title: Search Recipes
106117
isExpanded: false
107118
sections:
108119
- local: semantic_reranking_elasticsearch
109120
title: Semantic Reranking with Elasticsearch
121+
- local: vector_search_with_hub_as_backend
122+
title: Vector Search on Hugging Face with the Hub as Backend
110123

111124
- title: Agents Recipes
112125
isExpanded: false
113126
sections:
114127
- local: agents
115-
title: Build an agent with tool-calling superpowers using Transformers Agents
128+
title: Build an agent with tool-calling superpowers using smolagents
116129
- local: agent_rag
117130
title: Agentic RAG - turbocharge your RAG with query reformulation and self-query
118-
- local: agent_change_llm
119-
title: Create a Transformers Agent from any LLM inference provider
120131
- local: agent_text_to_sql
121132
title: Agent for Text-to-SQL with automatic error correction
122133
- local: agent_data_analyst
@@ -125,6 +136,8 @@
125136
title: Have several agents collaborate in a multi-agent hierarchy
126137
- local: multiagent_rag_system
127138
title: Multi-agent RAG System 🤖🤝🤖
139+
- local: mongodb_smolagents_multi_micro_agents
140+
title: MongoDB + SmolAgents Multi-Micro Agents to facilitate a data driven order-delivery AI agent
128141

129142
- title: Enterprise Hub Cookbook
130143
isExpanded: True

notebooks/en/advanced_rag.ipynb

+3-3
Original file line numberDiff line numberDiff line change
@@ -28,7 +28,7 @@
2828
"\n",
2929
"In this notebook, we will take a look into many of these blue notes to see how to tune your RAG system and get the best performance.\n",
3030
"\n",
31-
"__Let's dig into the model building!__ First, we install the required model dependancies."
31+
"__Let's dig into the model building!__ First, we install the required model dependencies."
3232
]
3333
},
3434
{
@@ -39,7 +39,7 @@
3939
},
4040
"outputs": [],
4141
"source": [
42-
"!pip install -q torch transformers transformers accelerate bitsandbytes langchain sentence-transformers faiss-cpu openpyxl pacmap datasets langchain-community ragatouille"
42+
"!pip install -q torch transformers accelerate bitsandbytes langchain sentence-transformers faiss-cpu openpyxl pacmap datasets langchain-community ragatouille"
4343
]
4444
},
4545
{
@@ -272,7 +272,7 @@
272272
"👀 As you can see, __the chunk lengths are not aligned with our limit of 512 tokens__, and some documents are above the limit, thus some part of them will be lost in truncation!\n",
273273
" - So we should change the `RecursiveCharacterTextSplitter` class to count length in number of tokens instead of number of characters.\n",
274274
" - Then we can choose a specific chunk size, here we would choose a lower threshold than 512:\n",
275-
" - smaller documents could allow the split to focus more on specific ideas.\n",
275+
" - Smaller documents could allow the split to focus more on specific ideas.\n",
276276
" - But too small chunks would split sentences in half, thus losing meaning again: the proper tuning is a matter of balance."
277277
]
278278
},

0 commit comments

Comments
 (0)