You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Revise workshop chapters for clarity, structure, and user engagement
- Updated chapter titles for consistency and clarity across all chapters.
- Enhanced learning objectives and prerequisites to guide readers effectively.
- Improved formatting and readability throughout chapters, including better spacing and organization of content.
- Incorporated actionable insights and real-world examples to enhance user understanding and engagement.
- Emphasized the importance of feedback mechanisms and performance metrics in RAG systems.
Copy file name to clipboardExpand all lines: docs/workshops/chapter0.md
+32-9Lines changed: 32 additions & 9 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,24 +1,31 @@
1
1
---
2
-
title: "Beyond Implementation to Improvement: A Product Mindset for RAG"
2
+
title: "0. Beyond Implementation to Improvement"
3
3
description: Rethinking RAG as a continuously evolving product rather than a static implementation
4
4
authors:
5
5
- Jason Liu
6
6
date: 2025-02-28
7
7
tags:
8
-
- productthinking
8
+
- product-thinking
9
9
- systems
10
-
- RAG
10
+
- rag
11
11
- fundamentals
12
12
- improvement
13
13
---
14
14
15
15
# Beyond Implementation to Improvement: A Product Mindset for RAG
16
16
17
-
Look, I've been building AI systems for over a decade, and I keep seeing the same mistake: teams ship a RAG system, pat themselves on the back, and then watch it slowly fail in production.
17
+
## Learning objectives
18
+
19
+
- Adopt a product mindset for RAG (continuous improvement vs one-off build)
20
+
- Reframe RAG as a recommendation engine with feedback
21
+
- Understand the improvement flywheel used across chapters
22
+
23
+
Look, I've been building AI systems for over a decade, and I keep seeing the same mistake: teams ship a RAG system, pat themselves on the back, and then watch it slowly fail in production.
18
24
19
25
This chapter is about avoiding that trap. We're going to talk about why the most successful RAG systems aren't the ones with the fanciest embeddings or the biggest context windows—they're the ones that get better every week based on what users actually do with them.
20
26
21
27
Here's what we'll cover:
28
+
22
29
- Why thinking of RAG as a "project" instead of a "product" dooms most implementations
23
30
- How to steal ideas from recommendation systems (because that's really what RAG is)
24
31
- A practical framework for turning user frustration into system improvements
@@ -37,13 +44,15 @@ I've built AI systems at Facebook, Stitch Fix, and worked with companies like Hu
37
44
Here's a quick way to tell which mindset a team has:
38
45
39
46
**Implementation Mindset:**
40
-
- "We need to implement RAG"
47
+
48
+
- "We need to implement RAG"
41
49
- Obsessing over embedding dimensions and context windows
42
50
- Success = it works in the demo
43
51
- Big upfront architecture decisions
44
52
- Focus on picking the "best" model
45
53
46
54
**Product Mindset:**
55
+
47
56
- "We need to help users find answers faster"
48
57
- Tracking answer relevance and task completion
49
58
- Success = users keep coming back
@@ -129,7 +138,7 @@ What's great about this is that it compounds. More data leads to better insights
129
138
130
139
### Optimizing Feedback Collection
131
140
132
-
**A quick story about feedback:** We spent weeks at one company getting almost no user feedback. Then we changed the prompt from "How did we do?" to "Did we answer your question?" Feedback rates went up 5x overnight.
141
+
**A quick story about feedback:** We spent weeks at one company getting almost no user feedback. Then we changed the prompt from "How did we do?" to "Did we answer your question?" Feedback rates went up 5x overnight.
133
142
134
143
Here's what actually works:
135
144
@@ -184,14 +193,17 @@ Without a systematic approach, teams face significant challenges:
184
193
Here's what happens in real meetings:
185
194
186
195
**"Make the AI better"**
196
+
187
197
- Without a system: Everyone looks nervous, suggests random ideas
188
198
- With a system: "Our top failure mode is date-related queries at 23% error rate. Here's our plan."
189
199
190
200
**"Where should we focus engineering time?"**
201
+
191
202
- Without a system: Whoever argues loudest wins
192
203
- With a system: "42% of failures are inventory problems. Let's start there."
193
204
194
205
**"Is this new embedding model worth it?"**
206
+
195
207
- Without a system: "The benchmarks look good?"
196
208
- With a system: "It improves our technical documentation queries by 15% but hurts on short questions. Not worth it."
197
209
@@ -215,7 +227,7 @@ The shift from engineer to product thinker is subtle but powerful. Here's how yo
215
227
216
228
This shift doesn't mean abandoning technical rigor. It means applying that rigor to problems that actually matter to your users, guided by data instead of assumptions.
217
229
218
-
**Quick story:** A restaurant chain spent months perfecting their voice AI's speech recognition. Then someone actually listened to the call recordings. Turns out 30% of callers were asking "What's good here?"
230
+
**Quick story:** A restaurant chain spent months perfecting their voice AI's speech recognition. Then someone actually listened to the call recordings. Turns out 30% of callers were asking "What's good here?"
219
231
220
232
They added a simple feature: when someone asks that, the AI recommends the three most popular items. Revenue went up 9%. They didn't improve the AI at all—they just paid attention to what people actually wanted.
221
233
@@ -232,6 +244,7 @@ Let me walk you through how this played out with a legal tech company I worked w
232
244
**Step 3:** Shipped it and watched what lawyers actually did. Added thumbs up/down buttons and tracked what they copied.
233
245
234
246
**Step 4:** After 2 months and 5,000 queries, patterns emerged. Three main query types:
247
+
235
248
- Case citations (worked great)
236
249
- Legal definitions (OK)
237
250
- Procedural questions (total failure)
@@ -245,12 +258,14 @@ End result: lawyers actually started using the system. Research time dropped 40%
245
258
**Pro tip:** When something's not working, first ask: "Is this an inventory problem or a capabilities problem?"
246
259
247
260
**Inventory problem:** You don't have the answer in your knowledge base
261
+
248
262
- Missing documents
249
263
- Outdated info
250
264
- Gaps in coverage
251
265
- Fix: Add more/better content
252
266
253
267
**Capabilities problem:** You have the answer but can't find it
268
+
254
269
- Bad retrieval
255
270
- Wrong search strategy
256
271
- Can't understand the query
@@ -261,13 +276,14 @@ I've seen teams waste months improving retrieval when they simply didn't have th
261
276
## Who This Is For
262
277
263
278
Based on who's shown up to my workshops, you're probably:
279
+
264
280
- A technical leader trying to figure out why your RAG system isn't getting better
265
281
- An engineer who built a RAG system and is now stuck maintaining it
266
282
- Part of a team (engineering, data science, product) trying to make AI actually useful
267
283
268
284
I've taught this to teams at tiny startups and big tech companies. The problems are surprisingly similar—everyone's trying to move from "we built RAG" to "our RAG system gets better every week."
269
285
270
-
## What's Coming Next
286
+
## What’s next
271
287
272
288
Each chapter builds on the last, taking you through the complete improvement flywheel. Everything includes code and examples you can steal for your own projects.
273
289
@@ -309,6 +325,13 @@ Here's what changes when you get this right:
309
325
310
326
The difference is night and day. Teams without a system spin their wheels. Teams with a system ship improvements every week.
311
327
328
+
## Summary
329
+
330
+
- Treat RAG as a product with a continuous improvement flywheel
331
+
- Focus on leading metrics (experiments, retrieval precision/recall)
332
+
- View RAG as a recommendation engine wrapped around an LLM
333
+
- Build feedback loops and monitoring from day one
334
+
312
335
## Reflection Questions
313
336
314
337
As you prepare for the next chapter, consider these questions about your current approach to RAG:
@@ -331,6 +354,6 @@ _Note: I've used this approach with companies across legal, finance, healthcare,
331
354
332
355
---
333
356
334
-
IF you want to get discounts and 6 day email source on the topic make sure to subscribe to
357
+
If you want to get discounts and 6 day email source on the topic make sure to subscribe to
0 commit comments