Skip to content

Commit 7795b5e

Browse files
committed
πŸ“ Expand prompt and eval guidance
1 parent 128a582 commit 7795b5e

File tree

2 files changed

+8
-0
lines changed

2 files changed

+8
-0
lines changed

β€ŽArtificial Intelligence Models.mdβ€Ž

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -56,6 +56,7 @@
5656

5757
- Using LLMs for coding is difficult and unintuitive, requiring significant effort to master.
5858
- English is becoming the hottest new programming language. [Use it](https://addyo.substack.com/p/the-70-problem-hard-truths-about).
59+
- [Prompts are code](https://mariozechner.at/posts/2025-06-02-prompts-are-code/). Markdown and JSON files are state.
5960
- Use comments to guide the model to do what you want.
6061
- Don't delegate thinking, delegate work.
6162
- Describe the problem very clearly and effectively.
@@ -66,6 +67,8 @@
6667
- Provide the desired function signatures, API, or docs. Apply the TDD loop and make the model write tests and the code until the tests pass.
6768
- Prioritize exploration over execution (at first). Iterate towards precision during the brainstorming phase. Start fresh when switching to execution.
6869
- Many LLMs now have very large context windows, but filling them with irrelevant code or conversation can confuse the model. Above about 25k tokens of context, most models start to become distracted and become less likely to conform to their system prompt.
70+
- [Use Progressive Disclosure](https://www.humanlayer.dev/blog/writing-a-good-claude-md) to ensure that the agent only sees tasks or project-specific instructions when it needs them.
71+
- Prefer pointers to files than copies (no code snippets, ...).
6972
- Make the model ask you more questions to refine the ideas.
7073
- Take advantage of the fact that [redoing work is extremely cheap](https://crawshaw.io/blog/programming-with-llms).
7174
- If you want to force some "reasoning", ask something like "[is that a good suggestion?](https://news.ycombinator.com/item?id=42894688)" or "propose a variety of suggestions for the problem at hand and their trade-offs".

β€ŽMachine Learning.mdβ€Ž

Lines changed: 5 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -26,6 +26,11 @@ These points are expanded with more details in courses like [Made With ML](https
2626
- Collecting good evals will make you understand the problem better.
2727
- Working with probabilistic systems requires new kinds of measurement and deeper consideration of trade-offs.
2828
- Don't work if you cannot define what "great" means for your use case.
29+
- [Evals replace LGTM-vibes development](https://newsletter.pragmaticengineer.com/p/evals). They systematize quality when outputs are non-deterministic.
30+
- [Error analysis](https://youtu.be/ORrStCArmP4) workflow: build a simple trace viewer, review ~100 traces, annotate the first upstream failure ([open coding](https://shribe.eu/open-coding/)), cluster into themes ([axial coding](https://delvetool.com/blog/openaxialselective)), and use counts to prioritize. Bootstrap with grounded synthetic data if real data is thin.
31+
- Pick the right evaluator: code-based assertions for deterministic failures; LLM-as-judge for subjective ones. Keep labels binary (PASS/FAIL) with human critiques. Partition data so the judge cannot memorize answers; validate the judge against human labels (TPR/TNR) before trusting it.
32+
- Run evals in CI/CD and keep monitoring with production data.
33+
- [Analyze β†’ measure β†’ improve β†’ automate β†’ repeat](https://newsletter.pragmaticengineer.com/p/evals).
2934
- Good eval metrics:
3035
- Measure an error you've observed.
3136
- Relates to a non-trivial issue you will iterate on.

0 commit comments

Comments
Β (0)