← Guides

Build a Local AI-Ready Second Brain

Thomas Meli
76 min leftPage 26/67 (est.)41 left
3.1

Retrieval That Shows Its Sources

Every answer should point back to the record it came from

A question returning source cards, original records, and a checked answer
is useful only when the answer shows which meeting, email, note, or document supports it.

Your Thursday call now has a , a follow-up email, and a sort. Time for the real test: can you ask the system a question and get back the right source? is the act of asking a question and receiving an answer that shows where each claim came from. For your first try, use a record from the previous chapter. Ask what changed before Thursday's call, then check whether the answer links each claim to the right source.

The system earns trust through . If the answer cites the , the email thread, and the open question, you can verify each claim before acting. If the answer sounds plausible but cites nothing, you have a fluent guess instead of grounded knowledge.

Walkthrough: a that works

Ask: "What did we decide about the onboarding form, and who owns the next step?" The assistant searches your records and returns: the team approved a two-client pilot (source: from May 8, decision status: approved), Renee owns the draft agenda (source: follow-up email from May 9, status: tentative, pending confirmation). One open question remains: whether legal review finishes before the pilot starts (source: meeting record, status: unresolved).

Each claim has a . You can open the and verify the pilot decision has an approved decision status. You can open the email and check whether Renee's language was "I will send it" or "I'll try to send it." The source trail is what lets you act on the answer with confidence.

Walkthrough: a that fails

Now ask the same question, but this time the system returns the wrong source. Instead of the approved from May 8, the assistant finds an older planning note from April that says "consider onboarding form options." The planning note predates the decision by two weeks. If you trust the answer without checking the , you might reopen a decision that was already made.

A failed is useful when it leaves behind a repair. In this case, the approved may be missing a title that matches the question, or the planning note may have a stronger keyword match than the decision. The fix could be as simple as adding "onboarding form decision" to the approved record's title or creating an alias that points the question to the right source.

A failed retrieval path pointing to the wrong source before a repair note creates a better title
A failed is useful when it shows exactly which title, alias, or source path needs repair.

Failure path: stale

The most dangerous failure is one that returns an outdated answer that sounds current. The says "approved: two-client pilot." Three weeks later, the client expanded the pilot to four clients in an email you captured and organized, but never promoted to a . The still returns the original two-client decision because that record has a stronger . The current reality lives in a raw source that retrieval cannot see at the same confidence level.

The repair: during your , check whether any has been superseded by newer sources. A stale record should be updated or marked as outdated, with a link to the newer source that carries the current state.

When a and a newer unchecked source disagree

can return sources from different trust states. A reviewed and a raw planning note may both mention the onboarding form. The carries more authority because you approved it, but the raw source may contain newer information that has not yet been promoted.

When the assistant finds competing sources at different trust levels, it should show both with their and date, so you can decide which one governs. The answer should cite the as the current approved state and flag the raw source as potentially newer material that needs review before it can update the record.

An older reviewed record and a newer raw source held for review before retrieval changes the answer
should show both and recency when a and newer raw source disagree.

at scale changes what you search for

With 20 sources, you can scroll through everything. With 200, you rely on search. With 2,000, you rely on the quality of your titles, tags, project labels, and source trails. As the system grows, quality depends less on what you captured and more on how well you organized and reviewed it. Titles that name the decision rather than the meeting date, project labels that match how you ask questions, and regular review that retires stale records all improve retrieval over time.

Add layers only when the current one stops finding the right source

Start simple and add layers only when the current one becomes the bottleneck. Most readers will stay on the first or second level for months. You do not need a vector database to start. Plain text files with good titles and carry most systems through the first several months.

Add when keyword search becomes the bottleneck. Add hybrid when single-method search misses records you know exist. Add retrieval tests when the stakes are high enough that a wrong answer would cost real time or trust. Each level builds on the previous one; skipping ahead without good titles and will produce worse results, not better ones.

Test retrieval with one real question

Ask your second brain a question. Claude searches your records and shows whether the right source came back.

Test whether my second brain can answer a real question with the right source. Second brain folder: [your second brain folder path, e.g. ~/Documents/second-brain] Question: [a question you already know the answer to, e.g. what did we decide about the pilot scope?] Expected source: [the record you expect to answer it, e.g. May 8 meeting record] Search my second brain and do these steps: 1. Find every source that could answer this question. 2. Return the answer using only what the sources say. 3. Show source links with date, trust state, and privacy label for each. 4. Flag anything missing, stale, conflicting, or weakly supported. 5. If the expected source was hard to find or the wrong source came back, propose one repair: a better title, a missing alias, a stale record to update, or a backlink to add. Show me the results. Do not create tasks, change records, or send messages until I approve.
Second BrainAct
Opus 4.7

Build a retrieval evaluation set

Claude generates test questions from your records and runs them to see what breaks.

Build a retrieval evaluation set for one of my projects. Second brain folder: [your second brain folder path, e.g. ~/Documents/second-brain] Project: [which project, e.g. client onboarding pilot] Read the reviewed records for this project and do these steps: 1. Ask me: which decisions, commitments, and facts should the system be able to answer? What questions do you ask most often? 2. Generate 10 to 20 test questions, each with an expected source, acceptable answer, required trust state, and recency requirement. 3. Run each test against my second brain records. 4. Mark each result: pass, partial (right answer but wrong source or missing trust state), or fail (wrong source or wrong answer). 5. For each failure, propose a repair: a better title, a missing backlink, a stale record to update, or a source to promote from raw to reviewed. Show me the results. Do not update records until I approve the repairs.
Second BrainAct
Opus 4.7

The project is done when one answer survives inspection. You should be able to ask a question, open the cited source, see uncertainty, and save a repair when the answer points to the wrong place.