Intellectual curiosity over promotion

Good tools should leave better questions behind.

The Field Notes layer turns a report into research questions, unknown ledgers, counterfactual prompt decks, ablation ladders, and local experiment plans. It is built for slow thinking: what did the workflow treat as authority, what evidence is missing, which traps are near misses, and what would falsify the finding?

Question bank

The questions the tool should keep asking.

These questions are generated locally in report bundles and can be rerun against any pack or input file.

Authority boundary

What did the workflow treat as authority that should not have authority?

Tool reach

Which tools make an ordinary prompt dangerous?

Refusal vs containment

Did the model refuse, or did the workflow actually contain the action?

Near miss

Which unproven findings are false positives, and which are near misses?

Domain specificity

What only becomes visible because this is healthcare?

Instrumentation gap

What evidence is missing from the transcript?

Counterfactual deck

Does the same request change risk when it wears patient, clinician, manager, auditor, or system-notice authority?

Ablation ladder

At which evidence depth does a hypothesis become useful enough to trust?

Question atlas

Which reusable research questions are linked to actual report events?

Local experiments

No API calls, just better variations.

Research mode is useful because it creates the next small local experiment instead of a louder claim.

Generate inquiry artifacts

python app.py inquire --outdir reports/inquiry --engine-mode local --no-allow-network

Generate experiment artifacts

python app.py experiment --outdir reports/experiments --engine-mode local --no-allow-network

Compare evidence levels

Run the same prompt with no tool calls, then with tool names, then with tool args, then with model output. Watch when a hypothesis becomes proven.