Q-Ledger is a machine-first observation ledger derived from edge observations, designed to make governance artefacts detectable, traceable, and chainable over time.
Scope: observation, not attestation. Q-Ledger proves neither identity, nor intent, nor legal compliance. It documents that a surface was observed as published and, in some cases, consulted within a declared window.
Why Q-Ledger exists
In an interpreted web, systems reconstruct context from partial signals. Q-Ledger publishes a weak but structured memory of that detectability:
- which machine-first entry points exist;
- which snapshots were published;
- what continuity is visible;
- which ruptures become detectable.
Q-Ledger does not settle truth. It preserves the conditions of observation from which later discussions about continuity, drift, or correction can become less anecdotal.
What Q-Ledger can show
Q-Ledger can show:
- that a set of machine-first artefacts was publicly published during a given window;
- that certain endpoints or artefacts were observed as consulted;
- that successive snapshots exist with a chaining logic;
- that continuity or rupture becomes visible.
Its purpose is to make publication historically legible.
What Q-Ledger cannot prove
Q-Ledger does not prove:
- the identity of the actor behind the artefacts;
- the intent behind a consultation;
- editorial or legal compliance;
- the absolute completeness of observation;
- the fidelity of a synthesis produced from those surfaces.
That is why Q-Ledger must remain connected to Observation vs attestation: why Q-Ledger is deliberately weak.
Published artefacts
Primary entry points:
- JSON:
/.well-known/q-ledger.json - YAML:
/.well-known/q-ledger.yml - Latest:
/.well-known/q-ledger-latest.json
Context artefacts often read together with Q-Ledger:
Chaining, continuity, and rupture
Q-Ledger relies on a logic of snapshots, hashes, and previous references. The interest is not cryptographic by itself. The interest is interpretive:
- make visible that a sequence of publications exists;
- make a rupture or lacuna easier to contest;
- prevent a silent correction or silent change from erasing all memory of what was previously published.
Continuity does not establish absolute truth. It makes a history more readable.
Relation to Q-Metrics and observations
Q-Ledger is not a dashboard. It is closer to a weak memory of observed conditions.
Q-Metrics then condenses some of those signals into comparable indicators. Observations serves as the descriptive hub that connects snapshots, windows, limits, and interpretations.
The proper chain is therefore:
published artefacts → weak observation → chaining → metric condensation → doctrinal reading
Why Q-Ledger matters in a machine-first device
Publishing governance files is not enough. It must still become possible to document their continuity, visibility, and publication stability.
In that logic, Q-Ledger helps make auditable the coupling between:
- machine-first architecture;
- governance files;
- identity and boundary surfaces;
- observation and measurement layers.
This is what connects Q-Ledger to Machine-first is not enough: why governance files change the reading regime and What each governance file actually does.
Limits and biases
- Observations depend on edge visibility, caches, and access conditions.
- Silence does not prove absence: an artefact may exist without being observed in the window.
- An observed consultation does not prove correct use of the surface.
- A published snapshot does not guarantee that every system will take it into account.
Q-Ledger therefore remains descriptive. Its strength comes precisely from the fact that it does not try to promise more than it can support.