Scan study: 30 popular OSS AI repos -- 202 high-confidence direct SDK call sites across 21 repos -- 0 with cryptographically verifiable evidence emission at those call sites. This is not "no logging" -- many have excellent observability. This measures a stricter property: whether a third party can verify the evidence artifact without access to the producer's infrastructure. Scope: Python source scan using direct SDK detection plus framework heuristics.
Proof run on pydantic-ai (commit-pinned): scan (5 call sites found) -> patch (2 lines auto-inserted) -> run (3 calls exercised) -> verify-pack PASS. https://github.com/Haserjian/assay/blob/280c25ec46afd3ae6938...
Try it:
pip install assay-ai
assay patch .
assay run -c receipt_completeness -- python your_app.py
assay verify-pack ./proof_pack_*
Tamper demo (5 seconds): pip install assay-ai && assay demo-challenge
assay verify-pack challenge_pack/good/ # PASS
assay verify-pack challenge_pack/tampered/ # FAIL (1 byte changed)
Full report + dataset (commit-pinned): https://github.com/Haserjian/assay/blob/280c25ec46afd3ae6938...The verifier is open source -- assay verify-pack is deterministic hash + signature checking. Read it, run it, or write your own.
https://pypi.org/project/assay-ai/
If I missed your instrumentation or a finding is a false positive, drop a commit link and I'll update the dataset.