I volunteer on boards and do governance work in my spare time. Last year I had to review ten years of board papers, audit reports, and committee transcripts to understand how a major project went wrong.
I tried ChatGPT. It found things that were "close enough" — paraphrased quotes, approximate summaries, contextually similar passages. But never verbatim text. When I asked for the exact page number, it confidently pointed to the wrong place.
I tried Claude with extended context. It could handle more documents but had the same problem: it would synthesize, it would interpret, it would "helpfully" rephrase. I still had to manually verify every single excerpt.
The problem was not that LLMs are bad at reading — they are incredible. The problem was that for accountability work, "close enough" is not good enough. If I am going to cite something in a report or put it in a story, I need the exact words, the exact page, and I need to know it is real.
So I built Accounter. It only surfaces what it can validate character-by-character. No inference. No paraphrasing. No helpful summaries. Just: "This exact text appears on this exact page of this specific document."
I built it in a month. It works for my use case. Now I need to know if it works for anyone else's.