The failure of “Simple Dredge” taught us a hard lesson: You cannot cheat physics. Computation requires energy. In an LLM, energy is tokens.
If you want the model to resolve a complex contradiction (“Auth vs. Network”), it must generate the tokens required to step through that logic. You cannot compress the thought process into 7 characters (“Network”), even if the result is 7 characters.
This is the Thinking Token Tax. To get the correct answer, you must pay for the reasoning tokens. If you suppress them (for the sake of density), you suppress the intelligence.
Native Function Calling won the “Reasoning” benchmark because it likely performs a “Hidden Chain of Thought” (internal reasoning) before generating the JSON. But this reasoning is invisible to us.
We matched this performance using Reasoning Dredge. We introduced a Compound Mold:
["ANALYSIS_TRACE: ", {name: "trace"}, "\nVERIFIED_ROOT_CAUSE: ", {name: "cause"}]
By forcing the model to fill the TRACE slot first, we gave it a “Scratchpad.” We authorized it to spend tokens on thinking. The result? It successfully tracked the conversation: "Alice blamed Auth -> Bob checked logs -> Charlie found Network flapping."
Once this trace existed in the context window, the final slot ([CAUSE]) became trivial. The “Auth” hallucination was impossible because the Trace explicitly debunked it.
Reasoning Dredge achieved what we call the Holy Trinity of Cognitive Engineering:
This destroys the trade-off. We don’t have to choose between “Smart but Chatty” and “Dumb but Dense.” We can have “Smart, Dense, and Auditable” by separating the Reasoning Phase (Trace) from the Extraction Phase (Result).
This discovery has profound implications for “High-Stakes AI” (Law, Medicine, Finance).
In these fields, an answer without a reason is worthless. A diagnosis of “Cancer” is malpractice; a diagnosis of “Cancer because X, Y, Z” is medicine.
By making the Trace a first-class citizen of the data structure, Dredge transforms AI from a “Black Box Oracle” into an “Auditable Analyst.” We don’t just trust the output; we verify the Trace. If the Trace cites a non-existent legal precedent, we catch it. This is the only path to regulatory compliance for AI agents.