It is 2 a.m. on a Wednesday. A deal team has two hundred target-company contracts open across four associates. They are looking for change-of-control consents, assignment restrictions, and exclusivity clauses. The signing date is Monday. The partner wants a single matrix showing every consent obligation, which counterparty holds it, the exact contract language, and whether any of it can be waived before close.
This is where most AI tools fall apart. Not because they cannot read a contract. They can. They fall apart because they were built for chat, and a chat interface is a profoundly wrong tool for the job a deal team actually needs done. AI for M&A due diligence document review is a structured-output problem, not a conversation problem. That distinction is the whole piece.
The real DD problem is scale, not speed
Every legal-AI deck opens with "3x faster review." The claim is true and also the wrong frame. A senior associate can read a single contract quickly. That is not where deals break.
Deals break because no single human can hold two hundred contracts in their head well enough to spot the one that requires board-level consent from a counterparty with a change-of-control put option priced above the deal. One contract out of two hundred. That contract is the deal. If you miss it, you blew the closing. If you find it late, you blew the purchase price.
The problem is not speed. The problem is structured coverage at scale with zero missed provisions that actually matter. Speed is a side effect of doing that well. Framing it as a speed problem leads you to tools that summarize faster. Framing it as a coverage problem leads you to tools that produce matrices.
Why chat-based AI fails at due diligence
A chat window is a one-document-at-a-time tool. You paste a contract in, you ask about consents, you get a paragraph. That works if you have five contracts. At two hundred, you would do two hundred paste-ask cycles, and your output is two hundred paragraphs in a transcript. That is not a deliverable. A partner cannot open a transcript and find the deal-breaker.
Three specific failure modes show up whenever deal teams try to push real DD through a chat interface.
No structure. A matrix has rows (contracts), columns (provisions), and cells with the exact language. A chat gives you prose. Converting prose back to a matrix is manual work. You are now using AI to generate text that a human has to parse into a spreadsheet. The work of parsing the output is the work you were trying to avoid.
Second, no cross-document comparison. The interesting finding in DD is almost never inside one contract. It is the outlier. The one supply agreement whose termination-on-change-of-control language is two standard deviations stricter than the other nineteen. A chat cannot see that, because it only saw one contract at a time. A matrix sees it instantly, because the outlier is the cell that does not match.
Third, no verification layer. A chat will happily paraphrase a clause. Paraphrase drifts. The attorney signing the disclosure schedule needs the exact words the contract uses, not the model's reading of them. Without verification against the source document, every paraphrase is a small liability the deal team has to catch in review. At scale, they will not catch all of them.
Cornell's Legal Information Institute defines due diligence as "the level of reasonable care or attention expected to avoid liability" in examining records and performing acquisitions. Reasonable care on two hundred contracts cannot be two hundred chat sessions. The tool has to match the shape of the work.
What structured output actually means
An attorney working on consent analysis across a target company asks one question: "For every contract in the target's VDR, identify whether the contract requires counterparty consent for change of control. Return the exact contractual language, the counterparty name, and a yes/no/conditional verdict."
The answer is not a paragraph. It is a matrix. Two hundred rows. Five columns. Each cell traceable to the exact page in the exact document. The attorney can sort the matrix by verdict and immediately see every "yes" and every "conditional." The partner can review the twelve contracts that matter instead of the two hundred that do not. The disclosure schedule writes itself from the matrix.
A rep-and-warranty comparison across five draft SPAs works the same way. Rows are R&W categories. Columns are the five drafts. Cells show which formulation each draft uses, which carve-outs, which knowledge qualifiers, which materiality thresholds. Divergences jump off the page. The partner walks into the negotiating session with a document that tells her where every draft agrees, where they differ, and where the differences matter.
A compliance gap analysis works the same way. Rows are regulatory requirements. Columns are the target's policies and agreements. Cells show whether each requirement is addressed, addressed but weak, or absent. Gaps are obvious. Priorities are obvious. The audit follow-up list writes itself.
The unifying pattern: the deliverable is a table, not a transcript. AI for DD is valuable only when the output has the shape of the work product a partner can actually use.
Why accuracy matters more in DD than anywhere else
In research, a hallucinated citation is a disaster, but a disaster the supervising attorney can usually catch before filing. In drafting, a bad paraphrase is a problem, but the drafter is reading every word anyway. In DD, the attorney is trusting the matrix. If the matrix says "no consent required" and the contract actually requires consent, nobody is going to re-read the two hundred contracts to catch the error. That is the entire reason the matrix exists.
So the accuracy question matters differently here. In DD, the cost of one missed provision can be the deal.
In internal testing, Aewita observed zero hallucinated outputs across 800 consecutive queries — statistically, a rough upper bound under 0.3% at 95% confidence. Ask your chat-based AI for its number. If they will not give you one, that is your answer.
The methodology behind that number is published at /blog/legal-ai-hallucination-rates. Short version: 800 queries, 22 practice areas, attorney review, binary scoring, standard statistical framing for the upper confidence bound. No fine print.
A missed consent requirement can blow up a deal. You want tools that do not miss.
What Aewita does differently
From the attorney's chair. Load the deal, which in practice means pointing Aewita at the virtual data room or the set of contracts already sitting in your DMS. Select the analysis type — consent mapping, R&W comparison, compliance gap, something custom the firm has standardized. Run it. Review the structured result.
The result is a matrix built on primary sources. Every cell in the matrix is grounded in retrieved contract text. Every quoted provision is a verbatim pull from the source document, with a pointer back to the exact page and paragraph. Every verdict — yes, no, conditional, silent — is supported by the provision that drove it.
What you do not see, but should want: every claim in the matrix is independently verified before the attorney sees it. If a quoted passage does not match the source document exactly, it does not ship. If a verdict is drawn from a provision that was retrieved but not on point, it gets flagged rather than asserted. This is the difference between a tool that can be trusted on its face and a tool whose output every associate still has to re-read start to finish.
We built this ourselves and we host it ourselves. Your deal documents never leave the Aewita environment to talk to an outside model provider. For a deal team that has been worrying about third-party data exposure since the first chat-based legal tools launched, this is not a nice-to-have. It is the only reason the tool can be used on live transactions at all. The security architecture is worth the five-minute read before any deal touches a new system.
Three categories of DD questions Aewita handles
Consent and assignment mapping. Across every contract in the target's VDR, identify change-of-control consents, anti-assignment clauses, and any related triggers. Produce a matrix sortable by counterparty, contract type, and verdict. This is the single most common AI-for-DD use case, and the one where the cost of a miss is highest.
Representation and warranty comparison. Side-by-side comparison of reps, warranties, and covenants across draft SPAs, LOIs, or term sheets. Matrix shows the exact formulation in each draft, the carve-outs, the materiality qualifiers, the knowledge qualifiers, and the survival periods. A partner can walk into negotiation with the full landscape in one document.
Compliance and regulatory gap analysis. Across the target's policies, agreements, and disclosures, identify coverage of a defined set of regulatory requirements — data privacy, anti-bribery, export control, labor, whatever is in scope. Matrix shows addressed, weakly addressed, or absent, with the supporting provision cited. The gap list becomes the remediation plan.
These are not exhaustive. They are the three that come up in almost every transaction. A firm that standardizes these three saves hundreds of associate-hours per deal and produces more consistent deliverables across practice groups.
What to ask before you trust any tool on a live deal
If you are a corporate partner weighing AI for DD, the questions are specific.
First, what does the output look like? Ask for a sample matrix on a sample deal. If the demo is a chat session, the tool is the wrong shape for your work.
Second, what is the accuracy number, with sample size and confidence interval? A point estimate from a small test is not an answer. A verified rate with a disclosed methodology is.
Third, where does my deal data go? If the vendor sends contracts to a third-party model provider, the deal team now has a data-exposure problem on top of a DD problem. The answer you want is a self-hosted system with the data path fully disclosed.
Fourth, is every provision reference traceable back to the source document? If the matrix cannot point to the exact page, the matrix is not grounded. Every cell should be independently verifiable by the associate in ninety seconds or less.
For how Aewita fits into drafting and research alongside DD, see /product/drafting and /product/research. For a head-to-head with the other systems corporate partners are considering, /compare walks through the comparison in detail.
The short version
Chat is the wrong tool for due diligence. Deal teams need structured matrices across entire contract populations, with every cell grounded in the source document and every quote verified before it reaches the attorney. The accuracy bar is higher than anywhere else in a firm's work, because the deliverable is trusted in a way a research memo is not.
AI for M&A due diligence document review, done right, is a quiet tool. It does not make you feel clever. It makes the matrix exist, on time, correct, one-click to the exact provision. The deal team spends its energy on the provisions that matter. The partner walks into the closing prepared. That is the whole product.
See a real DD matrix on a real deal
Thirty-minute demo with the product team. Bring a redacted VDR or a sample contract set. We will run it live.