Articles/Training & Evaluation

Case Study: How a Boutique Law Firm Built a Contract Risk Scoring Tool Without a Technical Team

The problem, the scope, the build, and the outcome. A pattern that applies to more firms than just law.

February 2026·8 min read

Note: this case study is a composite drawn from common patterns across multiple engagements with transactional legal practices. Details have been generalized. The workflow, the problem, the build process, and the outcome reflect real experience - not a single anonymized client.

The firm and the problem

Eight attorneys. A transactional practice covering M&A, commercial real estate, and general commercial work. No in-house technology staff, no dedicated operations team. The partners handled most of the senior work directly.

Contract review was the bottleneck. The firm's partners were spending three to four hours on each incoming contract - reading through, identifying risk clauses, categorizing them by severity, and producing a summary for client discussion. The work followed a consistent logic. The same clause types flagged high-risk for the same reasons. The same indemnification structures warranted the same conversation. The categories were stable.

But the execution required senior judgment to do correctly. So the work landed on partners, who did it well and slowly, because that was the only option.

The managing partner framed it clearly in the first discovery session: "I know what I'm looking for every time. It's the reading and finding it that takes three hours."

Why this workflow was the right starting point

The contract review workflow had properties that made it a strong candidate for automation - and firms evaluating their own options should recognize the pattern.

Clear inputs and outputs

Input: a contract document. Output: a structured summary with flagged clauses, risk scores, and notes for partner review. The format was consistent and the success criteria were specific.

Articulable decision logic

The partners could write down what made a clause high-risk. The criteria were not simple, but they were explicit. Automatic renewal without notice period - high risk. Uncapped liability for consequential damages - high risk. The logic existed in the partners' heads and could be extracted.

High volume, consistent execution

The firm reviewed dozens of contracts per month. Every one required the same process. The same three hours, applied repeatedly to work that followed the same logic every time.

A clear role for human judgment at the end

The tool would not replace the partner's judgment. It would isolate the judgment calls. Instead of reading the whole document to find what needs attention, the partner would review the flags and make decisions about each. The tool does the reading. The partner does the thinking.

This is the pattern worth identifying in any firm: work that is high-volume, logic-driven, and time-consuming in its mechanical execution - but that requires real expertise to evaluate the output. The tool handles the mechanical part. The expert handles the rest. For more on how to identify which workflow fits this pattern at your firm, see how to figure out what to build first.

Discovery: two weeks to map the risk criteria

Discovery ran for two weeks. The primary output was a written risk scoring framework - a document that captured what the firm's partners looked for in each clause category, what made a clause high-risk versus flagged versus acceptable, and what notes were useful to a partner in the review session with a client.

Getting that framework out of the partners' heads was the real work. They knew their criteria. They had never written them down in a form someone else could follow, because there had never been a reason to. The discovery sessions were essentially structured interviews with real contracts: "walk me through how you would look at this clause" repeated across enough examples to capture the patterns and the exceptions.

By the end of discovery, the risk criteria were specific enough to build from. The firm also had something useful independent of the build: a written methodology they could use to train associates and discuss standards internally. That documentation had value regardless of what came next.

The build: six weeks, tested against real work

The build took six weeks. The tool accepted a contract upload, extracted key clauses by type, scored each against the firm's risk criteria, and produced a structured summary - clause-by-clause flags, risk level, and a brief note explaining the flag in terms the partner could take directly into a client conversation.

Testing happened against thirty contracts from past matters, pulled from the firm's files. This was the part that mattered most. Constructed demo examples can look impressive and fail on real work. The thirty past contracts represented the actual distribution of complexity the firm handled - standard commercial agreements, unusual indemnification structures, contracts from industries the firm worked in regularly.

Several adjustments came out of that testing. The initial scoring logic was flagging certain standard limitation-of-liability clauses as high-risk when the partners would typically accept them in commercial real estate contexts. The note format was too brief in some cases - the partners wanted enough detail to use the note in a client call without going back to the contract. Two clause types that appeared frequently were not in the initial scope and needed to be added.

All of those fixes happened before deployment. They were inexpensive to address during the build and would have been significantly more disruptive to address after the tool was live.

The outcome

Contract review time dropped from three to four hours to approximately forty-five minutes. The partner still reviews every contract. The tool does not make the decisions and was never designed to. What it eliminates is the reading-and-flagging work - the mechanical execution of a repeatable process that consumed most of the three hours.

The forty-five minutes is now spent on the flags: evaluating each one, deciding how to address it, and preparing for the client conversation. That is the work that actually required a partner in the first place.

Before

  • 3-4 hours per contract review
  • Partner reads the full document
  • Flags identified manually during reading
  • Notes assembled from memory and marginalia
  • Bottleneck scales with volume

After

  • ~45 minutes per contract review
  • Partner reviews structured flag summary
  • Flags pre-identified and scored by clause type
  • Notes generated in client-ready format
  • Capacity scales without adding senior time

What the handoff looked like

The tool was deployed on the firm's own infrastructure - a straightforward setup that one of the partners managed with documentation provided. No ongoing hosting dependency on the development side.

The handoff included the working tool, the risk scoring framework document from discovery (updated to reflect adjustments made during the build), a maintenance guide covering how to update clause criteria, how to add new clause types, and what to do when the tool flags something the partners disagree with.

In the six months following deployment, the firm updated the risk criteria twice - once when they took on a new practice area and needed to add clause categories, and once when a regulatory change affected how they assessed a specific indemnification structure. Both updates were made by working through the maintenance guide. Neither required outside help.

That is what genuine handoff looks like. The firm is not dependent on the original developers to keep the tool working or to adjust it as the practice evolves. The documentation was specific enough to follow. The logic was visible enough to modify. Read more about why this matters in you own what we build.

The pattern beyond law

The specific workflow here is contract risk review. The underlying pattern applies across professional services: high-volume, logic-driven work that currently requires senior judgment to execute, where the judgment and the mechanical execution are bundled together in a way that makes the work expensive and slow.

Consulting firms have proposal and deliverable review workflows that follow the same structure. Accounting firms have documentation and exception-flagging work. Financial advisors have portfolio review processes. The specific clause types change. The underlying pattern - extract, classify, score, summarize, present for review - does not.

The question to ask at your firm: where is senior time being consumed by work that follows a consistent logic, produces a consistent output format, and could be tested against historical examples? That is the workflow worth examining first.

If you have a workflow that fits this pattern and want to talk through whether a build makes sense - scope, cost, timeline, and what the outcome would look like - see how the engagement works. See also: what happens in a 303 engagement for a detailed walkthrough of each phase.

Next step

Ready to build something your firm owns?

We scope each engagement around your highest-value workflow. The conversation starts with what you want to build and whether we are the right people to build it.