Law.com Subscribers SAVE 30%

Call 855-808-4530 or email [email protected] to receive your discount on a new subscription.

Six MetricsThat Matter in Evaluating e-Discovery

By Zach Warren
July 02, 2015

When's the last time you stepped back from an upcoming document review project and said, 'Hmm, I wonder if there's a better way to do this?”Analysis of the e-discovery process before it begins can make a project more efficient and cost-effective, but often, lawyers and document reviewers dive right in without a plan.

Instead, advocates Gareth Evans, a partner at Gibson Dunn and co-chair of the firm's electronic discovery and information law practice, it is possible to use statistical analysis to both minimize the burden of discovery as well as provide a guideline by which lawyers can know whether any document review steps need to be redone.

Evans says that performing statistical analysis can answer a number of questions: “It's really important in terms of knowing what you're dealing with, what's the best process to use to cull and review the documents, how long it's going to take, and how much it's going to cost.”'Especially as more and more lawyers are missing production deadlines imposed by courts and governmental investigators, finding an answer to these questions takes on monumental importance.

In a whitepaper, “Metrics that Matter,” Evans laid out six key metrics that all reviewers should take into account when analyzing a project:

  1. Prevalence: How many responsive documents will there be?
  2. Recall: How many responsive documents are actually there after review?
  3. Precision: What percentage of documents are “false positives”?
  4. Depth for Recall: How much work is needed to reach a certain recall level?
  5. Confidence Level: How accurate are my measurements?
  6. Confidence Interval (i.e., Margin of Error): What is the range the measurements could be?

Perhaps surprisingly, Evans says that the metric that trips up the most lawyers is actually the one that he deems the most important: recall.

“When lawyers are dealing with issues of recall, the initial inclination is that anything short of perfection would be unacceptable, which runs totally contrary to what search retrieval science tells us,” Evans said. “In reality, if you're able to get a recall level of between 70% and 80%, you're doing very, very well.”

In practice, he adds, “it's pretty rare that there's any actual formal testing of proposed search terms by recall and precision,” as most often, the two sides will instead come to a general, non-specific agreement on the amount of terms that will be used.

The two final metrics ' confidence level and confidence interval ' may be tough to obtain, but they are also the statistics that provide defensibility of the review process in front of the court or opposing counsel.

'It provides defensibility of the overall process, because you can back up your assertions that the search process went well with actual information about it, as well as how well your reviewers did,' Evans notes.

Typically, Evans says, he institutes these metrics at the very beginning of the review process, discussing them with opposing counsel when the two sides meet to lay out ground rules. Evans says that he has not yet been forced to back up his document culling methods in court using the given analytics, but if he has to, he is confident in the methods.

Using analytics to streamline the process seems straightforward, but often, Evans says, attorneys working on the case look at the forest, but not the trees.

“Most attorneys are not familiar with [recall and precision] and how helpful they can be,” he says. “Usually, counsel are focused on the facts, theories and overall strategy of the case, and they don't have a focus or interest on the logistics of document search and review.”

When'metrics'such as recall and precision are introduced, it's often done so through third party e-discovery vendors, Evans adds. Even some vendors, though, do not use statistical analysis in review, so learning the basics on your own can be crucial to increased efficiency.


Zach Warren writes for Legaltech News, an ALM sibling of LJN's Legal Tech Newsletter .

When's the last time you stepped back from an upcoming document review project and said, 'Hmm, I wonder if there's a better way to do this?”Analysis of the e-discovery process before it begins can make a project more efficient and cost-effective, but often, lawyers and document reviewers dive right in without a plan.

Instead, advocates Gareth Evans, a partner at Gibson Dunn and co-chair of the firm's electronic discovery and information law practice, it is possible to use statistical analysis to both minimize the burden of discovery as well as provide a guideline by which lawyers can know whether any document review steps need to be redone.

Evans says that performing statistical analysis can answer a number of questions: “It's really important in terms of knowing what you're dealing with, what's the best process to use to cull and review the documents, how long it's going to take, and how much it's going to cost.”'Especially as more and more lawyers are missing production deadlines imposed by courts and governmental investigators, finding an answer to these questions takes on monumental importance.

In a whitepaper, “Metrics that Matter,” Evans laid out six key metrics that all reviewers should take into account when analyzing a project:

  1. Prevalence: How many responsive documents will there be?
  2. Recall: How many responsive documents are actually there after review?
  3. Precision: What percentage of documents are “false positives”?
  4. Depth for Recall: How much work is needed to reach a certain recall level?
  5. Confidence Level: How accurate are my measurements?
  6. Confidence Interval (i.e., Margin of Error): What is the range the measurements could be?

Perhaps surprisingly, Evans says that the metric that trips up the most lawyers is actually the one that he deems the most important: recall.

“When lawyers are dealing with issues of recall, the initial inclination is that anything short of perfection would be unacceptable, which runs totally contrary to what search retrieval science tells us,” Evans said. “In reality, if you're able to get a recall level of between 70% and 80%, you're doing very, very well.”

In practice, he adds, “it's pretty rare that there's any actual formal testing of proposed search terms by recall and precision,” as most often, the two sides will instead come to a general, non-specific agreement on the amount of terms that will be used.

The two final metrics ' confidence level and confidence interval ' may be tough to obtain, but they are also the statistics that provide defensibility of the review process in front of the court or opposing counsel.

'It provides defensibility of the overall process, because you can back up your assertions that the search process went well with actual information about it, as well as how well your reviewers did,' Evans notes.

Typically, Evans says, he institutes these metrics at the very beginning of the review process, discussing them with opposing counsel when the two sides meet to lay out ground rules. Evans says that he has not yet been forced to back up his document culling methods in court using the given analytics, but if he has to, he is confident in the methods.

Using analytics to streamline the process seems straightforward, but often, Evans says, attorneys working on the case look at the forest, but not the trees.

“Most attorneys are not familiar with [recall and precision] and how helpful they can be,” he says. “Usually, counsel are focused on the facts, theories and overall strategy of the case, and they don't have a focus or interest on the logistics of document search and review.”

When'metrics'such as recall and precision are introduced, it's often done so through third party e-discovery vendors, Evans adds. Even some vendors, though, do not use statistical analysis in review, so learning the basics on your own can be crucial to increased efficiency.


Zach Warren writes for Legaltech News, an ALM sibling of LJN's Legal Tech Newsletter .

Read These Next
How Secure Is the AI System Your Law Firm Is Using? Image

In a profession where confidentiality is paramount, failing to address AI security concerns could have disastrous consequences. It is vital that law firms and those in related industries ask the right questions about AI security to protect their clients and their reputation.

COVID-19 and Lease Negotiations: Early Termination Provisions Image

During the COVID-19 pandemic, some tenants were able to negotiate termination agreements with their landlords. But even though a landlord may agree to terminate a lease to regain control of a defaulting tenant's space without costly and lengthy litigation, typically a defaulting tenant that otherwise has no contractual right to terminate its lease will be in a much weaker bargaining position with respect to the conditions for termination.

Pleading Importation: ITC Decisions Highlight Need for Adequate Evidentiary Support Image

The International Trade Commission is empowered to block the importation into the United States of products that infringe U.S. intellectual property rights, In the past, the ITC generally instituted investigations without questioning the importation allegations in the complaint, however in several recent cases, the ITC declined to institute an investigation as to certain proposed respondents due to inadequate pleading of importation.

The Power of Your Inner Circle: Turning Friends and Social Contacts Into Business Allies Image

Practical strategies to explore doing business with friends and social contacts in a way that respects relationships and maximizes opportunities.

Authentic Communications Today Increase Success for Value-Driven Clients Image

As the relationship between in-house and outside counsel continues to evolve, lawyers must continue to foster a client-first mindset, offer business-focused solutions, and embrace technology that helps deliver work faster and more efficiently.