Bits + Bytes Round-Up – News and updates from around the ediscovery landscape

The Bits + Bytes Round-Up provides capsule updates on developments and emerging stories in the ediscovery field. This week’s Round-Up covers trends in global litigation studies, the OPM data breach, social media evidence in divorce cases, and more. Global litigation survey highlights growing ediscovery costs, prevalence of TAR, mobile data’s rise – Issued last week, […]

Five ways to avoid getting sued for discovery malpractice

This article originally appeared on the Logikcull blog and is reproduced here courtesy of Robert Hilson, director for Logikcull. He can be reached at Data-related fiascos that lead to malpractice suits, or disputes that would otherwise spawn malpractice claims if they weren’t settled quietly out of court, happen all the time. You don’t hear much […]

Predictive Coding? Let’s Talk about Predictive Pricing

by Scott Hipsman Solutions Architect, DTI Over the last several years, the majority of discussions in the eDiscovery market have been around cost reduction through the use of predictive coding. While this technology, along with other data analysis techniques, can certainly drive down costs, they do not help predict costs early in the process. So, […]

In high-stakes world of FCPA internal investigations, ediscovery strategies rise in importance

By Daniela Guzman June 3, 2015   In the mid-2000s, the US Securities and Exchange Commission and the Department of Justice dusted off an almost 30-year old law called the Foreign Corrupt Practices Act (FCPA), using it to pursue companies that paid bribes to officials overseas to win government business or influence contracting decisions. […]

Using extrapolated precision for performance measurement

This article originally appeared at and has been republished here with the permission of its author, Bill Dimm.  This is a brief overview of my paper “Information Retrieval Performance Measurement Using Extrapolated Precision,” which I’ll be presenting on June 8th at the DESI VI workshop at ICAIL 2015. The paper provides a novel method […]

Robust metrics needed to counter impact of ‘data shift’ on predictive coding measures

Attorneys and judges often rely exclusively upon “precision” and “recall” thresholds for acceptance of dichotomous classification models in what is commonly referred to in the legal industry as “predictive coding.” Because these measures fail to provide a complete understanding of the proposed model’s characteristics and efficacy, this paper will argue that interested parties should go […]

The Single Seed Hypothesis

This article was originally published at Clustify Blog. It is republished here with the permission of its author, Bill Dimm.  This article shows that it is often possible to find the vast majority of the relevant documents in a collection by starting with a single relevant seed document and using continuous active learning (CAL). This […]

Major firm’s bid to reverse training video sanction airs deep-seated frustration over discovery gridlock

It has been said that a true test of one’s character is what he does when no one’s watching. That standard, according to some, applies with equal rigor to practitioners conducting discovery. “[T]he obligation of a lawyer is to act as if you are acting in court with a judicial officer present,” Christopher Duggan, representing the American […]

Can you really compete in TREC retroactively?

This post was originally published at It is republished here with the permission of its author, Bill Dimm.  I recently encountered a marketing piece where a vendor claimed that their tests showed their predictive coding software demonstrated favorable performance compared to the software tested in the 2009 TREC Legal Track for Topic 207 (finding […]