The E-Discovery Games: A Closer Look at Technology Assisted Document Review David D. Lewis, Ph.D., Information Retrieval Consultant Kara M. Kirkeby, Esq.,

Slides:



Advertisements
Similar presentations
1 Evaluation Rong Jin. 2 Evaluation  Evaluation is key to building effective and efficient search engines usually carried out in controlled experiments.
Advertisements

An Evaluation of MC/DC Coverage for Pair-wise Test Cases By David Anderson Software Testing Research Group (STRG)
Automated Software Testing: Test Execution and Review Amritha Muralidharan (axm16u)
Chapter 4 Design Approaches and Methods
Webinar Sponsorship Partner. Jason Velasco Jason Velasco is an electronic discovery industry veteran with more than 15 years of experience in electronic.
Technology Assisted Review: Trick or Treat? Ralph Losey, Esq., Jackson Lewis 1.
5 Vital Components of Every Custodian Interview David Meadows, PMP, Managing Director – Discovery Consulting, Kroll Ontrack Dave Canfield, EJD, Managing.
Avoiding Sanctions & Surprises The ethics of discovery Kat Meyer, Esq. President of Conquest eDiscovery, LLC.
Business Statistics for Managerial Decision
Reasoning Methodologies in Information Technology R. Weber College of Information Science & Technology Drexel University.
Technology-Assisted Review Can be More Effective and More Efficient Than Exhaustive Manual Review Gordon V. Cormack University of Waterloo
Evaluating Search Engine
INFO 624 Week 3 Retrieval System Evaluation
©2007 H5 Simultaneous Achievement of high Precision and high Recall through Socio-Technical Information Retrieval Robert S. Bauer, Teresa Jade
The Wharton School of the University of Pennsylvania OPIM 101 2/16/19981 The Information Retrieval Problem n The IR problem is very hard n Why? Many reasons,
COURSE ON PROFESSIONALISM ASOP #17 - Expert Testimony by Actuaries.
Web Search – Summer Term 2006 II. Information Retrieval (Basics Cont.) (c) Wolfgang Hürst, Albert-Ludwigs-University.
ISP 433/633 Week 6 IR Evaluation. Why Evaluate? Determine if the system is desirable Make comparative assessments.
9 C H A P T E R Transaction Processing and Enterprise Resource Planning Systems.
VENDORS, CONSULTANTS AND USERS
RANDOM SAMPLING PRACTICAL APPLICATIONS IN eDISCOVERY.
India Research Lab Auto-grouping s for Faster eDiscovery Sachindra Joshi, Danish Contractor, Kenney Ng*, Prasad M Deshpande, and Thomas Hampp* IBM.
Search and Retrieval: Relevance and Evaluation Prof. Marti Hearst SIMS 202, Lecture 20.
Get Off of My I-Cloud: Role of Technology in Construction Practice Sanjay Kurian, Esq. Trent Walton, CTO U.S. Legal Support.
TAR, CAR, Predictive Coding, Integrated Analytics What Does It All Mean? Huron Legal provides advisory and business services to assist law departments.
S/W Project Management
E-Discovery: Not Grandfather’s Litigation Finding Your Client’s Narrative.
Processing of large document collections Part 2 (Text categorization) Helena Ahonen-Myka Spring 2006.
Technology-Assisted Review Solutions Predictive Coding Language-Based Analytics SM (LBA) © 2013 RenewData.
11 C H A P T E R Artificial Intelligence and Expert Systems.
Marco Nasca Senior Director, Client Solutions TRANSFORMING DISCOVERY THROUGH DATA MANAGEMENT.
Analysis 360: Blurring the line between EDA and PC Andrea Gibson, Product Director, Kroll Ontrack March 27, 2014.
Presented by Rebecca Shwayri. Introduction to Predictive Coding and its benefits How can records managers use Predictive Coding Predictive Coding in Action.
Evaluating What’s Been Learned. Cross-Validation Foundation is a simple idea – “ holdout ” – holds out a certain amount for testing and uses rest for.
IPMA Executive Conference Value of IT September 22, 2005.
Audit Sampling: An Overview and Application to Tests of Controls
Jimmy Coleman.  The Sedona Conference  The Electronic Discovery Reference Model Project  The Federal Judicial.
The Challenge of Rule 26(f) Magistrate Judge Craig B. Shaffer July 15, 2011.
Greedy is not Enough: An Efficient Batch Mode Active Learning Algorithm Chen, Yi-wen( 陳憶文 ) Graduate Institute of Computer Science & Information Engineering.
WIRED Week 3 Syllabus Update (next week) Readings Overview - Quick Review of Last Week’s IR Models (if time) - Evaluating IR Systems - Understanding Queries.
Conducting Modern Investigations Analytics & Predictive Coding for Investigations & Regulatory Matters.
Evaluating Results of Learning Blaž Zupan
Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines  Measurement usually carried out.
Copyright © 2006 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin 8-1 Chapter Eight Audit Sampling: An Overview and Application.
Chapter 6: THE EIGHT STEP PROCESS FOCUS: This chapter provides a description of the application of customer-driven project management.
COURSE ON PROFESSIONALISM ASOP #17 - Expert Testimony by Actuaries.
Number Sense Disambiguation Stuart Moore Supervised by: Anna Korhonen (Computer Lab)‏ Sabine Buchholz (Toshiba CRL)‏
Performance Measures. Why to Conduct Performance Evaluation? 2 n Evaluation is the key to building effective & efficient IR (information retrieval) systems.
WERST – Methodology Group
Machine Learning Tutorial-2. Recall, Precision, F-measure, Accuracy Ch. 5.
Data Mining and Decision Support
Defensible Quality Control for E-Discovery Geoff Black and Albert Barsocchini.
Chapter. 3: Retrieval Evaluation 1/2/2016Dr. Almetwally Mostafa 1.
Information Retrieval Quality of a Search Engine.
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
Information Retrieval Lecture 3 Introduction to Information Retrieval (Manning et al. 2007) Chapter 8 For the MSc Computer Science Programme Dell Zhang.
Knowledge and Information Retrieval Dr Nicholas Gibbins 32/4037.
Beyond Predictive Coding – The True Power of Analytics.
Introduction to Machine Learning, its potential usage in network area,
What is Legal Analytics?
Audit Sampling: An Overview and Application
Background: The Big Data era
School of Computer Science & Engineering
Rule Induction for Classification Using
The Next Frontier in TAR: Choose Your Own Algorithm
Speakers: Ian Campbell, Claire Hass,
VENDORS, CONSULTANTS AND USERS
The Team Players and Play in a Complex Document Review Project: Past, Present and Future Ralph C. Losey, Jackson Lewis Principal and National e-Discovery.
INFORMATION RETRIEVAL TECHNIQUES BY DR. ADNAN ABID
I. Hypothetical A. Case Background: Relator, a former employee of Defendant, brought a case in federal court against a large defense contractor (Relator’s.
Presentation transcript:

The E-Discovery Games: A Closer Look at Technology Assisted Document Review David D. Lewis, Ph.D., Information Retrieval Consultant Kara M. Kirkeby, Esq., Document Review Manager, Kroll Ontrack 1

Dave Lewis, Ph.D. President, David D. Lewis Consulting Co-founder TREC Legal Track Testifying expert in Kleen Products, LLC, et al. v. Packaging Corp. of America, et al Fellow of the American Association for the Advancement of Science 75+ publications; 8 patents in: –e-discovery –information retrieval –machine learning –natural language processing –applied statistics Past research positions: University of Chicago, Bell Labs, AT&T Labs 2

Kara M. Kirkeby, Esq. Manager of Document Review Services for Kroll Ontrack Previously managed document reviews on complex matters for a large law firm Member: Minnesota State Bar Association (Civil Litigation Section), the Hennepin County Bar Association, the American Bar Association, Minnesota Women Lawyers (Communications Committee) Served as a judicial law clerk for Hon. Karen Klein, Magistrate judge of the U.S. District Court of North Dakota J.D., magna cum laude, Hamline University School of Law 3

Discussion Overview What is Technology Assisted Review (TAR)? Document Evaluation Putting TAR into Practice Conclusion 4

What is Technology Assisted Review? 5

Why Discuss Alternative Document Review Solutions? Document review is routinely the most expensive part of the discovery process. Saving time and reducing costs will result in satisfied clients. Traditional/Linear Paper-Based Document Review Online Review Technology Assisted Review 6

Why Discuss Alternative Document Review Solutions?  Conducting a traditional linear document review is not particularly efficient anymore  Focus instead on a relevance driven review process involving lawyers and technology working together 7 Traditional Document Review Inaccurate Time- Consuming Expensive

What Is Technology Assisted Review (TAR) ? Three major technologies: Supervised learning from manual coding Sampling and statistical quality control Workflow to route documents, capture manual decisions, and tie it all together in a unified process 8 recall: 85% +/- 4% precision: 75% +/- 3% Presented by Dave Lewis

Supervised Learning: The Backbone of TAR Better Documents Better Reviewing Better Machine Learning Better Documents Better Reviewing By iterating supervised learning, you target documents most likely to be relevant or on topic, creating a virtuous cycle: 9 Presented by Dave Lewis

Software learns to imitate human actions For e-discovery, this means learning of classifiers by imitating human coding of documents Any content-based sorting into classes can be imitated –Responsive vs. Non-responsive –Privileged vs. Non-privileged –Topic A vs. Topic B vs. Topic C Widely used outside e-discovery: –Spam filtering –Computational advertising –Data mining 10 Supervised Learning: The Backbone of TAR Presented by Dave Lewis

Text REtrieval Conference (“TREC”), hosted by National Institute of Standards and Technology (“NIST”) since 1992 o Evaluations open to academics and industry TREC Legal Track (since 2006) provides simulated review for responsiveness task Focus is on comparing technology assisted approaches o Not a human vs. machine bakeoff o Not a product benchmark However, results suggest advantages to technology assisted review 11 Research & Development: TREC Legal Track Presented by Dave Lewis

1.High effectiveness of TAR runs o Best T-A runs in TREC 2009 examined 0.5% to 4.1% of collection while finding an estimated 76.7% of responsive documents with 84.7% precision 2.Low effectiveness of manual review o Substantial effort needed by TREC organizers to clean up manual review to point it can be used as gold standard 3.An argument can be made (Grossman & Cormack, 2011) that 2009 data shows TAR results better than pre-cleanup manual review 12 Research & Development: TREC Legal Track Presented by Dave Lewis

Analyze What is Technology Assisted Review? 13 Train START: Select document set Identify training set Knowledgeable human reviewers train system by categorizing training set System learns from training; prioritizes documents and suggests categories Evaluate Evaluate machine suggestions END: Produce documents Presented by Dave Lewis Quality control production set Human reviewers:

SELECT Manually review documents for training o Key docs from your side or opponent o Docs found by searches on key terms o Docs prioritized for review o Random (non-QC) docs o Docs difficult for previous iteration's classifier (active learning) Effectiveness increases as training set grows various docs for training random docs for QC manual review train classifiers auto-code documents compare coding with elite coding on random sample good enough to produce? estimate effectiveness for entire set review for privilege PRODUCTION priority docs for review YESNO 14 Learning and Classification Presented by Dave Lewis

Manually review prioritized documents o Needs of case o Classifier predictions If classifier is accurate enough, trust its call on responsiveness? Privilege is more sensitive o Manually select some subsets for 100% privilege review o Employ sampling for other subsets o Classifiers can also help identify likely privileged docs SELECT various docs for training random docs for QC priority docs for review manual review train classifiers auto-code documents compare coding with elite coding on random sample good enough to produce? estimate effectiveness for entire set review for privilege PRODUCTION YESNO 15 Production Presented by Dave Lewis

Any binary classification can be summarized in a 2x2 table o Linear review, automated classifier, machine-assisted... o Responsive v. non-responsive, privileged v. non-privileged... Test on sample of n documents for which we know answer o TP + FP + FN + TN = n "Truth" YesNo Prediction Yes TP (true positives) FP (false positives) No FN (false negatives) TN (true negatives) 16 Classification Effectiveness Presented by Dave Lewis

True Negatives False Positives True Positives False Negatives Classifier Says "Yes" "Yes" is Correct All Documents 17 Classification Effectiveness Presented by Dave Lewis

Recall = TP / (TP+FN) o Proportion of interesting stuff that the classifier actually found High recall of interest to both producing and receiving party "Truth" YesNo Prediction Yes TP (true positives) FP (false positives) No FN (false negatives) TN (true negatives) 18 Classification Effectiveness

Precision = TP / (TP+FP) o Proportion of stuff found that was actually interesting High precision of particular interest to producing party: cost reduction! "Truth" YesNo Prediction Yes TP (true positives) FP (false positives) No FN (false negatives) TN (true negatives) 19 Classification Effectiveness

Seminal 1985 study by Blair & Maron Review for documents relevant to 51 requests related to BART crash Boolean queries used to select documents for review o Process iterated until reviewer satisfied 75% of responsive documents found Sampling showed recall of less than 20% B&M has been used to argue for everything from exhaustive manual review to strong AI o Real lesson is about need for sampling! 20 Research & Development: Blair & Maron Presented by Dave Lewis

Want to know effectiveness without manually reviewing everything. So: o Randomly sample the documents o Manually classify the sample o Estimate effectiveness on full set based on sample Type of estimates: o Point estimate, e.g. F1 is 0.74 o Interval estimate, e.g. F1 in [0.67,0.83] with 95% confidence Sampling is well-understood o Common in expert testimony in range of disciplines 21 Sampling and Quality Control Presented by Dave Lewis

SELECT various docs for training random docs for QC priority docs for review manual review train classifiers auto-code documents compare coding with elite coding on random sample good enough to produce? estimate effectiveness for entire set review for privilege PRODUCTION Manually review random sample for QC o Use best reviewers here Estimate recall, precision, etc. o Of auto-coding, manual review, or both combined –Estimates used in: o Deciding when finished o Tuning classifiers (and managing reviewers) o Defensibility – Auto-coding can also be used to find likely mistakes (not shown) YESNO 22 Sampling and Quality Control Presented by Dave Lewis

23 Putting TAR into Practice

Barriers to Widespread Adoption  Industry-wide concern: Is it defensible?  Concern arises from misconceptions about how the technology works in practice »Belief that technology is devoid of any human interaction or oversight »Confusing “smart” technologies with older technologies such as concept clustering or topic grouping »Limited understanding of underlying “black box” technology  Largest barrier: Uncertainty over judicial acceptance of this approach »Limited commentary from the bench in the form of a court opinion »Fear of being the judiciary’s “guinea pig” 24

Developing TAR Case Law  Da Silva Moore v. Publicis Groupe »Class-action suit: parties agreed on a protocol signed by the court »Peck ordered more seeding reviews between the parties »“Counsel no longer have to worry about being the first ‘guinea pig’ for judicial acceptance of computer-assisted review … [TAR] can now be considered judicially approved for use in appropriate cases.”  Approximately 2 weeks after Peck’s Da Silva Moore opinion, District Court Judge Andrew L. Carter granted plaintiff opportunity to submit supplemental objections »Plaintiff later sought to recuse Judge Peck from the case  Stay tuned for more…. 25

Developing TAR Case Law  Kleen Products v. Packaging Corporation of America »Defendants had completed 99% of review, Plaintiffs argue that they should use Predictive Coding and start document review over »Not clear whether Defendants did more than keyword search  Other notable points from Kleen Products »Defendants assert they were testing their keyword search queries, not just guessing –Argue they did not use Predictive Coding because it did not exist yet  Stay tuned for more…. 26

Technology Assisted Review: What It Will Not Do  Will not replace or mimic the nuanced expert judgment of experienced attorneys with advanced knowledge of the case  Will not eliminate the need to perform validation and QC steps to ensure accuracy  Will not provide a magic button that will totally automate document review as we know it today 27

Technology Assisted Review: What It Can Do  Reduce: »Time required for document review and administration »Number of documents to review; if you choose an automated categorization or prioritization function »Reliance on contract reviewers or less experienced attorneys  Leverage expertise of experienced attorneys  Increase accuracy and consistency of category decisions (vs. unaided human review)  Identify the most important documents more quickly 28

TAR Accuracy TAR must be as accurate as a traditional review Studies show that computer-aided review is as effective as a manual review (if not more so) Remember: Court standard is reasonableness, not perfection: “[T]he idea is not to make it perfect, it’s not going to be perfect. The idea is to make it significantly better than the alternative without as much cost.” -U.S. Magistrate Judge Andrew Peck in Da Silva Moore 29

30 What is Intelligent Review Technology (IRT) by Kroll Ontrack? Intelligent Prioritization Intelligent Categorization Automated Workflow Reviewing Efficiently, Defensibly & Accurately Augments the human-intensive document review process to conduct faster and cheaper discovery

1.Cut off review after prioritization of documents showed marginal return of responsive documents for specific number of days 2.Cut off review of a custodian when, based on prioritization statistics that showed only non-responsive documents remained 3.Used suggested categorizations to validate human categorizations 4.Used suggested categorizations to segregate documents as non- responsive at >75% confidence level. After sampling that set, customer found less than.5% were actually responsive (and only marginally so). Review was cut off for that set of documents 5.Used suggested categorizations to segregate categories suggested as privilege and responsive at >80% confidence. Sampled, mass categorized 6.Use suggested categorizations to mass categorize documents and move them to the QC stage, by-passing first-level review 7.Used suggested categorizations to find documents on a new issue category when review was nearing completion 31 Successes in the Field: Kroll Ontrack’s IRT

32

33 Conclusion

34 Parting Thoughts  Automated review technology helps lawyers focus on resolution – not discovery – through available metrics »Complements human review, but will not replace the need for skillful human analysis and advocacy  We are on the cusp of full-bore judicial discussion of Automated Review Technologies »Closely monitor judicial opinions for breakthroughs »Follow existing best practices for reasonableness and defensibility  Not all Technology Assisted Review solutions are created equal »Thoroughly vet the technology before adopting

35 Q & A

36