Open The Pod Bay Door HAL …. Tom O “Interviews” a Google Notebook LM Robot.

February 12, 2025

Now most of you who know me are aware that I am not on the AI bandwagon. But Google NotebookLM is a fascinating AI tool that caused me to say “wow” for the first time in years about a new technology. Why? Because it does something innovative that, although early in the development cycle, shows great promise in working with electronic documents.

Sure it summarizes and analyzes and allows you to make queries and make notes to those summaries. But it also generates an audio “discussion” about the material you summarize, with two AI “moderators” AND it allows you to interrupt the moderators and ask questions.

How are the answers you ask? Well as deep as the source material being summarized and surprisingly articulate.

So to that end, I pointed it to Ralph Losey’s latest blog post comparing ChatGPT to Gemini and DeepSeek. https://e-discoveryteam.com/2025/02/12/breaking-new-ground-evaluating-the-top-ai-reasoning-models-of-2025/ You can find the full audio recording of the summary below and a short recording of my interaction with the “moderators” (who Ralph calls the “Gemini Twins”)on the EDicovery Channel at https://youtu.be/WMTOCteLaX8 .

Audio Overview of Ralph Losey Blog Post

Drawbacks? Sure. Data size in the free version is small and data types do not include any documents that come from Redmond. I’m shocked. Once you’re in the Interactive Mode window you need to stay there … returning to the prior window will restart the recording. Name pronunciations are iffy. It mispronounced Ralph’s last name, promised to correct it when I pointed out the error then did so intermittently. And the Gemini Twins are a little too bubbly and effusive in their commentary for my taste.

Still, an exciting start to what promises to be a richly featured product.


UNDERSTANDING LARGE DATA SETS

January 28, 2025

Big data, the larger, more complex data sets from multiple, often non-traditional, sources, is prevalent in eDiscovery. It has overwhelmed the management ability of traditional data processing software and driven the development of advanced computer-based solutions such as TAR and AI.

Let’s look at exactly what big data is and how we can handle it in the eDiscovery process.

  1. DATA
  1. WHAT IS LARGE?  

What exactly is “big data”?  One company that should know is Oracle, which says:

“The definition of big data is data that contains greater variety, arriving in increasing volumes and with more velocity. This is also known as the three “Vs.”

(Source: https://www.oracle.com/big-data/what-is-big-data)

Resource center Domo offers this quick and easy calculation:

“The most basic way to tell if data is big data is through how many unique entries the data has. Usually, a big dataset will have at least a million rows. A dataset might have less rows than this and still be considered big, but most have far more.”

(Source: www.domo.com/learn/article/4-ways-to-tell-if-your-data-is-big-data )

The problem here is not just size. Many programs have limits on how much data they can display or analyze. Large datasets are difficult to upload because of their size but firms need to analyze the whole dataset at once. They can’t just look at portions of the data, so they need a tool that will allow them to inspect everything at once but not take so much time in doing that they are not practical to use. .

  • DATA EXPLOSION

The origins of large data sets go back to the 1960s and ‘70s with the establishment of the first data centers and the development of the relational database. By 2005, people began to realize just how much data users generated through Facebook, YouTube, and other online services. Hadoop, an open-source framework created specifically to store and analyze big data sets, was developed that same year.

The development of open-source frameworks was essential for the growth of big data because they make big data easier to work with and cheaper to store. In the years since then, the volume of big data has skyrocketed. With the advent of the Internet of Things (IoT), more objects and devices are connected to the internet, gathering data on customer usage patterns and product performance. The emergence of machine learning has produced still more data, and the COVID-19 pandemic brought a rapid increase in global data creation since 2020, as most of the world population had to work from home and used the internet for both work and entertainment.

Users, both human and machine, are generating huge amounts of data. So just how much data is there in the world today? Some estimates suggest that the total amount of data on the internet reached 175 zettabytes in 2022.[1] Some studies show that 90% percent of the world’s data was created in the last two years and that every two years, the volume of data across the world doubles in size.

Additionally, there is a substantial amount of replicated data. By the end of this year, 2024, the unique/replicated data ratio is projected to change from 1:9 to 1:10.

(Source:Statista.com)

The world’s data volume has increased dramatically in the past twenty years for several reasons. What led to this explosion in data? First, according to Moore’s Law, digital storage becomes larger, cheaper, and faster with each successive year. Second, with the the advent of cloud databases, previous hard limits on storage size became obsolete as Internet giants such as Google and Facebook used cloud infrastructures to collect massive amounts of data. Companies around the world soon adopted similar big data tactics. And finally, billions of new users gained internet access across the globe, pushing more and more data accumulation.

What does that look like in real terms? The graphic below is instructive:

  • NEW DATA TYPES

The emergence of newer types of data beyond the traditional word processing and spreadsheet platforms common for years in litigation matters is also a fact.

Some common types of emerging data include:

  • Mobile data.
  • Messaging data. 
  • Marketing
  • Medical data. 
  • IoT data. 
  • MORE USERS

Out of the nearly 8 billion people in the world, 5.35 billion of them, or around 66% of the world’s population, have access to the internet. By Q3 of 2023, it was estimated that almost 96 percent of the global digital population used a mobile device to connect to the internet. Global users spend almost 60 percent of their online time browsing the web from their mobile phones. The most popular app activities on mobile were chatting and communicating, as well as listening to music and online banking activities.

(Source: Statista)

  1. LITIGATION
  1. HISTORY

How does all this large data fit into the historical timeline of litigation? The first “large” document case I was directly involved with was a coordinated action in Sacramento Ca in 1986. I maintained an index (no images) of 5m pages loaded into the DOS version of Summation on a Compaq 386 with a 20 MHz Intel 80386 CPU, 1 MB RAM, 16 KB ROM, two 1.2 MB 5¼-inch floppy drives and a 40 MB hard disk drive. The PC cost $7,999 

By the mi-90’s, I was working for the Texas Attorney General in their tobacco litigation.  I ran a coding shop of 3 shifts of 35-50 coders per shift using a LAN with Gravity software. We eventually housed a database of 13 million pages, again no images.

The image below is of the Minnesota tobacco archives where the AG there collected 28,455 boxes stacked in rows up to 12 high, four wide, 70 deep, containing 93 million pages of paper. It was the largest single records collection in the history of tobacco litigation.

In 2011, I began working with an iCONECT database here in New Orleans for the Plaintiffs in the BP case. We had 1 billion pages of emails, word processing documents, spreadsheets, proprietary data applications, and instrumentation reports. The documents were in a private cloud which was accessed by more than 100 outside law firms and their experts representing more than 116,000 individual plaintiffs and at any given point in time, 300 reviewers from over 90 law firms representing various case teams as well as several state attorneys were accessing the database. 

Today there is a Relativity database in the Jan 6 insurrection cases with files for both the US Attorney data and over half of the 1200+ defendants, all of which exceeds 10TB. It contains predominantly emails, texts, video and audio files taken from seized cell phones which were then produced to the defense.

  • STRATEGIES

So, what strategies can we use to handle all this data?  And do any of the ones we used in 1986 still apply?

  1. The Z Factor

The most important is not technological at all. It is what Bruce Markowitz. The SVP at Evolver Legal Services, likes to call “the Z factor”.  We all know about the X factor, the great unknown, but the Z factor is one which is often unarticulated. (https://www.youtube.com/watch?v=k1AEFTdVzy0)

In a nutshell it’s the question I always ask my clients when we start a project. “What is you want to do?”  Bruce says, “what is the end result you need?”.  Take the answer to that question and build your workflow around it.

  • Map Your Data.

I’ve said it over and over, many times. Get with the IT staff and generate a data map. In the old days it was easy …you simply asked where the warehouse with the boxes was. Now you need knowledgeable IT staff to show you the way. You’re Lewis and Clark, they’re Sacajawea. You’re not going to get where you need to go without them.

Once you have a data map, you can decide:

  • What data might potentially be relevant;
  • Where that data is located;
  • Who is in charge of managing that data; and
  • How to make sure it is preserved.
  • Litigation Holds

This is not a one and done. You need a comprehensive hold that is managed by someone who crafts a comprehensive hold and follows up on it periodically.

  • Analytics

Not something we had in the paper days. The sooner you can use data analytics to quantify the key issues of potential litigation, the better. A good analytics assessment, even if just with a significant sampling of data, can be crucial in developing a case strategy.

  • Standardization

Again, not a factor in the early days of large data cases as we struggled to work with paper documents and early formats of electronic data which each required their own OS or specialty database.

Now a key feature of ESI processing is to put all the data in a common format for review.  Agreeing on this format can be problematic however so remember that Rule 26(b)(1) of the Federal Rules of Civil Procedure, requires the parties to conduct a pre-trial meeting to agree on a proposed discovery plan which should include this component.

Although including determining their approach to eDiscovery. To make a strong case for favorable proportionality, parties must understand their data, build strong collaboration across business units, and utilize eDiscovery software to enhance collaboration and streamline the process.

  • Data Exchange Protocol

Although not required by Rule 26(b)(1), it is encouraged by many observers including the Sedona Conference and the EDRM.  Having a specific agreement on data exchange makes standardization easier and more efficient.

III CONCLUSION

My current favorite tool, Nextpoint, takes a proactive approach to large data sets by using two of the tools I mention above, Mapping and Analytics, in a process they call Early Data Assessment. EDA allows legal teams to sift through a mountain of electronic data to find potential evidence, thus reducing the data size and providing valuable insights which allow informed decisions for a more productive document review.


[1] A zettabyte is equal to 1,000 exabytes, or 1 trillion gigabytes.

  1. Megabytes = 1 Gigabyte

1000 Gigabytes = 1 Terabyte.
1000 Terabytes = 1 Petabyte.
1000 Petabytes = 1 Exabyte.
1000 Exabytes = 1 Zettabyte.


Is AI The Fight Club of Legal Technology ?

May 15, 2024

Artificial Intelligence has become the biggest buzz word in legal technology since, well the last biggest buzzword.  ECA, TAR, Blockchain, Analytics, Big Data, Collaboration, Disruption, Innovation.

Every 6 months we have a new “big thing” and right now it’s AI.

Articles are constantly harping on the rise of the machines that AI portends. Gartner includes AI on their list of the top 10 strategic technology trends of 2019 and even estimates that 80% of emerging technologies will be built on a foundation of artificial intelligence by 2021. In a Law Technology Today article, Andrew Ng, Co-Founder of Coursera and Adjunct Professor of Computer Science at Stanford University, says AI is the new electricity. “Just as electricity transformed almost everything 100 years ago,” he explains, “today I actually have a hard time thinking of an industry that I don’t think AI will transform in the next several years.”

https://www.lawtechnologytoday.org/2019/04/artificial-intelligence-will-change-e-discovery-in-the-next-three-years

And in that same article, Ajith Samuel technologist and co-founder of eDiscovery company Exterro, says that:

  1. Using AI will become “frictionless,” meaning that it will be ever more seamlessly integrated into the e-discovery process.
  2. AI will move out of the review phase, earlier in the EDRM, helping legal teams get to the facts of the matter faster, cheaper, and smarter than ever before.

And let’s not forget the 2015 survey by Altman Weil of 320 firms with at least 50 lawyers on staff which found that 35 percent of the leaders at those firms (responding anonymously) believed some form of AI would replace first-year associates in the coming decade. 20 percent of those same respondents said second- and third-year attorneys could also be replaced by technology over the same period and half said that paralegals could be killed off by computers. (See graphic below)

But if we are to believe the latest ILTA survey, that simply isn’t happening. The annual survey of the International Legal Technology Association was released on Nov 5 2020 and reported answers from 537 firms, representing more than 116,000 attorneys and 240,000 total users.

With regards to IA, it finds that just over 50% of respondents are not “…presently pursuing any IA option” and only 25% are actively researching an AI option. Respondents with active pilot projects or study groups were 7% and 4% respectively with only 10% reporting an active AI tool deployment.  See graphic below

https://www.iltanet.org/resources/publications/surveys/2019ts?ssopc=1

So what is going on here? We hear lots of talk about AI but not much actual usage. Part of the problem is, I believe, definitional and actual definitions of AI are in short supply.

One recent article broke out AI into 6 categories

  • Due diligence – Litigators perform due diligence with the help of AI tools to uncover background information. We’ve decided to include contract review, legal research and electronic discovery in this section.
  • Prediction technology – An AI software generates results that forecast litigation outcome.
  • Legal analytics – Lawyers can use data points from past case law, win/loss rates and a judge’s history to be used for trends and patterns.
  • Document automation – Law firms use software templates to create filled out documents based on data input.
  • Intellectual property – AI tools guide lawyers in analyzing large IP portfolios and drawing insights from the content.
  • Electronic billing – Lawyers’ billable hours are computed automatically.

And an actual standard (or standards) for AI has been slow to develop with the first just recently published by the Organization for Economic Co-operation and Development (OECD) which adopted and published its “Principles on AI”   on the Law and AI blog.  http://www.lawandai.com/ 

But in all that discussion, where are the AI use propositions for eDiscovery? Well the problem there is that eDiscovery vendors are traditionally close mouthed about their systems. And since a primary feature of AI as mentioned by Ajith Samuel above is its “frictionless” adoption, then AI implementation is hidden by both design and practice. Legal technology has become more Fight Club than computer lab and AI has become the worst example of that proposition.

I’ve written before that all this emphasis on new technology reminds me of my old friend, the late Browning Marean. He was a great fan of the writings of Ray Kurzweil, the technologist and futurist who wrote The Age of The Intelligent Machine. Browning’s favorite Kurzweil was The Singularity Is Near: When Humans Transcend Biology, which posited that technological advances would irreversibly transform people as they augment their minds and bodies with genetic alterations, nanotechnology, and artificial intelligence.

I however am more mindful of another tenet of the Singularity, that exponential increase in technologies will lead to a point where progress is so rapid it outstrips humans’ ability to comprehend it. To me we are losing sight of the proposition that people are slow and computers fast but people are smart and computers are dumb.

And in fact, some of today’s greatest minds in technology fell the same way Stephen Hawking has stated, in an op-ed which appeared in The Independent in 2014, “Success in creating AI would be the biggest event in human history. Unfortunately, it might also be the last, unless we learn how to avoid the risks..” His fear? As posted in a separate interview  with BBC, it was simply stated: “humans, limited by slow biological evolution, couldn’t compete and would be superseded by A.I.”

Hawking recently joined Elon Musk, Steve Wozniak, and hundreds of others in issuing a letter unveiled at the International Joint Conference Buenos Aires, Argentina warning that artificial intelligence can potentially be more dangerous than nuclear weapons. Even Bill Gates has expressed concerns and during a Q&A session on Reddit in January 2015, said “I am in the camp that is concerned about super intelligence. First, the machines will do a lot of jobs for us and not be super intelligent. That should be positive if we manage it well. A few decades after that though the intelligence is strong enough to be a concern. I agree with Elon Musk and some others on this and don’t understand why some people are not more concerned.”

Sound far-fetched? Well then, consider it from our perspective as attorneys. What is the ethical dilemma of bestowing legal responsibilities on robots? Does not all this talk of AI undermine our ethical duties to manage our client’s matters if we don’t really understand how these programs work?

As far back as 2013, Peter Geraghty (Director of the ETHICSearch, ABA Center for Professional Responsibility) and Susan J. Michmerhuizen (ETHICSearch Research Counsel) wrote an article for Your ABA Enews called Duty to Supervise Nonlawyers: Ignorance is Not Bliss. Although the article focused on issues with paralegals and support staff, I would suggest that computers also qualify as non- lawyers and the concerns mentioned in the article should apply to them and the technical experts who use them as well

This issue arises constantly when vendors run computer searches of documents and then produce directly to opposing counsel. The non-supervised release of privileged material can be an enormous problem for a firm, so much so that Geraghty and Michmerhuizen noted an excerpt from Comment [3] to Rule 5.3 which states:

… Nonlawyers Outside the Firm

[3]A lawyer may use nonlawyers outside the firm to assist the lawyer in rendering legal services to the client. Examples include the retention of an investigative or paraprofessional service, hiring a document management company to create and maintain a database for complex litigation, sending client documents to a third party for printing or scanning, and using an Internet-based service to store client information. When using such services outside the firm, a lawyer must make reasonable efforts to ensure that the services are provided in a manner that is compatible with the lawyer’s professional obligations.

Keep this in mind when retaining a technical expert or using software to search and produce. Do you really understand what is going on? How much work being done by computers are you actively supervising in a knowledgeable manner? In these days of a duty of technical competence, attorneys cannot simply delegate to others, even their clients, the responsibility of understanding technology. I would suggest that blindly relying on AI or other computer intelligence to make decisions does not rise to that necessary level of understanding.

Always remember that technology is a tool and humans use tools not vice versa. The ultimate decision-making about what tool to use and how to use it resides with you, the attorney. As I have said before, we need to keep the attorney in AI.                                                                                                        

It’s not enough to be aware of AI, we have to understand AI. Or, as the great technologist Elvis Aaron Presley once said, “A little less conversation, a little more action please.”


ELECTRONIC DISCOVERY IN CRIMINAL CASES

July 15, 2022

ORIGINAL STANDARD

Discussions about electronic discovery have traditionally focused on civil matters because civil cases have historically been more document intensive and thus that discovery process has been more richly nuanced. In addition, when criminal matters do have relevant documents, the discovery standard has been the so-called Brady rule, named after the landmark Supreme Court ruling Brady v. Maryland (373 U.S. 83 (1963) which held that prosecutors are only required to share evidence deemed exculpatory of the defendant.

In 2012, eDiscovery expert Craig Ball wrote in a column called Thoughtful Guidelines for E-Discovery in Criminal Cases “… apart from meeting Brady obligations, I think most lawyers regard criminal law as an area where there is no discovery, let alone this new-fangled e-discovery.More recently, a 2019 article in the New York Times mentioned that defense lawyers often called the Brady approach the “blindfold” law because it kept them in the dark and forced them into plea bargains without any knowledge of precisely what documents the prosecutors had in their possession. 

But that standard is changing.

CHANGING STANDARD

In 2005, Judge Marcia Pechman of the Western District of Washington had just concluded a white-collar criminal case brought by the government against Kevin Lawrence and his company, Znetix.  That case had nearly 1.5 million scanned electronic documents, at the time an extremely high document volume which caused both logistical and budgetary problems for the Court. 

Judge Pechman decided to convene a group of attorneys from the U. S. Defenders Office and the US Attorney in Seattle to discuss more efficient and cost-effective ways to deal with electronic documents in large cases.  This group included Russ Aoki, then a Criminal Justice Act (CJA) Panel attorney and now Coordinating Defense Attorney in complex matters for the Defenders, who had represented Mr. Lawrence.

I was privileged to be the only non-case attorney in the group, which went on to create a set of best practices policies for large document cases and wiretap surveillance evidence. Those policies were in effect in the Seattle federal court as a local rule for many years before other similar groups began meeting around the country. The eventual result of those meetings was a protocol released in 2012 by a Joint Technology Working Group of federal criminal practitioners created by the Director of the Administrative Office of the United States Courts (the supervising agency of the U.S. Defenders Office) and the U.S. Attorney General.

Then last year came the culmination of several years of meetings and discussions on this subject in which Atty Aoki was actively involved: an actual rule change.

Effective Dec 1, 2019, Federal Rule of Evidence 16.1 Pretrial Discovery Conference; Request for Court Action was changed to read:

(a) Discovery Conference. No later than 14 days after the arraignment, the attorney for the government and the defendant’s attorney must confer and try to agree on a timetable and procedures for pretrial disclosure under Rule 16.

(b) Request for Court Action. After the discovery conference, one or both parties may ask the court to determine or modify the time, place, manner, or other aspects of disclosure to facilitate preparation for trial.

Additionally, the note to the new rule references the 2012 Joint ESI Protocol, the first time that protocol has been referenced in a rule.  At roughly the same time, the Western District of Washington local criminal rules were amended to also reference the Joint Protocol and specifically use the Meet & Confer task checklist from that protocol. This was the first time either the ESI Protocol is referenced and the checklist used in a local rule. As Attorney Aoki told me, “We know because we looked at every local rule in all 94 jurisdictions.”

At the same time, many states are beginning to undertake similar changes to criminal discovery rules. Foremost in that movement has been New York which passed changes which went into effect this month requiring all prosecutorial material to be shared early in the case. For more on the New York changes see this article.

DRIVERS OF CHANGE

What is driving these changes? More than 90 percent of documents created today are generated in electronic format, so data size is exploding. At the same time, the diversity of data types continues to increase as 88 percent of the US population uses the Internet every day and 91 percent of adults use social media regularly.

Social media, mobile device content and cloud storage content are all now routinely part of the definition of relevant ESI.  See Social Media Evidence in Criminal Proceedings: An Uncertain Frontier from Georgetown Law here as well as a recent article in California Lawyer by Atty. Robert Hill, an associate at Eisner Gorin LLP, a boutique criminal defense firm in Los Angeles.

And even more exotic types of ESI loom in the future of litigation.  In a November 2019 article in Law.com, Judge Andrew Peck (Ret.), of counsel at DLA Piper, noted that  “There are already news reports of Fitbits and pacemakers being looked at in criminal cases to show that the digital data contradicts the defendant’s story of what happened.”  The same article mentions a conversation with Andrew Hessel, president of Humane Genomics, a seed stage company that makes cancer-fighting viruses, in which he discusses data definition of genomic material and the fact that viruses can now be defined so precisely as to allow printing of their data “maps”.

EXCEPTION TO NEW STANDARD

I should note, however, the exception that courts have carved out for material on cell phones. In Riley v. California, 134 S. Ct. 2473 (2014), the US Supreme Court unanimously ruled that police may not search the cell phones of criminal suspects upon arrest without a warrant. In Carpenter v U.S, No. 16-402, 585 U.S. (2018), the court held in a 5-4 decision that accessing historical records containing the physical locations of cellphones without a search warrant violates the Fourth Amendment to the United States Constitution.

Both opinions were authored by Chief Justice Roberts, and both hold that that smartphones and other electronic devices are not in the same category as wallets, briefcases, and vehicles which are subject to a limited initial examination upon arrest. Rather, said Chief Justice Roberts in the Riley opinion, cell phones are “now such a pervasive and insistent part of daily life that the proverbial visitor from Mars might conclude they were an important feature of human anatomy.” 

He went on to note that cellphones “are based on technology nearly inconceivable just a few decades ago” when the Court had upheld the search of an arrestee’s pack of cigarettes. Today, he wrote, citizens today have a reasonable expectation of privacy for information on their cell phones noting “Our answer to the question of what police must do before searching a cell phone seized incident to an arrest is accordingly simple — get a warrant.”

Another exception to these developing standards is seizures at the US border.  Entry into the US has long been a separate category as many observers argue the US Constitution cannot apply to non-citizens who are still outside the US.  That discussion is beyond the scope of this article but a good starting point is a blog post I did on the subject for CloudNine.  

CONCLUSION

In addition to pointing out the clear differences between civil and criminal discovery standards, the secondary point of this article is to note that while criminal defense attorneys have a duty to preserve and produce electronically stored information (ESI) just as their civil counterparts do, that most state and federal criminal discovery is statutory, or rule-based.  But constitutional concepts apply in both arenas in order to ensure a fair trial and due process, including the right against self-incrimination and against unreasonable searches and seizures.

For a very good overview of all the issues involved in criminal eDiscovery practice can be found in Criminal Ediscovery: A Pocket Guide for Judges, a 2015 publication of the  Federal Judicial Center. It focuses on a number of issues and in greater detail than we can cover in this post and is an excellent resource. It is authored by Sean Broderick, National Litigation Support Administrator, Administrative Office of the U.S. Courts, Defender Services Office; Donna Lee Elm, Federal Defender Middle District of Florida; Andrew Goldsmith, Associate Deputy Attorney General & National Criminal Discovery Coordinator U.S. Department of Justice; John Haried, Co-Chair, eDiscovery Working Group — EOUSA U.S. Department of Justice and Kirian Raj, Senior Counsel to the Deputy Attorney General U.S. Department of Justice. I highly recommend it.


CAN WE KEEP THE ATTORNEY IN AI OR IS ALL THE HYPE JUST FIGHT CLUB 2?

July 7, 2022

Great post this week by Doug Austin on his eDiscovery Today blog. Called Setting Realistic Expectations About AI in eDiscovery: eDiscovery Best Practices, it covered a new article in ILTA’s Peer to Peer Magazine entitle The Humans Stay in the Picture: 4 Realities of AI in Modern eDiscovery. The author of that article, Dr. Gina Taranto of ProSearch, made four keys points about AI but one in particlualr caught my eye.

Her third bullet poitn was called “Humans Stay In the Picture” and basically said that no matter how sophisticated the technology, you still need planning, training, QC and analysis and that “Humans are required for all of that.”

That brought to my mind an article I wrote in 2020 called “IS AI THE FIGHT CLUB OF LEGAL TECHNOLOGY?” I’ve reposted it below and even though some of the surveys and stats are out of date, the emphasis is still the same as what Dr. Taranto wrote. … let’s keep the attorney in AI.


Artificial Intelligence has become the biggest buzz word in legal technology since, well the last biggest buzzword. ECA, TAR, Blockchain, Analytics, Big Data, Collaboration, Disruption, Innovation.

Every 6 months we have a new “big thing” and right now it’s AI. Articles are constantly harping on the rise of the machines that AI portends. Gartner includes AI on their list of the top 10 strategic
technology trends of 2019 and even estimates that 80% of emerging technologies will be built on a foundation of artificial intelligence by

In a Law Technology Today article , Andrew Ng, Co-Founder of Coursera and Adjunct Professor of Computer Science at Stanford University, says AI is the new electricity. “Just as electricity transformed
almost everything 100 years ago,” he explains, “today I actually have a hard time thinking of an industry that I don’t think AI will transform in the next several years.”

And in that same article, Ajith Samuel technologist and co-founder of eDiscovery company Exterro, says that using AI will become “frictionless,” meaning that it will be ever more seamlessly integrated into the e-discovery process. He feels that AI will move out of the review phase, earlier in the EDRM, helping legal teams get to the facts of the matter faster, cheaper, and smarter than ever before. Ultimately AI will play an increasing role in orchestrating the e-discovery process, streamlining the process and improving efficiency.

And let’s not forget the 2015 survey by Altman Weil of 320 firms with at least 50 lawyers on staff which found that 35 percent of the leaders at those firms (responding anonymously) believed some form of AI would replace first-year associates in the coming decade. 20 percent of those same respondents said second- and third-year attorneys could also be replaced by technology over the same period and half said that paralegals could be killed off by computers. (See graphic below)

But if we are to believe the latest ILTA survey, that simply isn’t happening. The annual survey of the International Legal Technology Association was released on Nov 5, 2020, and reported answers from 537
firms, representing more than 116,000 attorneys and 240,000 total users. With regards to IA, it finds that just over 50% of respondents are not “…presently pursuing any IA option” and only 25% are actively
researching an AI option. Respondents with active pilot projects or study groups were 7% and 4% respectively with only 10% reporting an active AI tool deployment. (See graphic here )


So, what is going on here? We hear lots of talk about AI but not much actual usage. Part of the problem is, I believe, definitional and actual definitions of AI are in short supply.

One recent article broke out AI into 6 categories
 Due diligence – Litigators perform due diligence with the help of AI tools to uncover background information. We’ve decided to include contract review, legal research and electronic discovery in this section.
 Prediction technology – An AI software generates results that forecast litigation outcome.
 Legal analytics – Lawyers can use data points from past case law, win/loss rates and a judge’s history to be used for trends and patterns.
 Document automation – Law firms use software templates to create filled out documents based on data input.
 Intellectual property – AI tools guide lawyers in analyzing large IP portfolios and drawing insights from the content.
 Electronic billing – Lawyers’ billable hours are computed automatically. (See )

And an actual standard (or standards) for AI has been slow to develop with the first just recently published by the Organization for Economic Co-operation and Development (OECD) which adopted
and published its “Principles on AI” on the Law and AI blog.

But in all that discussion, where are the AI use propositions for eDiscovery? Well, the problem there is that eDiscovery vendors are traditionally close mouthed about their systems. And since a primary
feature of AI as mentioned by Ajith Samuel above is its “frictionless” adoption, then AI implementation is hidden by both design and practice.


Legal technology has become more Fight Club than computer lab and AI has become the worst example of that proposition.

I’ve written before that all this emphasis on new technology reminds me of my old friend, the late Browning Marean. He was a great fan of the writings of Ray Kurzweil, the technologist and futurist who wrote The Age of The Intelligent Machine. Browning’s favorite Kurzweil was “The Singularity Is Near: When Humans Transcend Biology”, which posited that technological advances would irreversibly transform people as they augment their minds and bodies with genetic alterations, nanotechnology, and artificial intelligence.


I however am more mindful of another tenet of the Singularity, that exponential increase in technologies will lead to a point where progress is so rapid it outstrips humans’ ability to comprehend it. To me we are
losing sight of the proposition that people are slow and computers fast, but people are smart and computers are dumb.

And in fact, some of today’s greatest minds in technology fell the same way Stephen Hawking has stated, in an op-ed which appeared in The Independent in 2014, “Success in creating AI would be the biggest
event in human history. Unfortunately, it might also be the last, unless we learn how to avoid the risks.” His fear? As posted in a separate interview with BBC, it was simply stated: “humans, limited by slow
biological evolution, couldn’t compete and would be superseded by A.I.”

Hawking recently joined Elon Musk, Steve Wozniak, and hundreds of others in issuing a letter unveiled at the International Joint Conference Buenos Aires, Argentina warning that artificial intelligence can
potentially be more dangerous than nuclear weapons. Even Bill Gates has expressed concerns and during a Q&A session on Reddit in January 2015, said “I am in the camp that is concerned about super intelligence. First, the machines will do a lot of jobs for us and not be super intelligent. That should be positive if we manage it well. A few decades after that though the intelligence is strong enough to be a concern. I agree with Elon Musk and some others on this and don’t understand why some people are not more concerned.”

Sound far-fetched? Well then, consider it from our perspective as attorneys. What is the ethical dilemma of bestowing legal responsibilities on robots? Does not all this talk of AI undermine our ethical duties to manage our client’s matters if we don’t really understand how these programs work?

As far back as 2013, Peter Geraghty (Director of the ETHICSearch, ABA Center for Professional Responsibility) and Susan J. Michmerhuizen (ETHICSearch Research Counsel) wrote an article for Your ABA Enews called “Duty to Supervise Nonlawyers: Ignorance is Not Bliss”. Although the article focused on issues with paralegals and support staff, I would suggest that computers also qualify as non-lawyers and the concerns mentioned in the article should apply to them and the technical experts who use them as well.

This issue arises constantly when vendors run computer searches of documents and then produce directly to opposing counsel. The non-supervised release of privileged material can be an enormous problem
for a firm, so much so that Geraghty and Michmerhuizen noted an excerpt from Comment [3] to Rule 5.3 which states:


… Nonlawyers Outside the Firm
[3]A lawyer may use nonlawyers outside the firm to assist the lawyer in
rendering legal services to the client. Examples include the retention of
an investigative or paraprofessional service, hiring a document
management company to create and maintain a database for complex
litigation, sending client documents to a third party for printing or
scanning, and using an Internet-based service to store client
information. When using such services outside the firm, a lawyer must
make reasonable efforts to ensure that the services are provided in a
manner that is compatible with the lawyer’s professional obligations.

Keep this in mind when retaining a technical expert or using software to search and produce. Do you really understand what is going on? How much work being done by computers are you actively supervising in a knowledgeable manner? In these days of a duty of technical competence, attorneys cannot simply delegate to others, even their clients, the responsibility of understanding technology. I would suggest that blindly relying on AI or other computer intelligence to make decisions does not rise to that necessary level of understanding.

Always remember that technology is a tool and humans use tools not vice versa. The ultimate decision-making about what tool to use and how to use it resides with you, the attorney. As I have said before, we
need to keep the attorney in AI.


It’s not enough to be aware of AI, we have to understand AI. As that great technologist Elvis Aaron Presley once said, “A little less conversation, a little more action please.”


New Relativity One UI Makes It Extremely Easy for Users to Get to Work Right Away.

June 22, 2020

RelOne AERO

Relativity held their annual Relativity Fest London event virtually in May this year and the keynote speaker, Relativity chief product officer Chris Brown, spoke about both their recently announced pay as you go pricing model and the new, currently under soft release, UI for RelOne called Aero.

RelOne has been around for four years and while changes to the interface have been going on for about 3 years, the Advanced Access Group came into play in early to mid-April and began working with this completely new UI. The group consists of 2 channel partners, two corporations, and two law firms, all of which have been instrumental in guiding the development of the UI with their enhanced feedback.

Relativity has been saying that Aero is more than just a fresh coat of paint and current users are being quoted as saying the new “ease of use and simplicity” is “… already having an impact.”

All this discussion of course piqued my interest, so I cast around, watched several of their webcasts and was finally able to arrange a personalized demo firsthand. Aero won’t be officially released until September, but it is commercially available now through providers in the Aero Advance Access program. Here’s what it looks like.

Overall, the 3 main goals of Aero set out by Relativity are:

Intuitive Workflow

Designed to get you to what you need faster, RelativityOne delivers an intuitive and streamlined platform, reducing unnecessary clicks and decisions so you have exactly what you need to accomplish your work.

Light-Speed Performance

Aero delivers what you need fast. Whether you’re flying doc-to-doc, running batch operations, or moving across the platform, everything is available when and where you need it.  Documents with large page counts load much faster now rendering on a page by page basis rather than waiting for the entire doc to render.

Easy Navigation

With logical workflows, step-by-step navigation, and simplified processes you can move through the platform without thinking about where you go next.  The modernized aesthetics have removed ~70k clicks and has minimized cursor travel to increase efficiency.

My specific impressions of the feature set are:

  1. First major change that you will see is that the tabs on the top now become categories on the left
  2. There are no default categories yet but there eventually be some based on a user profile or case defaults
  3. Document previews show in a viewer window which is a view only mode, but you can click on the DocID to bring up the full document and perform coding
  4. The full doc viewer has the complete doc listing on the left and you can jump to any document
  5. You can also pop up document history or image thumbnails as you scroll
  6. The dashboard is collapsible
  7. Ability to save searches as well as the long overdue ability for searching over mass searches feature and a mass copy/move/delete feature
  8. Filtering is available by person or by date
  9. Search enhancements include:
    • Searching for emojis or emoticons
    • Persistent highlights
    • Search for ASCII symbols
    • Highlight one term and focus search
    • Find conceptually similar in a paragraph
    • Display zero hits
  10. Direct loading of documents
    • Can drag and drop up to 100 “loose documents”
    • With large files, can look at pages that have loaded while the remainder of the loading continues.  Large docs are now in essence rendered on a page by page basis
  11. Adjust extracted text size in a manner that is similar to resizing columns in Excel
  12. Hardware agnostic
  13. Browser agnostic
  14. May have some version requirements especially with regards to the working version of Windows
  15. Field creation can occur on the fly
  16. Automatic workflows including:
    • Automated DT search updating as data is loaded
    • Analytics
    • Privilege lists
    • These will require setting a rule simultaneous to loading
  17. Predictive coding
  18. Azure
    • Hosting
    • Invariant processing

A general release was originally planned for September although it remains to be seen if the COVID-19 pandemic has any effect on that. As the graphic below shows, however, Aero is available now. Pricing is a currently said to be a flat subscription fee plus a user charge or pay as you go based on usage.

If you’d like to chat more about Aero or arrange for a demo the way I did, just contact me at toconnor@gulfltc.com.  


O365 eDiscovery Search Part 2 with Rachi Messing and Tom O’Connor

May 29, 2019

In our previous installment on Content Search we discussed basic searching and how to work with the results. This session covers some of the deeper filtering functionality that can be performed in a Review Set along with advanced search techniques and basic ECA functionality with those techniques. In addition, Rachi mentions an exciting development regarding the new O365 ability to download data directly from Facebook. in the main workloads.


Louisiana Misses The Mark On Technical Competence? Not So Fast My Friend!

May 28, 2019

Corso 1

My adopted home state of Louisiana has come under a lot of fire recently for not being up to speed on the duty of technical competence. Bob Ambrogi rather pointedly called his blog post on the subject,  A Tech Ethics Opinion that Misses the Mark and in that column referred to another post by Nicole Black who had previously written about the same subject in a Legal News column.  I know and respect both Bob and Nicole but I’m going to go all NCAA College Game Day on them here and channel my inner Lee Corso.

Basically I think their comments miss the mark on several points. First, Bob states that

“ ….  the ABA’s first opinion to address Model Rule 1.1, Comment 8 — Formal Opinion 477 issued in 2017 — makes the point repeatedly that the duty of technology competence encompasses the ability to understand how the client uses technology, what technology systems the client uses, and the client’s degree of technology sophistication.”

So first let’s recall that ABA Model Rule 1.1, Comment 8 was passed in August of 2012 and merely says  “ … a lawyer should keep abreast of changes in the law and its practice, including the benefits and risks associated with relevant technology.” As I noted at the time, it was horribly vague and it languished, rightfully so,  for several years before states started incorporating some form of it into their Rules of Professional Conduct. Taken by itself it doesn’t mention anything repeatedly.

ABA Formal Opinion 477 was issued, as Bob notes,  on May 1, 2017. This was well after the state technical competence bandwagon had started rolling but it dealt specifically with Securing Communication of Protected Client Information, the actual title of the Opinion. It does talk about technical competence quite a bit but, in my opinion, it is within the framework of protecting confidential client information, a specific technical discussion and not that of an overall duty of technical competence.

Next, both Bob and Nicole take the position that in today’s world, technical competence is a given.  Their comments seem to echo this line:

For example, a lawyer would have difficulty providing competent legal services in today’s environment without knowing how to use email or create an electronic document. 

But that quote isn’t from the Model Rules either. It’s from the ABA Commission on Ethics 20/20 Report issued in August of 2012. And in that report, Comment 8 of Formal Opinion 477 notes that the Commission said, in commenting on the proposed change to Model Rule 1.1:

The 20/20 Commission also noted that modification of Comment [6] did not change the lawyer’s substantive duty of competence: “Comment [6] already encompasses an obligation to remain aware of changes in technology that affect law practice, but the Commission concluded that making this explicit, by addition of the phrase ‘including the benefits and risks associated with relevant technology,’ would offer greater clarity in this area and emphasize the importance of technology to modern law practice. The proposed amendment, which appears in a Comment, does not impose any new obligations on lawyers. Rather, the amendment is intended to serve as a reminder to lawyers that they should remain aware of technology, including the benefits and risks associated with it, as part of a lawyer’s general ethical duty to remain competent.”  (my emphasis added)

Finally, Bob, much like everyone discussing this topic, quotes the The State Bar of California’s Formal Opinion No. 2015-193 with it’s “6 things ever lawyer needs to know about technology” emphasis. I hasten to point out that the California opinion was written in August of  2016 and specifically refers ONLY to eDiscovery matters.

Further, the last para of the opinion specifically states:

This opinion is issued by the Standing Committee on Professional Responsibility and Conduct of the State Bar of California. It is advisory only. It is not binding upon the courts, the State Bar of California, its Board of Trustees, any persons or tribunals charged with regulatory responsibilities, or any member of the State Bar

It is not a discussions of a general duty of technical competence and I am not aware that the California Bar has either offered an opinion regarding such a duty or offered a competency rule change on such. Nor has the California Supreme Court or Legislature amended their rules to reflect such a change.

So Louisiana amended its Code of Professionalism (not ts Rules of Professional Conduct) to reflect a duty of tech competence and in one para seemed to follow Model 1.1 by saying:

“I will stay informed about changes in the law, communication, and technology which affect the practice of law.”

Remember that ABA Model Rule 1/1 says

“ … a lawyer should keep abreast of changes in the law and its practice, including the benefits and risks associated with relevant technology.”

So, we left out the “benefits and risk” clause. That is a major departure form the tone of the ABA Commission on Ethics noted in their 20/20 Report comment. Really?

Sorry but I don’t think so. Long before this Code change, whenever I asked LSBA people   ( and I speak at LSBA CLE events quite a bit … I’m talking to folks at the LSBA all the time) about the duty of tech competence I was basically told, “well we feel the duty of competence INCLUDES tech competence and that’s what we tell people who ask.”  Exactly as the ABA Commission On Ethics 20/20 Report stated. It’s a reminder of an  already existing duty.

Finally, Bob generally and Nicole specifically interpret the Louisiana wording as implying a choice of whether or not to use technology. I personally think that’s splitting hairs in a way the Code doesn’t intend.  Perhaps a better word would have been “when” not “if”,  but still, do we seriously think anyone is NOT using technology? I mean they could choose not to use a phone either but I’m guessing their work would diminish. Rapidly.

Honestly, you think the Louisiana Bar is telling people they don’t have to use technology? Why not ask all the states who haven’t implemented the recommendations of Comment 18 to ABA Model Rule 1.6 why they aren’t taking ″reasonable efforts″ to protect against the inadvertent or unauthorized disclosure of, or access to, client information? Or ask California why they don’t have any opinion at all.  Or ask all the states who passed these nice vague rules if they have passed a subsequent CLE requirement for gaining that tech competence? I think we all know that the answer to that question is two.

But Louisiana has missed the mark? Come on man!!


Is ESI Production Rigged Against Plaintiffs In Mass Tort Cases? One Expert Thinks So

May 20, 2019

The eDiscovery Daily Blog has been running  a series I wrote on the challenges facing Plaintiff’s counsel in the eDiscovery arena.  The most recent chapter was posted this morning on their site here and starts with a quote from Bob Eisenberg talking about forms of production.

Bob is well known in the eDiscovery community. He is the Director, eDiscovery & Information Governance at Larson Security LLC, Program Director at the Cleveland-Marshall College of Law’s (CMLaw) eDiscovery Professional Certificate Program and eDiscovery Counsel at Marc J. Bern & Partners LLP in New York City.  

Bob has strong feelings about dealing with productions from the Defense side and feels that in many cases, especially those in the field of Mass Torts,  the litigation game is rigged against the Plaintiffs. Here’s his full exposition of that position which I think you will find interesting.

In mass tort litigation the eDiscovery playing field is tilted in favor of the defense and it seems few are complaining.

Legal disputes, more often then not, succeed or fail as a result of the practice of eDiscovery.  However, in Mass Torts the advantage has been, and continues to gravitate, relentlessly, to the defense.  The access to vital evidentiary information, critical to sustaining plaintiffs’ cases, is being whittled away by virtue of the pro forma production of data to plaintiffs that are almost entirely in static imaged-based form.

Every effort is made by Mass Tort defense firms, in virtually every litigation (no matter how varied the types of ESI; no matter how limiting to plaintiffs) to assure that records produced in discovery are delivered by defendants to plaintiffs in an imaged-based format (Tiff or PDF) with load files for searchable text and metadata; and, practically never provided (except for a tiny proportion that are considered worthless as evidence in image format) as files produced in the manner in which they have been created and stored; that is, in their ultimately most utilizable incarnation; in native form.

The difference between native and imaged-based productions is stark, contrary to some commentators that are, by practice and inclination, defense-oriented.  There is a wealth of potentially invaluable information that is lost and likely forever irretrievable by plaintiffs as a result of a non-native production. Primarily, this information takes the form of inaccessible, yet critical, metadata. Here I must recommend an extraordinarily articulate and comprehensive article exploring the superiority of native over image-based data sets, authored by the prolific and profoundly knowledgeable, eDiscovery expert, Craig Ball (see “The Case for Native Production”, Craig Ball, Esq., Practicallaw.com, October/November 2014.)

 In particular, with enormous productions, by the defense, of complex, often medically and scientifically esoteric documents, being commonplace, Mass Tort plaintiffs’ are severely hampered in the successful prosecution of their claims, by being rendered incapable, as a result of the defense’s ESI production format, of optimizing the use of Advanced Document Analytics, and Augmented Document Review (TAR) in the search and review of production sets.  These tools are commonly deployed on today’s eDiscovery platforms. It is productions in native (or near-native) format that are able to maximize the efficacy of these tools; and it is these tools that are capable of achieving a just outcome in disputes of critical importance – indeed, sometimes of life and death significance – to Mass Tort plaintiff-litigants.

The inarguable “bottom line” is that, in proving their case Mass Tort plaintiffs have become increasingly reliant upon Advanced Analytics and TAR, as the datasets produced by the defense become increasingly voluminous and complex in their make-up.

The arguments advanced by the defendant corporations, in multiple Mass Tort actions, in support of an imaged-based production, are, curiously, similar in scope and details; and, in fact, alike, as well, in lack of persuasiveness.

Below is what this writer has encountered as the most prevalent “supporting” arguments raised by defense counsel in defending, and insisting upon, the production of image-based productions in lieu of those in native format.

None can be categorized as persuasive. These enshrined obstructive orthodoxies of the defense in Mass Tort litigations cannot be permitted to continue. They render it impossible to ever truly level the playing field, and provides impermissible armament to a defense that, in most instances, from the inception of the litigation, maintains an enormous advantage in financial prowess and other resources, and, accordingly, can effectively inhibit the realization of justice and equity for thousands who have been grievously injured through no fault of their own.  The above is an especially powerful counter-argument in light of recent revelations of the failure of the Federal Drug Administration (FDA) to appropriately distinguish between its pharmaceutical manufacturing stakeholders and those that should constitute its most important constituency: The Consumers (see NY Times editorial column “When Medical Devices Go Unscreened”, NYT, Sunday, May 5, 2019.)

In Mass Torts, in an effort to avoid ESI productions in native form Defense counsel frequently advance the following seven arguments:

1.  The Pre-Existence of a Defendant’s Discovery Database containing all documents to be produced. This is often an attempt to short-circuit any meaningful meet-and-confer discussion in connection with native format production with the defense asserting that the entirety of the documents whose production is demanded have been produced previously and are housed (and sometimes alleged as stored for time immemorial) in a Legacy Discovery or Litigation database as an imaged-based collection.   Accordingly, the argument goes, requiring production now in native can only result in an undue burden and cost.

      Counter-Argument: The documents so housed in these existing databases were      produced in separate actions in which the Requester had not required production in native form, or, having done so, had, for its own reason, accepted a production that met their minimal requirements; a production that was, in effect, “good enough”.     In the matter in which you provide representation, the tools you will use and the document searches and filtering you will conduct, require nothing other than ESI in its native format.

2. It is problematic and unduly costly to redact a production collection in native form.

Counter-argument: This is a red-herring argument grounded in ancient eDiscovery           history. In fact, redaction in native can often be accomplished inexpensively and efficiently (by, for instance, altering and saving the redacted document to a new file)  and, besides, redacted records, in the overwhelming proportion of matters, constitute a  very small percentage of the production; accordingly, redaction should not be a “Tail  Wagging the Dog”, and the determining factor in selecting the production format of vital  evidentiary material.  However, since the number of records are typically small, where redaction in native is burdensome, the redacted record can be produced in image- format with load files;

3. It is unduly burdensome and costly to require an entirely new review of relevant documents now being produced in a new production format

Counter-argument – The continuing, unaltered existence of the identical records to now be produced in their native form is a given. These are, most likely, the same             discoverable records that have already been produced in dozens, hundreds or even      thousands of cases comprising the Mass Tort litigation.  Indeed, their non or altered    existence would constitute an admission of spoliation of the most egregious nature.     A native production is, in virtually all matters, not unduly burdensome for the defense to execute.  It is usually strikingly less costly than the earlier image-based  production. Native ESI is capable of being saved to and produced upon inexpensive media that has remained in common usage for decades;

4. Bates identifiers are diabolically difficult to affix to a production in native format.

Counter-argument –  An anachronistic argument that dates back to the pioneering days   of eDiscovery.  Now Bates identifiers are frequently affixed to native files by changing the name of the file to reflect the Bates number.  Also, legends indicating   confidentiality, or the existence of a Protective Order, or other category, can also be    endorsed on records in their native format.  Another common approach is to affix Bates  identifiers on documents only at the time they are needed, such as for a deposition or  trial, where at that time an image bearing the needed endorsements can be generated.  Typically, these endorsed exhibits will impact only a small proportion of the entire collection, and entail limited effort.  Again, to allow the need to “Bates Stamp”  a very small proportion of the documents in a discovery collection to inhibit the entire production in native, is highly counter-productive, to say the least.

5. Federal Rule of Civil Procedure 34 does not require a production in native (an argument, of course, advanced in the federal system, or in those state courts where eDiscovery regulations reflect a similar Rule.)

Counter-argument: This position is actually very much beside the point, and calls for  a “so what” response.  Plaintiffs are not arguing to the contrary.  Yes, this is what the Rule does not require; so, it is true, but irrelevant, if the production in native does not constitute an undue burden or cost to Respondent, and the production in  image  format can be shown to render the production to Requester one that, under the   circumstances , is not reasonably useable.

6. The image-based production has, in a de facto sense, been accepted by the court (through the vehicle of time and repetition) as the “default mode of production”; that is, courts in other jurisdictions presiding over cases concerning the same medical devices, vaccines, medicines, etc., or even the judge presiding in the present Mass Tort litigation, has routinely permitted an image-based production with load files.

Counter-argument: This is, likewise, beside the point; since your position in the  matter you are engaged in paints a portrait of a non-burdensome production in native,  starkly contrasted to an eDiscovery platform technology that can be, unnecessarily and critically crippled by a production format that the defense is coercing plaintiffs  to accept in a stark effort to assure that the use of analytical tools will not be  optimized or may be negated altogether. 

7. Productions in image format are of equal value to a native production when analytical tools are applied. Here the core argument by the defense, is that the defense’s image-based production does not undermine the use of plaintiffs’ most important eDiscovery technical tools; and, that an image-based production metadata load file, as opposed to the complete and unrestricted existence of metadata values in a document collection in its native form, are of equal efficacy when using analytics.

Counter-argument: This argument is, simply, not true.  A collection rendered without        the stripping of metadata values (native) will always bear more potentiality to reveal        vital evidence than one that relies on the fashioning of a metadata load file (image-          based). All too frequently, the load file created for an imaged-based production                  bears defects in multiple aspects, or is incapable of being adapted for use for the                analytics used by a particular eDiscovery platform.  Some knowledgeable have                  estimated that between 70% to 80% of metadata load files are marred by defects. In          actuality in order for the search and review of a static image-based production to              fully benefit from analytical tools, the extraction of metadata and its re-introduction         into a load file must be virtually perfect.  A standard rarely achieved in any discipline      

  All of this confrontation may, or may not, come down to a vigorously argued Meet-and-  Confer before the court or even a hotly advocated Motion to Compel.

The arguments rendered above by the defense in justification of its refusal to produce in native, are a cocktail of the dubious, bogus and unfounded.  A stew of junk or half-baked technical science and disingenuous advocacy, seeking to rationalize the unreasonable, while tilting that proverbial playing field as far as possible in support of  the defense in Mass Torts.

 

 

 

 


Rachi Messing & Tom O’Connor Are Back with Searching in O365 EDiscovery

May 8, 2019

This is the fifth installment of the Office 365 series with Rachi Messing, Senior Program Manager at Microsoft Legal  and Tom O’Connor of the Gulf Coast Legal Tech Center and this week Rachi and Tom discuss the basic search features of O365.  You can use Content Search to search mailboxes, public folders, SharePoint Online sites, and OneDrive for Business locations. There are no limits on the number of mailboxes and sites that you can search. There are also no limits on the number of searches that run at the same time. After you run a search, the number of content sources and an estimated number of search results are displayed in the details pane on the search page. You can preview the results, or export them to a local computer and if your organization has an Office 365 Enterprise E5 subscription, you can prepare the results for analysis using the powerful analytics features of Office 365 Advanced eDiscovery.