Three big questions about AI in financial services

White & Case LLP

White & Case LLP

The success of artificial intelligence (AI) algorithms hinges on the ability to gain easy access to the right kind of data in sufficient volume. Put more simply, AI depends on good data. Even Google—which is famous for the pioneering work in AI that underpins its standard-setting search-based advertising business—makes no bones about the critical role of data in AI. Peter Norvig, Google's director of research, has said: "We don’t have better algorithms, we just have more data."

Companies increasingly realize that data is critical to their success—and they are paying striking sums to acquire it. Microsoft's US$26 billion purchase of the enterprise social network LinkedIn is a prime example. But other technology companies are also seeking to acquire data-related assets, typically to acquire more than just identity-linked information from social media sources by focusing instead on vast troves of anonymized consumer data. Think, for example, of Oracle pursuing an M&A-led strategy for its Oracle Data Cloud data aggregation service, or IBM buying, within the past two years, both The Weather Company and Truven Health Analytics.

To unlock the full value of AI algorithms, companies must have access to large data sets, apply abundant data-processing power, and have the skills to interpret results strategically

Early returns for companies making such investments are promising. Still, to unlock the full value of AI algorithms, companies must have access to large data sets, apply abundant data-processing power, and have the skills to interpret results strategically. Increasingly, those three elements are in the hands of the largest technology companies, fueling their market value. According to Kleiner Perkins Internet Trends Report 2017, seven of the 10 most valuable companies in the world are technology companies, compared with just three out of 10 in 2012.

Outside of tech, other industries are struggling to determine how AI can help power their future. But many financial services institutions start from a position of comparative advantage—they have large data sets and decades of experience using analytical tools, building models and employing large teams of software developers. More recently, they have also begun to incorporate data scientists into their ranks. Well positioned to leverage AI, financial services institutions have already begun to incorporate AI in parts of their business, such as algorithmic trading.

So how can financial services institutions best integrate AI into their operations—and, in turn, accelerate and improve their yield on AI investments? What we know is that, at the outset of their AI initiatives, companies will have to address three core questions: Who owns the data that is essential to AI? Who is responsible for AI decisions and actions? And what are AI’s implications for cybersecurity? Answering these questions will often require grappling with complex issues that affect multiple stakeholders, sometimes with competing interests. And because the space is evolving rapidly, companies’ answers are likely to evolve as well. 

Who owns the data that is essential to AI?

Who owns this massive volume of data that companies are seeking to mine for insight?  The answer to this question is an increasingly complex one, particularly given the explosion of unstructured data, the rise of partnerships that expand access to data, and the increasing ability to identify individuals when data sets are combined. Consumers, technology companies, third-party data providers, regulators and the panoply of institutions looking to realize AI's potential—all are stakeholders with complicated and competing interests in questions about who owns the data and for what purpose, what it's worth, and how it can rightfully be used.

Consumer privacy and consent

Consumers have a stake in which of their data is used, where and how it is used, and toward what ends it is used. But precisely what rights do they have? Legally, the answer may depend on where consumers live and what they agreed to when they gave access to their data. In the context of a global economy (and global data sets), organizations may often have to conduct complicated cross-border analyses and risk assessments, and resolve complex jurisdictional conflicts to determine the right way forward regarding consumer rights.

In response to this high level of complexity and competing demands, a variety of initiatives have been undertaken. The World Economic Forum, for example, launched a project, called Rethinking Personal Data, that maps out what it describes as the "personal data ecosystem." The project calls for greater transparency into how data-driven polices are communicated, increased accountability for companies that harness data, and greater empowerment for individuals seeking to control how that personal data is used. But concrete results have yet to materialize.

Compounding the challenge, companies have begun to invest large sums to access consumer data, frequently through M&A. But, absent careful consideration, multiple deals, that may be layered on multiple and varied data-sharing arrangements, can sometimes muddy data provenance, increasing the risk that data will be duplicated and commingled. Separating commingled data can be extremely costly and difficult, if not impossible, because the data are often anonymized and aggregated into a pool.

This challenge may be even greater when it comes to unstructured data—such as text from emails, documents, social media postings, call-center transcripts and other sources. Unstructured data can be stored in relational databases alongside structured, transactional data, but extracting or associating it with structured data records can be difficult, not only as a matter of law but also technically. AI can help with this, providing efficient ways to organize troves of unstructured data by adding and ordering metadata, essentially transforming them into structured formats.

Companies are working to ensure their data governance programs account for the explosion of unstructured data, which has become the feedstock for AI

AI and data governance

Financial services institutions have invested heavily over the years in data governance programs. They do so for a variety of reasons, including ensuring regulatory compliance; enabling data-driven decision-making; improving customer service; bolstering risk management; and addressing issues related to mergers, acquisitions and divestitures. But approaches to data governance can vary considerably from institution to institution. Some focus on data quality as part of data governance, whereas others view governance mainly as a set of management policies that are overseen by various councils. Still others consolidate everything, incorporating IT and related policies into one overarching data governance program.

But not all data in financial institutions is equal. Much of the data contains customers' personally identifiable information (PII), such as names, addresses, phone and social security numbers, information about their behavior regarding their personal finances, contractually protected data (e.g., information under a non-disclosure agreement), negotiated instruments and material nonpublic information.

Given these competing and complicated types and uses of data, financial services institutions are prioritizing the development of a cogent approach to data governance. Companies are working to ensure their data governance programs account for the explosion of unstructured data, which has become the feedstock for AI. This work includes identifying and mapping PII in structured and unstructured data, de-identifying and anonymizing customer information where necessary, and applying encryption and pseudonomization (a procedure by which the most identifying fields are replaced by one or more artificial identifiers) to comply with data privacy requirements.

Companies are reckoning with how AI affects their policies and processes. Banks, investment institutions and insurance companies are developing (or recognize the need to develop) policies that spell out AI-driven approaches to a variety of functions such as credit scoring and compliance, which increases processing speeds, minimizes labor and other resource costs, reduces human error and improves customer service.

Financial services institutions recognize that they can mine customer data to unearth opportunities for developing desirable, tailored products, such as providing customers with dashboards that aggregate and analyze data from disparate financial accounts and offering accounting services. Although banks enjoy a higher degree of customer trust than other industries, trust can erode if they do not protect customer data or provide third parties with unauthorized access to it.

Business partners

Partnering with other entities to collect, process and store customer data can further complicate questions of data ownership. Leasing data is also complicated. Data agreements are difficult to construct and interpret, and strict limitations on data usage can be difficult to enforce. Blockchain technologies may offer solutions by establishing data marketplaces that create auditable trails of provenance and ownership, logging appropriate use and attempted misuse.

The use of a third-party business partner does not relieve a financial institution of its obligations to comply with laws and regulations. A financial services institution retains responsibility for compliance, even when it outsources data processing and other functions.

Consider autonomous vehicles. Who owns the data about the car—the owner, the manufacturer or the software provider (if different from the manufacturer)? And if the same data is being used by multiple parties or algorithms simultaneously, how effectively can an individual ensure the data is used only for agreed-upon purposes?

Similar issues apply in the payments sphere. Multiple stakeholders—including consumers, merchants, payment providers and processors, and banks—may have interests in the data that is generated through payment transactions. When payments are made on mobile apps, the list of stakeholders might also include app developers, OS providers and wireless carriers. Sorting through who has which rights to what data—even when the customer has provided consent to its use—can be extremely difficult.

The answer hinges on data governance. Increasingly, financial services institutions must shift the emphasis from merely managing the life cycle of data to governing its use in context, including safeguarding data to ensure it is being used for its intended purpose—with the consent of the data subject and in compliance with global privacy laws, intellectual property laws, or other applicable laws, rules or regulations.

AI and big data have expanded the notion of business partners and data services. Data brokers have long traded personal and other data—such as age, gender and income data—in transactions largely comprising static lists and databases. Today any product vendor can potentially collect and provide access to data generated through use of its products.

The proliferation of the Internet of Things has only increased the number and type of products that generate data. Moreover, organizations increasingly use unstructured data that is available via real-time flows. It can be difficult to value data in real time, not to mention ensure that it is fit for its intended purpose. Real-time data use poses risks related to data provenance sovereignty as well as legal and regulatory compliance, whether an organization leases it or acquires it through M&A.

Valuing data in M&A

Mergers and acquisitions are commonplace in the financial services sector, and companies have become accustomed to the HR and IT issues involved. But one area that is far from agreed upon is how to value data in M&A deals—including the potential future value of data.

Consider a hypothetical situation in which Bank A buys Bank B, particularly because it wants to acquire B's data. Bank A has its own trove of valuable data, as well as a deep understanding of AI and a staff that is able to combine the data sets to increase their aggregate value far beyond the sum of their parts. How can Bank B arrive at an appropriate valuation that ensures it gets a fair price for its data, given that the value of its data depends on factors that it may not understand or even be aware of? These and related data monetization, privacy and cybersecurity considerations are leading to an increase in the level of expertise, time and attention required during transactional due diligence to properly ascertain the value of (or liability embedded in) the data that is part of a transaction.

M&A deals also expose institutions to risks when acquirers do not fully comprehend what is in the data they are acquiring. During the 2007-2008 financial crisis, many data sets changed hands very quickly, and participants often did not have the time to sufficiently assess the risk the data may have presented. The situation is even more complex today given the emergence of powerful AI and deep learning algorithms that were not in use at the time of the financial crisis. In the future, such technologies will be even more common—and they will also be an important part of drawing up detailed risk profiles for data belonging to potential M&A targets.

Because AI algorithms have an increasing ability to act independently, it may become difficult to assign responsibility to humans for decisions or actions AI takes

Who's responsible for AI decisions and actions?

As AI becomes more sophisticated, companies will increasingly automate complex processes, including those that involve decision-making that requires judgment informed by experience. AI that incorporates machine learning and deep learning can leverage what it learns in order to write, in essence, new algorithms autonomously—and these new algorithms will affect the future decision-making and actions of this evolved AI.

Understanding what an AI program has learned and why it does what it does can be incredibly difficult. And because AI algorithms have an increasing ability to act independently, it may become difficult to assign responsibility to humans for decisions or actions AI takes. In some cases, existing laws and regulations clearly apply, but the grey areas are increasing, and anticipating the types of complications that will emerge as the scope and scale of AI-driven automation expands can be very difficult.


In financial services—particularly in consumer finance—decisions about individuals' creditworthiness have traditionally been made using a transparent process with defined rules and relatively limited data sets. This transparency, however, may not always be achievable when AI drives big data. As AI is incorporated into financial operations, institutions run the risk that their algorithms may inadvertently make biased decisions or take actions that discriminate against protected classes of people—leaving financial institutions accountable, even if the alleged discrimination is unintentional.

There are three primary sources of bias in the AI process: data, training and programming. Bias can inhere in the data used to train machine-learning models—for example, in the assumptions used to create data sets for such models. It can arise when the size and scope of a data set used to train models is insufficiently large, and thus give rise to conclusions that are misleading or erroneous. Data sampling can compound this problem, as a given sample might not be representative of the data set as a whole. Bias outcomes can result from algorithms that input data reflecting bias and perpetuate those biases in the output.

The training of algorithms using what is known as "supervised machine learning" is also susceptible to bias. Even expert human trainers can potentially skew an algorithm by providing inputs and interpreting algorithmic results in ways that unintentionally reflect personal bias.

Most data scientists and developers go to great lengths to create objective algorithms. But even the most careful expert trainers are subject to contextual influences—cultural, educational, geographic—that can affect the assumptions that inform machine-learning code and skew algorithmic results. Beyond this, as algorithms learn on their own, they can draw facially neutral connections (i.e., connections that do not appear to be discriminatory on their face) that, nevertheless, adversely impact protected classes.

This kind of unconscious bias poses an important challenge, because bias does not need to be intentional to land financial institutions in hot water. Courts and regulators are expanding their use of the "disparate impact" theory of liability, which focuses more on discriminatory effects of credit decisions and policies than on a financial institution's rationale or motivation. Under disparate impact theory, if bias is seen to result in an alleged discriminatory act, there is no need to show that discrimination was intentional to establish liability—only that the discrimination occurred.

AI-powered applications should include mechanisms to safeguard customer and consumer information and provide a process for consumers to protest an algorithmic denial of financial services


Many hope that AI—particularly deep learning—will enable automatic and algorithmic trading to evolve beyond a focus on containing costs to driving profits by informing trading and financial planning. For example, AI can facilitate the use of big data to inform decisions on opening customer accounts and extending credit to customers. But it may be difficult for financial institutions to accept the risks of AI if algorithms are not accountable for results.

First, unaccountable algorithms can undermine US and global data privacy principles and requirements, leading to fines, penalties and other regulatory actions. The Federal Trade Commission (FTC) Fair Information Practice Principles (FIPPs) provide guidelines for entities that collect and use personal information. The FIPPs include safeguards to ensure that information processing in manual and automated systems is fair and provides adequate privacy protection.

In the European Union (EU), under the Charter of Fundamental Rights of the European Union and the Treaty on the Functioning of the European Union, persons have a fundamental right to protection regarding the processing of their personal data. Under the EU’s General Data Protection Regulations (GDPR), personal data must be processed in a fair and transparent manner, often with the consent of the data subject. Among other rights, EU data subjects may obtain confirmation of whether a controller is processing their personal information, the recipients (or categories of recipient) with whom their data is being shared, and the existence and significance of any automated decision-making.

Second, it will be perilous for financial institutions to negatively affect consumer access to financial services via algorithms without justification or recourse to review a negative outcome. This will warrant the attention of the FTC, the European Commission and other consumer watchdogs that will investigate suspect algorithms, training sets and the underlying data.

Third, the costs of building and maintaining AI-powered applications should include mechanisms to safeguard customer and consumer information and provide a process for consumers to protest an algorithmic denial of financial services. If consumers have no recourse to a negative outcome of an AI algorithm, they may not give their consent to apply their personal and financial data to an algorithm. Financial institutions must ensure that all data used by algorithms to determine customer service is accurate and valid and that there are appropriate technical controls, including encryption or pseudonymization, to safeguard personal and private information while it is being processed and stored.

Machine-learning technology is powered by complex mathematics. Deep-learning technology seeks to create overlapping networks of algorithms that are analogous to a brain's neural networks. Researchers are a long way from knowing which data is processed where in an AI neural network. That said, work is underway to test neural networks and determine how data flows through various nodes and layers, which will ultimately help clarify how decisions are made.

Competition law policy will dictate that regulators evolve their analysis of tacit collusion to accommodate non-human decision makers

Antitrust, collusion

AI could also have important ramifications for competition law, as pricing mechanisms shift from people, whose actions are more readily covered by existing competition law, to computer algorithms. Competition law has historically focused on corporate actors who are seen to be complicit in limiting or distorting competition, in particular through collusion—agreement and intent to engage in efforts such as price-fixing in an oligopoly or across competing products.

In an AI-driven landscape, new questions arise, such as: What constitutes collusion in an algorithm-dominated environment? What are the boundaries of legality and collusion? What are the antitrust liabilities for developers and users of algorithms? What technology can monitor and constrain AI? If human intelligence can be defined and simulated by a machine, is it possible to create moral, ethical and law-abiding algorithms?

Even when computer algorithms are proxies for market players, overt collusion remains a violation of competition laws. An agreement to utilize algorithms to coordinate prices or reduce competition remains illegal. It is more difficult to determine legality when the collusion is not explicit but may be tacit—such as when prices are coordinated but there was no explicit agreement to coordinate them among participants.

Suppose similar computer algorithms operated by different companies to promote an optimal pricing strategy in which each predicts another's reaction to fluctuating prices. Suppose further that the algorithms determine that the best way to maximize profits for their individual corporate owners is to each set the same optimum price for a particular product. Would the algorithms' concerted action amount to tacit collusion to fix prices?

Will lawmakers allow similar algorithms to run in a competitive industry until their corporate owners learn of the concerted effort to set an optimal price among competing algorithms? Perhaps lawmakers will redefine "agreement" and "intent" for computer actors and introduce proscriptive elements for algorithms and limit the big data available for analysis. At minimum, competition law policy will dictate that regulators evolve their analysis of tacit collusion to accommodate non-human decision makers and consider demanding more transparency into the decisions made by algorithms to permit monitoring and alerting mechanisms for potentially anticompetitive behavior.

What are AI's implications for cybersecurity?

Organizations are already beginning to use AI to bolster cybersecurity by, for example, automating complex processes for detecting attacks and reacting to breaches. And these applications are becoming increasingly sophisticated as learning AI is deployed in the security context. But AI can open vulnerabilities as well, particularly when it depends on interfaces within and across organizations that inadvertently create opportunities for access by nefarious agents. And attackers are beginning to deploy AI, too, meaning they will increasingly develop automated hacks that are able to learn about the systems they target, and to identify vulnerabilities, on the fly.

White-hat defenses, partner vulnerabilities

Financial institutions engage in a broad range of activities that require them to collect, store and use sensitive information about customers, their finances and their behavior toward a variety of ends, from account creation to detecting criminal behavior in banking transactions. Sensitive data includes PII, while less sensitive but valuable information includes customer transaction data.

Besides information collected directly from customers, financial businesses also acquire data, including transaction data, from third parties and the internet, such as social network data, to inform AI-driven efforts to deepen their understanding of their customers. Each financial activity or line of business may use AI differently. Banks may use chatbots, for example, to improve customer experiences. Wealth planning and management services may use robo-advisors to add investment options to customer portfolios. Insurance companies may use AI in claims processing to improve workflows and identify fraud. Financial businesses might use AI to identify relationships that could give rise to new services or identify fraud and money-laundering activities.

But commingling a variety of data in the service of AI can be fraught with risk. First, big-data collection creates a growing attack surface that increases vulnerability to hackers, who look to breach security controls and access PII without authorization. Second, AI provides hackers with a tool to land and expand data breaches (although conversely, AI will also become essential in detecting and preventing breaches). Third, AI can help financial institutions identify new relationships, but in doing so it can sometimes recreate identities that have been masked to comply with privacy requirements. Fourth, commingled data makes it difficult to track whether a data set includes PII, who owns the data, and how the data can be used—in light of restrictions related to data subjects' consent and applicable laws and regulations.

Organizations that use commingled data are still required to know what data they have, where it is located and who owns it. Different laws and regulations apply to various data types across regions. For example, the GDPR will provide EU data subjects with rights regarding companies' processing and holding of their personal data regardless of the organization's location. Individuals in EU member countries can exercise these rights to identify the information collected about them and understand how it is used, as well as to access, review and demand the correction and deletion of that data under certain circumstances.

For compliance, financial institutions can track the provenance of data down to the record level and apply policy management and security controls, such as pseudonymizing and encrypting data. But the micromanagement of large data sets used in AI can be costly and require additional tools, such as applying metadata enhancement techniques to identify data under compliance requirements, tag it appropriately, and apply policy management to control data access and usage. Sometimes, businesses may destroy entire data sets rather than invest the time and resources to micromanage data provenance at the record level.

So-called "personal data vaults" have sprung up in recent years, enabling individuals to store personal data and manage its use, and even to grant anonymized access to it for a fee. Similar efforts have been undertaken with personal health data in recent years. So far, few if any of these efforts have proven particularly commercially successful, although health data is conceivably one area where personal data management could eventually proliferate.

The large and increasing volume of data that financial institutions use for AI presents a large attack surface

Black-hat attacks

Much of the data used in the financial services sector is of great interest to hackers. The large and increasing volume of data that financial institutions use for AI presents a large attack surface. The security challenge may be compounded by partnerships with third-party data providers, particularly when the third parties provide real-time data via API and may not have the hardened, secure interfaces found in financial services.

Financial services institutions are generally required to conduct risk assessments of security measures and design an information security program to protect nonpublic personal data and sensitive business information. Financial institutions are also increasingly responsible for the security measures used by third-party partners and affiliates; from a regulatory and compliance standpoint, insured depository institutions are jointly accountable with third-party vendors.

The repercussions for things like data breaches and allowing unauthorized access to PII go beyond regulatory fines and penalties. Following a data breach, the erosion of customer trust and the tarnishing of a financial brand could shake—if not shatter—a company’s ability to compete in the financial sector, where the secure handling of financial data is imperative. On the other hand, failing to take advantage of new technologies, including AI, is just as certain to leave a company trailing others currently poised to disrupt the financial sector.

The key is to move forward, but smartly. Fortunately, the cybersecurity industry itself is taking advantage of AI, leading to automated detection and response capabilities that were unheard of only a few years ago.

Technology is a tool, and humans must oversee the machines they deploy to ensure choices made by algorithms make sense and align with social principles and regulatory rules

Reality checks

The challenges of implementing AI are real but the benefits are great—including increased speed and efficiency, reduced labor and resource costs, reduced human error, the ability to tailor products and services and otherwise improve the customer experience, and improved security. Financial institutions must proceed with care in the AI era, but few, if any, can afford to sit back or ignore AI.

To stay focused amidst the welter of activity in this space, companies should adopt three broad guidelines:

  • Set out clear principles and document strategies and processes. Any data set may contain the seeds of bias, and learning machines may take unwanted actions that no company can anticipate. Companies that articulate their objectives and show that they have a made concerted effort to comply with regulations and respect consumers can put themselves in a position of strength when challenges arise. This includes rigorously testing systems, and analyzing and documenting their performance. Those that do not do this will leave themselves vulnerable to reputational and legal complications, even when they have done their best to meet high standards.
  • Manage technology with technology. Machines will increasingly be used to monitor other machines, whether that involves regulatory technology (regtech) applications that enable companies to automate compliance or cybersecurity applications that enable companies to identify potential vulnerabilities or trigger responses to actual breaches. AI can enable these kinds of meta-technologies to learn on the fly so they can respond effectively even as the machines they are designed to oversee or defend against evolve in real time. Technology will become increasingly important as a means of managing technological complexity as data flows continue their exponential growth trajectories, and increasingly sophisticated AI goes mainstream.
  • Keep people front and center. AI programs are increasingly able to make sophisticated judgments and decisions and even write their algorithms based on insights they learned through experience to guide their future choices and actions. But most experts agree that we are a long way from a future in which computers can operate autonomously in contexts requiring sophisticated, dynamic use of judgment. It is always critical to remember that technology is a tool, and humans must oversee the machines they deploy to ensure choices made by algorithms make sense and align with social principles and regulatory rules. In the vast majority of cases, there is no substitute for human judgment. Companies that cede too much control to technology will increase their risk exposure significantly, if not catastrophically.

We are at the beginning of a journey that will take some time to unfold. The financial services sector will be one of the most important sectors forging new commercial applications for AI in the coming years, but it may also be among the most vulnerable to AI solutions that go sideways. Companies that balance the various interests of stakeholders will produce significant benefits for business and consumers.

Click here to download PDF.

DISCLAIMER: Because of the generality of this update, the information provided herein may not be applicable in all situations and should not be acted upon without specific legal advice based on particular situations.

© White & Case LLP | Attorney Advertising

Written by:

White & Case LLP

White & Case LLP on:

Readers' Choice 2017
Reporters on Deadline

"My best business intelligence, in one easy email…"

Your first step to building a free, personalized, morning email brief covering pertinent authors and topics on JD Supra:
*By using the service, you signify your acceptance of JD Supra's Privacy Policy.
Custom Email Digest
- hide

JD Supra Privacy Policy

Updated: May 25, 2018:

JD Supra is a legal publishing service that connects experts and their content with broader audiences of professionals, journalists and associations.

This Privacy Policy describes how JD Supra, LLC ("JD Supra" or "we," "us," or "our") collects, uses and shares personal data collected from visitors to our website (located at (our "Website") who view only publicly-available content as well as subscribers to our services (such as our email digests or author tools)(our "Services"). By using our Website and registering for one of our Services, you are agreeing to the terms of this Privacy Policy.

Please note that if you subscribe to one of our Services, you can make choices about how we collect, use and share your information through our Privacy Center under the "My Account" dashboard (available if you are logged into your JD Supra account).

Collection of Information

Registration Information. When you register with JD Supra for our Website and Services, either as an author or as a subscriber, you will be asked to provide identifying information to create your JD Supra account ("Registration Data"), such as your:

  • Email
  • First Name
  • Last Name
  • Company Name
  • Company Industry
  • Title
  • Country

Other Information: We also collect other information you may voluntarily provide. This may include content you provide for publication. We may also receive your communications with others through our Website and Services (such as contacting an author through our Website) or communications directly with us (such as through email, feedback or other forms or social media). If you are a subscribed user, we will also collect your user preferences, such as the types of articles you would like to read.

Information from third parties (such as, from your employer or LinkedIn): We may also receive information about you from third party sources. For example, your employer may provide your information to us, such as in connection with an article submitted by your employer for publication. If you choose to use LinkedIn to subscribe to our Website and Services, we also collect information related to your LinkedIn account and profile.

Your interactions with our Website and Services: As is true of most websites, we gather certain information automatically. This information includes IP addresses, browser type, Internet service provider (ISP), referring/exit pages, operating system, date/time stamp and clickstream data. We use this information to analyze trends, to administer the Website and our Services, to improve the content and performance of our Website and Services, and to track users' movements around the site. We may also link this automatically-collected data to personal information, for example, to inform authors about who has read their articles. Some of this data is collected through information sent by your web browser. We also use cookies and other tracking technologies to collect this information. To learn more about cookies and other tracking technologies that JD Supra may use on our Website and Services please see our "Cookies Guide" page.

How do we use this information?

We use the information and data we collect principally in order to provide our Website and Services. More specifically, we may use your personal information to:

  • Operate our Website and Services and publish content;
  • Distribute content to you in accordance with your preferences as well as to provide other notifications to you (for example, updates about our policies and terms);
  • Measure readership and usage of the Website and Services;
  • Communicate with you regarding your questions and requests;
  • Authenticate users and to provide for the safety and security of our Website and Services;
  • Conduct research and similar activities to improve our Website and Services; and
  • Comply with our legal and regulatory responsibilities and to enforce our rights.

How is your information shared?

  • Content and other public information (such as an author profile) is shared on our Website and Services, including via email digests and social media feeds, and is accessible to the general public.
  • If you choose to use our Website and Services to communicate directly with a company or individual, such communication may be shared accordingly.
  • Readership information is provided to publishing law firms and authors of content to give them insight into their readership and to help them to improve their content.
  • Our Website may offer you the opportunity to share information through our Website, such as through Facebook's "Like" or Twitter's "Tweet" button. We offer this functionality to help generate interest in our Website and content and to permit you to recommend content to your contacts. You should be aware that sharing through such functionality may result in information being collected by the applicable social media network and possibly being made publicly available (for example, through a search engine). Any such information collection would be subject to such third party social media network's privacy policy.
  • Your information may also be shared to parties who support our business, such as professional advisors as well as web-hosting providers, analytics providers and other information technology providers.
  • Any court, governmental authority, law enforcement agency or other third party where we believe disclosure is necessary to comply with a legal or regulatory obligation, or otherwise to protect our rights, the rights of any third party or individuals' personal safety, or to detect, prevent, or otherwise address fraud, security or safety issues.
  • To our affiliated entities and in connection with the sale, assignment or other transfer of our company or our business.

How We Protect Your Information

JD Supra takes reasonable and appropriate precautions to insure that user information is protected from loss, misuse and unauthorized access, disclosure, alteration and destruction. We restrict access to user information to those individuals who reasonably need access to perform their job functions, such as our third party email service, customer service personnel and technical staff. You should keep in mind that no Internet transmission is ever 100% secure or error-free. Where you use log-in credentials (usernames, passwords) on our Website, please remember that it is your responsibility to safeguard them. If you believe that your log-in credentials have been compromised, please contact us at

Children's Information

Our Website and Services are not directed at children under the age of 16 and we do not knowingly collect personal information from children under the age of 16 through our Website and/or Services. If you have reason to believe that a child under the age of 16 has provided personal information to us, please contact us, and we will endeavor to delete that information from our databases.

Links to Other Websites

Our Website and Services may contain links to other websites. The operators of such other websites may collect information about you, including through cookies or other technologies. If you are using our Website or Services and click a link to another site, you will leave our Website and this Policy will not apply to your use of and activity on those other sites. We encourage you to read the legal notices posted on those sites, including their privacy policies. We are not responsible for the data collection and use practices of such other sites. This Policy applies solely to the information collected in connection with your use of our Website and Services and does not apply to any practices conducted offline or in connection with any other websites.

Information for EU and Swiss Residents

JD Supra's principal place of business is in the United States. By subscribing to our website, you expressly consent to your information being processed in the United States.

  • Our Legal Basis for Processing: Generally, we rely on our legitimate interests in order to process your personal information. For example, we rely on this legal ground if we use your personal information to manage your Registration Data and administer our relationship with you; to deliver our Website and Services; understand and improve our Website and Services; report reader analytics to our authors; to personalize your experience on our Website and Services; and where necessary to protect or defend our or another's rights or property, or to detect, prevent, or otherwise address fraud, security, safety or privacy issues. Please see Article 6(1)(f) of the E.U. General Data Protection Regulation ("GDPR") In addition, there may be other situations where other grounds for processing may exist, such as where processing is a result of legal requirements (GDPR Article 6(1)(c)) or for reasons of public interest (GDPR Article 6(1)(e)). Please see the "Your Rights" section of this Privacy Policy immediately below for more information about how you may request that we limit or refrain from processing your personal information.
  • Your Rights
    • Right of Access/Portability: You can ask to review details about the information we hold about you and how that information has been used and disclosed. Note that we may request to verify your identification before fulfilling your request. You can also request that your personal information is provided to you in a commonly used electronic format so that you can share it with other organizations.
    • Right to Correct Information: You may ask that we make corrections to any information we hold, if you believe such correction to be necessary.
    • Right to Restrict Our Processing or Erasure of Information: You also have the right in certain circumstances to ask us to restrict processing of your personal information or to erase your personal information. Where you have consented to our use of your personal information, you can withdraw your consent at any time.

You can make a request to exercise any of these rights by emailing us at or by writing to us at:

Privacy Officer
JD Supra, LLC
10 Liberty Ship Way, Suite 300
Sausalito, California 94965

You can also manage your profile and subscriptions through our Privacy Center under the "My Account" dashboard.

We will make all practical efforts to respect your wishes. There may be times, however, where we are not able to fulfill your request, for example, if applicable law prohibits our compliance. Please note that JD Supra does not use "automatic decision making" or "profiling" as those terms are defined in the GDPR.

  • Timeframe for retaining your personal information: We will retain your personal information in a form that identifies you only for as long as it serves the purpose(s) for which it was initially collected as stated in this Privacy Policy, or subsequently authorized. We may continue processing your personal information for longer periods, but only for the time and to the extent such processing reasonably serves the purposes of archiving in the public interest, journalism, literature and art, scientific or historical research and statistical analysis, and subject to the protection of this Privacy Policy. For example, if you are an author, your personal information may continue to be published in connection with your article indefinitely. When we have no ongoing legitimate business need to process your personal information, we will either delete or anonymize it, or, if this is not possible (for example, because your personal information has been stored in backup archives), then we will securely store your personal information and isolate it from any further processing until deletion is possible.
  • Onward Transfer to Third Parties: As noted in the "How We Share Your Data" Section above, JD Supra may share your information with third parties. When JD Supra discloses your personal information to third parties, we have ensured that such third parties have either certified under the EU-U.S. or Swiss Privacy Shield Framework and will process all personal data received from EU member states/Switzerland in reliance on the applicable Privacy Shield Framework or that they have been subjected to strict contractual provisions in their contract with us to guarantee an adequate level of data protection for your data.

California Privacy Rights

Pursuant to Section 1798.83 of the California Civil Code, our customers who are California residents have the right to request certain information regarding our disclosure of personal information to third parties for their direct marketing purposes.

You can make a request for this information by emailing us at or by writing to us at:

Privacy Officer
JD Supra, LLC
10 Liberty Ship Way, Suite 300
Sausalito, California 94965

Some browsers have incorporated a Do Not Track (DNT) feature. These features, when turned on, send a signal that you prefer that the website you are visiting not collect and use data regarding your online searching and browsing activities. As there is not yet a common understanding on how to interpret the DNT signal, we currently do not respond to DNT signals on our site.

Access/Correct/Update/Delete Personal Information

For non-EU/Swiss residents, if you would like to know what personal information we have about you, you can send an e-mail to We will be in contact with you (by mail or otherwise) to verify your identity and provide you the information you request. We will respond within 30 days to your request for access to your personal information. In some cases, we may not be able to remove your personal information, in which case we will let you know if we are unable to do so and why. If you would like to correct or update your personal information, you can manage your profile and subscriptions through our Privacy Center under the "My Account" dashboard. If you would like to delete your account or remove your information from our Website and Services, send an e-mail to

Changes in Our Privacy Policy

We reserve the right to change this Privacy Policy at any time. Please refer to the date at the top of this page to determine when this Policy was last revised. Any changes to our Privacy Policy will become effective upon posting of the revised policy on the Website. By continuing to use our Website and Services following such changes, you will be deemed to have agreed to such changes.

Contacting JD Supra

If you have any questions about this Privacy Policy, the practices of this site, your dealings with our Website or Services, or if you would like to change any of the information you have provided to us, please contact us at:

JD Supra Cookie Guide

As with many websites, JD Supra's website (located at (our "Website") and our services (such as our email article digests)(our "Services") use a standard technology called a "cookie" and other similar technologies (such as, pixels and web beacons), which are small data files that are transferred to your computer when you use our Website and Services. These technologies automatically identify your browser whenever you interact with our Website and Services.

How We Use Cookies and Other Tracking Technologies

We use cookies and other tracking technologies to:

  1. Improve the user experience on our Website and Services;
  2. Store the authorization token that users receive when they login to the private areas of our Website. This token is specific to a user's login session and requires a valid username and password to obtain. It is required to access the user's profile information, subscriptions, and analytics;
  3. Track anonymous site usage; and
  4. Permit connectivity with social media networks to permit content sharing.

There are different types of cookies and other technologies used our Website, notably:

  • "Session cookies" - These cookies only last as long as your online session, and disappear from your computer or device when you close your browser (like Internet Explorer, Google Chrome or Safari).
  • "Persistent cookies" - These cookies stay on your computer or device after your browser has been closed and last for a time specified in the cookie. We use persistent cookies when we need to know who you are for more than one browsing session. For example, we use them to remember your preferences for the next time you visit.
  • "Web Beacons/Pixels" - Some of our web pages and emails may also contain small electronic images known as web beacons, clear GIFs or single-pixel GIFs. These images are placed on a web page or email and typically work in conjunction with cookies to collect data. We use these images to identify our users and user behavior, such as counting the number of users who have visited a web page or acted upon one of our email digests.

JD Supra Cookies. We place our own cookies on your computer to track certain information about you while you are using our Website and Services. For example, we place a session cookie on your computer each time you visit our Website. We use these cookies to allow you to log-in to your subscriber account. In addition, through these cookies we are able to collect information about how you use the Website, including what browser you may be using, your IP address, and the URL address you came from upon visiting our Website and the URL you next visit (even if those URLs are not on our Website). We also utilize email web beacons to monitor whether our emails are being delivered and read. We also use these tools to help deliver reader analytics to our authors to give them insight into their readership and help them to improve their content, so that it is most useful for our users.

Analytics/Performance Cookies. JD Supra also uses the following analytic tools to help us analyze the performance of our Website and Services as well as how visitors use our Website and Services:

  • HubSpot - For more information about HubSpot cookies, please visit
  • New Relic - For more information on New Relic cookies, please visit
  • Google Analytics - For more information on Google Analytics cookies, visit To opt-out of being tracked by Google Analytics across all websites visit This will allow you to download and install a Google Analytics cookie-free web browser.

Facebook, Twitter and other Social Network Cookies. Our content pages allow you to share content appearing on our Website and Services to your social media accounts through the "Like," "Tweet," or similar buttons displayed on such pages. To accomplish this Service, we embed code that such third party social networks provide and that we do not control. These buttons know that you are logged in to your social network account and therefore such social networks could also know that you are viewing the JD Supra Website.

Controlling and Deleting Cookies

If you would like to change how a browser uses cookies, including blocking or deleting cookies from the JD Supra Website and Services you can do so by changing the settings in your web browser. To control cookies, most browsers allow you to either accept or reject all cookies, only accept certain types of cookies, or prompt you every time a site wishes to save a cookie. It's also easy to delete cookies that are already saved on your device by a browser.

The processes for controlling and deleting cookies vary depending on which browser you use. To find out how to do so with a particular browser, you can use your browser's "Help" function or alternatively, you can visit which explains, step-by-step, how to control and delete cookies in most browsers.

Updates to This Policy

We may update this cookie policy and our Privacy Policy from time-to-time, particularly as technology changes. You can always check this page for the latest version. We may also notify you of changes to our privacy policy by email.

Contacting JD Supra

If you have any questions about how we use cookies and other tracking technologies, please contact us at:

- hide

This website uses cookies to improve user experience, track anonymous site usage, store authorization tokens and permit sharing on social media networks. By continuing to browse this website you accept the use of cookies. Click here to read more about how we use cookies.