chevron-down Created with Sketch Beta.

GPSolo Magazine

GPSolo March/April 2025 (42:2): AI for Lawyers

Privacy Risks with AI: Your Data, Their Knowledge

Gabriel Buehler

Summary

  • The more an artificial intelligence (AI) system learns, the more comprehensive its knowledge becomes, making it possible for AI to predict future actions with unsettling precision.
  • One of the most significant privacy risks associated with AI is the question of who has access to the data that powers these systems.
  • Carefully review the privacy policies of the AI services and apps you use. Many services allow you to opt out of certain data collection practices.
  • Both the EU and the United States have recently proposed comprehensive measures to govern AI to ensure privacy, increase accountability, and avoid misuse.
Privacy Risks with AI: Your Data, Their Knowledge
oonal/E+ via Getty Images

Jump to:

Artificial Intelligence (AI) is reshaping industries by improving efficiency, solving complex problems, and providing personalized experiences. But as AI becomes an integral part of our everyday lives, its extensive data collection practices raise profound privacy concerns. Whether through the use of AI in health care, the legal industry, or law enforcement, individuals’ data is increasingly becoming a commodity used to train algorithms, predict behaviors, and even make decisions that can impact our lives. The trade-off between convenience and privacy has never been more evident, and it’s crucial to examine the risks involved.

This article explores the various privacy risks presented by AI, examines recent instances of abuse in key industries, and offers practical advice for safeguarding personal information in an AI-driven world.

The Growing Scope of AI and Privacy Concerns

AI systems rely heavily on data, and the more data they can access, the more accurate and insightful their predictions become. From machine learning models predicting consumer behavior to deep learning models diagnosing diseases, AI systems collect data from a wide range of sources: online browsing, wearable devices, smart home technology, medical records, and more.

The problem with this data-centric model is not only its volume but also its sensitivity and the sensitivity of the data on which these AI systems are trained. As AI continues to evolve, it is becoming more capable of making highly accurate predictions about individuals’ personal habits, preferences, and even more sensitive aspects of their lives, such as their health and financial stability. For instance, AI can predict when a person is likely to make a purchase, when they might suffer a health issue, or even when they are considering a job change.

While these predictions provide tremendous benefits in terms of convenience, efficiency, and targeted services, they also raise significant questions about how much companies and governments can know about individuals—and who gets access to that information.

     

How Much Does AI Really Know?

AI’s ability to create highly accurate profiles of individuals is built on the vast amounts of data it consumes. It is not just data about what you’ve bought or what websites you’ve visited but also about your interactions with AI-powered devices. For instance, smart speakers, such as Amazon’s Alexa or Google Home, continuously listen to conversations and queries, potentially collecting sensitive personal information even when they’re not actively in use.

AI also uses non-obvious data to make inferences about individuals. Machine learning algorithms can predict health outcomes based on data gathered by wearable devices, which can reveal patterns related to a person’s activity level, location, and even sleep habits. Similarly, AI can forecast financial stability by analyzing your spending habits, subscription services, and even your social media posts. The more an AI system learns, the more comprehensive its knowledge becomes, making it possible for AI to predict future actions with unsettling precision.

The implications of this predictive power are far-reaching, as AI systems now hold personal data that can be used to create detailed profiles of individuals. These profiles might contain information that an individual never intended to share or even be aware of. The convenience of AI comes with a loss of control over personal information, and the resulting questions about privacy have yet to be fully addressed.

Who Has Access to Your Data?

One of the most significant privacy risks associated with AI is the question of who has access to the data that powers these systems. The information collected by AI systems is often shared with third parties, such as marketers, advertisers, and other companies that pay for access to user data. In many cases, individuals are unaware of the full extent of the data being collected and shared.

Data brokers, for instance, buy and sell personal information to organizations that use it for targeted advertising or to assess creditworthiness. Governments also have the power to collect personal data through AI systems, often without the consent of individuals. In some countries, government surveillance programs monitor citizens’ online activity, while in others, AI-driven facial recognition tools are used to track movements in public spaces.

This lack of transparency about who has access to personal data is a significant concern. In many cases, individuals are unaware that their data is being used or sold. Even when individuals do have control over their data, opting out of these data practices is often complicated and time-consuming.

Criminals are also targeting this data. In recent years, data breaches have exposed millions of individuals’ personal information, including names, addresses, financial details, and even biometric data. One of the most concerning aspects of AI is the way it can enhance the value of stolen data. AI systems often use machine learning algorithms to create highly detailed profiles of individuals, which can be used to assist in identity theft or to help hackers plan more effective attacks.

Case Studies: AI Abuse Across Industries

The ethical and privacy concerns surrounding AI are not hypothetical. There have been numerous real-world cases where AI systems have been used in ways that violate individuals’ rights. Below, we examine three examples from the legal industry, health care, and law enforcement.

Legal Industry: GitHub Copilot and Copyright Infringement

GitHub Copilot, an AI-powered tool developed by GitHub and OpenAI, assists developers by suggesting code snippets based on prior examples. In 2022, developers filed a lawsuit against GitHub and OpenAI, alleging that Copilot had been trained on vast amounts of open-source code without proper attribution or permission from the original authors.

The lawsuit, Software Freedom Conservancy v. GitHub, claimed that GitHub’s use of Copilot violated copyright laws by using proprietary code to train the AI without compensating or informing the original creators. The plaintiffs argued that Copilot’s suggestions were based on code that was copyrighted, leading to concerns about the erosion of intellectual property rights in an AI-driven world.

This case highlights a critical issue in AI development: the lack of clarity regarding the ownership of data used to train AI systems. While some argue that AI models can be trained on publicly available data under the doctrine of “fair use,” others believe this practice violates the intellectual property rights of creators. The case is still ongoing, but it underscores the growing legal and ethical concerns surrounding AI in industries such as software development.

Health Care: The Risks of Bias in AI Diagnosis

AI systems in health care promise to revolutionize diagnostics, but there have been significant concerns about their fairness and accuracy. A notable example is a 2019 study published in Science that found that an AI system used in health care had significant biases. The system, designed to predict health care needs and allocate resources, was trained on data that disproportionately reflected the needs of white patients, leading to a discriminatory allocation of health care resources.

In this case, the AI system exhibited a bias that favored individuals with more expensive health care needs, which typically corresponded to wealthier, often white, populations. As a result, the AI system underrepresented the needs of Black and Hispanic patients, exacerbating existing health care disparities. The findings of this study sparked widespread concern over the fairness of AI systems in sensitive sectors such as health care, where biased outcomes can have life-or-death consequences.

Law Enforcement: The Misuse of Facial Recognition Technology

Facial recognition technology has been deployed by law enforcement agencies around the world, raising significant concerns about privacy and civil liberties. In 2018, the American Civil Liberties Union tested Amazon’s facial recognition system, Rekognition, and found that it misidentified members of Congress as individuals who had been arrested for crimes.

The technology was used to match photos of people in a database of criminal mugshots, but it disproportionately misidentified people of color. This led to widespread criticism of facial recognition tools, with many arguing that they contribute to racial profiling and the violation of individuals’ right to privacy.

In response to growing concerns about the technology’s impact on civil rights, several cities in the U.S. have passed legislation banning the use of facial recognition in public spaces, and companies such as Amazon have been encouraged to suspend sales of the technology to law enforcement agencies.

Steps to Safeguard Your Privacy

Given the pervasive risks associated with AI-driven data collection, individuals must take steps to protect their privacy. The following strategies can help reduce your exposure and minimize the risks of AI misuse:

  1. Understand privacy policies. Start by carefully reviewing the privacy policies of the services and apps you use. Many services allow you to opt out of certain data collection practices, but you need to be proactive in setting your preferences.
  2. Use privacy-focused tools. Leverage privacy-enhancing tools, including VPNs, encrypted messaging apps such as Signal, and privacy-focused search engines such as DuckDuckGo. These tools can reduce your exposure to data collection by third parties and help anonymize your online activity.
  3. Opt out of data collection. Many websites and services offer ways to limit or opt out of data collection. Several organizations help consumers remove their data from commercial databases, limit its use in AI training, and opt out of unwanted marketing. Some examples of these websites are DeleteMe (a privacy service for removing personal information from online databases), OptOutPrescreen (a resource for opting out of prescreened credit and insurance offers), and Privacy Rights Clearinghouse (a nonprofit focusing on privacy education and resources).
  4. Advocate for stronger regulations. Support privacy regulations, such as the General Data Protection Regulation (GDPR) in the European Union (EU) and the California Consumer Privacy Act (CCPA), which offer individuals more control over their personal data.
  5. Limit data sharing. Be selective about what you share online and avoid linking multiple accounts to minimize data aggregation.

EU AI Act and U.S. Federal Proposals on AI Regulation

As AI technologies advance, regulatory frameworks are being developed to address the ethical, privacy, and safety concerns they raise. Both the EU and the United States have recently proposed comprehensive measures to govern AI to ensure privacy, increase accountability, and avoid misuse.

The EU AI Act: A Leading Regulatory Framework

The EU AI Act, proposed in April 2021, is the first comprehensive attempt to regulate AI at a systemic level. It categorizes AI applications into four risk levels—unacceptable, high, limited, and minimal—and implements varying degrees of oversight based on the potential risks posed by these technologies.

Key provisions:

  1. Risk-based categorization.
    1. Unacceptable risk. AI applications deemed harmful, such as social scoring systems by governments or real-time biometric surveillance in public spaces, are outright banned.
    2. High risk. AI used in critical areas such as health care, law enforcement, and employment is subject to strict regulatory requirements, including transparency, bias mitigation, and accuracy testing.
    3. Limited/minimal risk. Applications such as chatbots must simply inform users that they are interacting with AI.
  2. Accountability and transparency. The Act requires high-risk AI systems to maintain detailed documentation and transparency reports, making it easier for regulators to audit their operations.
  3. Data privacy. The regulation works in tandem with the GDPR to ensure that AI applications respect individual privacy and data security. For instance, it prohibits the use of sensitive personal data for AI training without explicit consent.

The EU AI Act aims to establish a gold standard for AI regulation by prioritizing ethical development and minimizing harmful societal impacts. However, critics argue that the stringent requirements may stifle innovation and create significant compliance costs, particularly for small businesses and start-ups.

Proposed U.S. Federal AI Regulations

The United States, traditionally a technology innovation hub, has taken a less centralized approach to AI regulation compared to the EU. However, recent federal initiatives signal a growing recognition of the need for comprehensive oversight.

Key proposals and frameworks:

  1. Blueprint for an AI Bill of Rights. Released by the White House Office of Science and Technology Policy in 2022, the Blueprint for an AI Bill of Rights outlines principles aimed at protecting individuals from the potential harms of AI. These include:
    1. Safe and effective systems: AI systems should undergo rigorous testing to avoid discrimination and bias.
    2. Data privacy: Individuals should have control over how their data is collected and used, with strong protections against misuse.
    3. Transparency: AI systems should clearly explain their processes and decision-making logic to users.
  2. Algorithmic Accountability Act. This proposed legislation would require companies to evaluate the impact of automated systems on factors such as bias, discrimination, and privacy. It emphasizes the need for transparency in AI applications used in critical areas such as hiring, lending, and law enforcement.
  3. Federal Trade Commission (FTC) enforcement. The FTC has increased its focus on AI, warning companies against deploying biased or deceptive algorithms. In 2021, the FTC issued guidance emphasizing that businesses could face penalties for using AI in ways that violate consumer protection laws.

EU vs. U.S. Approaches

The U.S. approach to AI regulation is more fragmented than the EU’s, with a reliance on sector-specific rules and existing consumer protection laws. Critics argue that this piecemeal strategy lacks the comprehensiveness needed to address the full scope of AI risks. However, proponents suggest that it provides flexibility, allowing innovation to flourish while addressing specific harms as they arise.

 

Comparison of EU and U.S. Approaches to AI Regulation

 

EU AI Act

U.S. Federal Proposals and State-by-State Regulations

Regulatory Scope Broad, unified framework covering all industries. Sector-specific guidance and principles.
Risk Categorization Risk-based, with strict oversight for high-risk AI. Focused on harms, such as bias and discrimination.
Data Privacy Strong protections aligned with GDPR. Emphasis on individual control over data.
Transparency Mandatory for high-risk systems. Encouraged but not consistently enforced.

Challenges and Opportunities

Both the EU AI Act and U.S. federal proposals represent critical steps toward addressing the ethical and privacy risks posed by AI. However, challenges remain:

On the positive side, these regulatory efforts provide an opportunity to set global standards for responsible AI development. By addressing issues such as data privacy, accountability, and bias, they pave the way for ethical and transparent AI technologies that benefit society while minimizing harm.

For more information, refer to the following resources:

Immense Potential, Significant Risks

AI technologies offer immense potential to improve many aspects of life, but the growing reliance on personal data to fuel these systems comes with significant risks. From data breaches to discriminatory outcomes, the misuse of AI across industries is an ongoing concern. By understanding these risks and taking proactive measures, individuals can protect their privacy in an increasingly AI-driven world. While the balance between convenience and privacy is complex, staying informed and taking action can help mitigate some of the dangers associated with AI.

    Author