The Unhinged Librarian

By Sam Chada

Library technology consultant with 20 years in library tech. I've been on both sides of the sales call - I know what data vendors are collecting and what could go wrong.

13 min read

Surveillance or Service? AI Privacy for Vulnerable Patrons

AI tools collect patron behavior data that can harm vulnerable communities. Libraries must assess privacy impact differently for different populations.

Why I wrote this: I just scrubbed a leaked search log that put a domestic violence survivor at risk; I don't want another library to repeat that.

Any vendor that won't sign data minimization and breach response terms is selling you patron danger, not personalization.

Design Launch Breach Drill 87% 49% 46%
Original chart I sketched while writing: rough checkpoints for AI Surveillance Vulnerable Patrons. Mark your own numbers on top of mine.

Libraries serve vulnerable populations. That's our actual mission. Not recommendations. Not personalization. Not vendor revenue. People who come to us because they have no other safe place to ask their questions.

TL;DR
  • Library AI (surveillance, behavior tracking, recommendation algorithms) disproportionately impacts vulnerable patrons: unhoused, immigrant, low-income communities who lack privacy alternatives.
  • AI privacy issues: data collection without informed consent, algorithmic bias in recommendations, integration with law enforcement surveillance, and patron data sold to third parties.
  • Vulnerable patrons rely on libraries as privacy-safe spaces. AI that enables tracking, identifies behavioral patterns, or enables law enforcement access breaches this implicit trust.
  • Board-level policy needed: minimum data collection, explicit vendor restrictions, opt-in vs. automatic enrollment, and transparency about what AI tools actually do with patron information.

When you implement an AI recommendation engine, a chatbot, or search analytics, vendors tell you it's "privacy-preserving." They use phrases like "aggregated, anonymized data." It sounds safe. It's not.

Here's what happens in practice: Your vendor collects search queries, reading history, and conversation logs. Then, six months later, the vendor gets hacked. Data leaks. Real people's searches are now exposed.

For the affluent patron searching for book club recommendations, it's annoying. For the immigrant patron who searched "asylum petition timeline," it's ICE showing up at their door. For the domestic violence survivor who searched "confidential shelter near me," it's an abuser finding their location. For the LGBTQ+ youth who searched "transgender healthcare," it's a family or religious institution weaponizing that data against them.

I know this sounds dramatic. I've cleaned up the fallout. I've notified a survivor that her search history was exposed. The risk isn't theoretical - it's what I watch happen when libraries collect vulnerable people's information without understanding who could weaponize it.


What AI Tools Actually Collect (And What Vendors Know About It)

Here's what your vendor is collecting, even when they claim it's "privacy-preserving":

Vendors claim they "de-identify" this data. That's a marketing statement, not a technical guarantee. Researchers have repeatedly shown that "anonymized" datasets can be re-identified by correlating multiple data points. An IP address + search timestamp + library location is often enough to pinpoint an individual.

But here's what vendors are counting on you not asking: Why does the algorithm need to keep individual-level data at all? If they're truly just improving recommendations, they could aggregate patterns without storing who searched for what. They don't, because retention creates value. Your patron data is an asset they can sell, license, or exploit.

Breaches happen. Every vendor gets breached eventually. And when they do, vulnerable people who trusted your library suddenly have their information in the hands of people who want to harm them.


Real Harms (Not Hypothetical)

Immigration Enforcement

An undocumented patron comes to your library because it's the only place they can ask questions without fear. They search "How to apply for asylum" or "Are undocumented immigrants eligible for services?" or "What happens if I get pulled over without papers?" If your AI tool logs these searches and the vendor gets breached, ICE now has a targeting list they didn't have before.

This isn't speculative. ICE agents have used library browsing records, checkout histories, and digital footprints to identify and track immigrants. Your recommendation engine data, if breached, becomes another data source they can use. You're not just documenting information need - you're creating a breadcrumb trail to vulnerable people.

Law Enforcement Targeting

A patron searches for information about bail, criminal defense, protest tactics, or their own arrest. If that data is breached and accessed by law enforcement, it becomes evidence or targeting information.

Librarians have faced subpoenas for patron records. If you have detailed behavior logs, those subpoenas get a lot of useful data.

Domestic Violence

A survivor comes to your library to plan their escape. They search for shelter locations, legal aid for protective orders, child custody resources, and how to safely leave. They use the library because they can't search at home - the abuser controls the devices and accounts. If your AI system logs these searches and the vendor gets breached, the abuser now has a roadmap: shelter names, legal strategies, timing. A data breach becomes a hunting guide.

Survivors choose libraries because we're supposed to be safe. Implementing surveillance systems - even well-intentioned ones - breaks that safety. It sends a message: your research here might be exposed. And for someone running from violence, that risk is lethal.

Sexual Orientation/Gender Identity

A teenager from a conservative religious household uses your library to explore who they are. They search for "Is it normal to be trans?" and "LGBTQ+-friendly therapists near me" and look at the library's digital collection of books about gender identity. They trust your library because librarians protect privacy. If your AI system logs these searches and the vendor gets breached - or if a parent demands the data through a legal request - that teenager's parents now know they're exploring their identity before they're ready to come out.

Outing is dangerous. Conversion therapy is still happening. Religious institutions and parents use any information they can find to intervene. When you implement surveillance systems in libraries, you're telling LGBTQ+ youth: this isn't a safe place to explore. You're forcing them to choose between the library and their safety.

Medical/Mental Health

A patron searches for HIV treatment, mental health resources, or specific medical conditions. If that data is breached, it becomes identifiable health information that insurers, employers, or others could misuse.


Vulnerable Populations Matrix

Population Risk If Data Breached Harm Type
Immigrants/Undocumented Targeted for ICE deportation efforts Removal from country
LGBTQ+ Youth Outed to family/community/institutions Family rejection, conversion therapy
Domestic Violence Survivors Abuser accesses location/shelter information Physical harm, re-victimization
Political Activists State surveillance of activism/organizing Targeting, arrest, harassment
People w/ Undisclosed Health Issues Employers/insurers access health searches Discrimination, denial of services
Asylum Seekers Information about asylum search history Denial of asylum, deportation
People in Legal Trouble Law enforcement accesses legal research Additional charges, conviction

The hard truth: You cannot consent your way out of these risks. Vendors know this. That's why they keep collecting. A domestic violence survivor can't "opt out of tracking" if opting out means losing access to resources she needs. Consent only works when people have a genuine choice. But vulnerable people don't have a choice - they choose between using the library with surveillance or not using it at all. And if they choose to use it, a breach doesn't ask for their consent. It just happens.


How to Protect Your Mission (3-Step Template)

Before you implement any AI tool - and I mean before you sign anything - run this assessment. This isn't a box to check. This is protecting the vulnerable people who depend on your library.

Step 1: Demand Truth About Data Collection

Don't ask the vendor's sales rep. Ask the vendor's privacy officer or data engineer: What data does your tool collect? How long is it retained? Who has access? Can it be tied back to individual patrons? If they hedge, push back. "Eventually de-identified" is not a commitment. "May be accessed by" is not acceptable.

Document what they tell you - everything:

Step 2: Assess Harm by Population (Be Specific)

For each vulnerable population your library serves, ask: If this data is breached tomorrow, what happens to them? Don't be abstract. Get specific. Get angry about it.

Example: Recommendation Engine + Immigrant Community in Your Area

Step 3: Make a Real Decision

You have three options. Pick one. Own it.

Option A: Don't Implement

If the harm is severe and can't be mitigated, don't use the tool. This is a legitimate decision. More than legitimate - it's the right decision if the risk to vulnerable people is too high. Not every vendor feature is worth the risk. A recommendation engine is nice. Protecting your community's safety is essential.

Option B: Proceed with Aggressive Safeguards

If the tool serves a genuine need, demand contractual safeguards. Don't ask nicely. Demand them. Here's what you require:

• Data de-identified within 48 hours (not "eventually") • No retention of individual transaction history • Encryption both in transit and at rest • Immediate breach notification (24 hours max) • Right to audit vendor security practices • Subpoena response protocol: Vendor notifies library before responding • Limitation on vendor employee access to patron data

Option C: Proceed with Informed Consent

If you think transparency is enough, require informed consent where patrons understand exactly what's collected, who might access it, and can opt out without losing core services. Be clear: "Opting out means you don't use the recommendation engine, but you can still search, check out books, and use all other library services." (Note: This is often insufficient for vulnerable populations, but it's better than collecting data without asking. Many vulnerable people will opt out once they understand the real risk.)


Vendor Contract Language (Copy-Paste Ready)

Add these clauses to your vendor contracts:

DATA DE-IDENTIFICATION Vendor shall de-identify all personally identifiable patron information within 48 hours of collection. De-identification must comply with HIPAA Safe Harbor method or equivalent. Vendor shall not retain individual transaction records beyond this period. BREACH NOTIFICATION Vendor shall notify Library within 24 hours of discovering any unauthorized access, loss, or theft of patron data. Notification must include: affected population size, data types exposed, actions taken to contain breach. SUBPOENA RESPONSE If Vendor receives a subpoena, regulatory request, or legal inquiry requesting patron data, Vendor shall: (1) Notify Library immediately; (2) Not respond without Library consent or court order; (3) Provide Library opportunity to seek protective order. NO PROFILING Vendor shall not use patron behavioral data to create profiles, scores, or predictive models of individual patrons without explicit written consent. Vendor shall not sell, license, or share patron data with third parties. AUDIT RIGHTS Library retains right to audit Vendor's security practices, data handling procedures, and vendor subcontractor access to patron data upon 30 days notice, no less than annually.

Implementation Checklist

Before going live with any AI tool:


Talking Points for Your Board

"One data breach could put vulnerable patrons at risk. Not someday - now."

Concrete example: "If our recommendation engine data is breached, every undocumented patron who searched 'visa petition' or 'asylum' is now identifiable. ICE agents have used library data to target deportations. We would be giving them a targeting list. This isn't theoretical - it's what happened with ICE and library records."

"Our core obligation is patron safety. AI tools create new risks to that."

Libraries have always protected patron privacy. It's in the ALA Code of Ethics. It's in our intellectual freedom principles. AI tools force us to choose: Do we serve vendor profit or patron safety? We can't do both.

"Privacy is the foundation of access. Vulnerable people won't use a library that surveils them."

Immigrants won't search for visa information if they fear the data will be used against them. Domestic violence survivors won't seek shelter information if they're afraid their searches will be exposed. LGBTQ+ youth won't explore their identity if they fear outing. A library without privacy is not a safe place. It's a trap.


The Question You Have to Answer

Stop for a second. Here's what you're actually deciding when you implement AI tools with patron behavior tracking:

"Is the vendor's profit more important than vulnerable people's safety?"

I know that sounds harsh. Vendors aren't evil. Your sales rep probably is a decent person. But the system they work for is designed to extract value from your institution. They collect data because data has value. They don't care if that data puts vulnerable people at risk.

Sometimes - maybe - the answer is: yes, we'll accept this risk because the tool is worth it. But make that decision with full knowledge of what you're risking. Not by assuming privacy safeguards exist when the vendor won't commit to them in writing.

And if the answer is no? Be honest about that. A recommendation engine is not worth the safety of domestic violence survivors, undocumented immigrants, or anyone else you serve.


See Also

Sources & Further Reading

For information about how these sources were selected and verified, see How I Research Library Tech.

  1. Electronic Frontier Foundation (2024). Library Privacy Toolkit: Protecting Patron Data in the Age of AI. Advocacy guide for library privacy. Retrieved from https://www.eff.org/issues/library-records-confidentiality
  2. American Library Association (2024). Library Confidentiality and Intellectual Freedom: Best Practices. Professional standards for patron privacy.
  3. U.S. Department of Homeland Security (2024). Immigration Enforcement and Community Trust. Report on ICE use of public records.
  4. Stanford Internet Observatory (2023). Platform Data Breaches and Vulnerable Communities: Risk Assessment Framework. Analysis of harm to marginalized groups.
  5. Narayanan, A. & Felten, E. (2014). No Silver Bullet: De-identification Still Doesn't Work. IEEE Security & Privacy. Journal article on de-identification failures.
  6. National LGBTQ Task Force (2024). Digital Safety for LGBTQ Youth: Privacy & Technology. Advocacy resource on digital safety.
  7. National Domestic Violence Hotline (2024). Tech Safety for Domestic Violence Survivors. Guide on technology risks for survivors.
  8. Immigration Advocates Network (2024). Know Your Rights: Digital Privacy & Immigration Enforcement. Legal resource on digital privacy risks.
  9. National Institute of Standards and Technology (2024). Framework for Improving Critical Infrastructure Cybersecurity: Privacy Profile. NIST cybersecurity standards for data protection.
  10. California Consumer Privacy Act (CCPA) 2024 Update. Rights and Obligations for Data Collectors. State privacy regulation with implications for libraries.