“Algorithmic transparency” sounds like a solution to Big Tech’s power. Lawmakers propose it. Advocates demand it. Companies promise it. But what does it actually mean? And would transparency alone fix the problems algorithms create?
The answer is more complicated than most transparency advocates acknowledge.
Transparency isn’t one thing
When people demand algorithmic transparency, they usually mean one of several different things:
Input transparency: What data does the algorithm use? Does it consider race, gender, or other protected characteristics? Does it use personal information users didn’t knowingly provide?
Process transparency: How does the algorithm work? What weights does it assign to different factors? What mathematical operations transform inputs into outputs?
Output transparency: What did the algorithm decide? Why was a specific person denied a loan, shown certain content, or flagged for investigation?
Audit transparency: Can independent researchers access the algorithm to test for bias, errors, or manipulation?
These are related but distinct concepts. A system can be transparent in one dimension but opaque in others.
Trade secrets and competitive advantage
Companies resist transparency partly because algorithms represent competitive advantages. Facebook’s News Feed algorithm, Google’s search ranking, Netflix’s recommendation system—these are valuable precisely because competitors don’t know how they work.
Requiring full disclosure would eliminate this advantage. If every recommendation algorithm were public, companies would quickly copy the best features, reducing innovation incentives.
This tension is real. We want companies to compete on quality (which requires innovation), but we also want accountability (which requires transparency). Balancing these goals is harder than “just make it transparent” suggests.
Understanding doesn’t equal accountability
Even if algorithms were fully transparent, most people couldn’t understand them. Modern machine learning systems involve billions of parameters, complex mathematical transformations, and emergent behaviors that even their creators don’t fully comprehend.
Publishing the algorithm doesn’t help if understanding requires a PhD in computer science and access to massive datasets for testing.
This is different from traditional transparency. When a government agency publishes its rules, citizens can understand them. “To qualify for this benefit, your income must be below $X” is comprehensible. An algorithm with 10 billion parameters isn’t.
The “black box” problem
Machine learning algorithms learn patterns from data without being explicitly programmed. Even their creators often can’t explain why the algorithm makes specific decisions.
You can know the training data, the architecture, and the mathematical operations—and still not understand why the algorithm flagged one loan application as high-risk but approved another nearly identical application.
This “black box” problem means full transparency might still not provide meaningful explanations for individual outcomes. The algorithm itself doesn’t “know” why in any human-interpretable sense.
Gaming and manipulation
Publishing how algorithms work enables gaming. If everyone knows Facebook prioritizes engagement, content creators optimize for engagement over accuracy. If loan algorithms are public, applicants manipulate applications to maximize approval odds.
This isn’t hypothetical. SEO (search engine optimization) exists because Google’s algorithm is partially known. People game it constantly. Full transparency would accelerate gaming, potentially making algorithms less accurate and more manipulable.
Bad actors benefit most from transparency. Foreign influence operations, spammers, and scammers would use algorithmic knowledge to evade detection and amplify harmful content.
What useful transparency looks like
Effective algorithmic accountability requires specific, targeted transparency:
Input audits: Independent researchers should verify that algorithms don’t use prohibited inputs like race or gender (except where legally permitted for remedial purposes).
Bias testing: Algorithms should be tested for disparate impact on protected groups, even if individual inputs seem neutral.
Output explanations: When algorithms make consequential decisions (loan denials, hiring, criminal sentencing), affected people deserve explanations in plain language.
Adverse outcome reporting: Companies should report aggregate outcomes (approval rates by demographic group, error rates, false positive/negative rates) without revealing proprietary details.
Regulatory access: Government regulators should have access to algorithms for enforcement purposes, with confidentiality protections to prevent disclosure to competitors.
Academic research access: Vetted researchers should be able to audit algorithms for bias and accuracy without full public disclosure.
The limits of transparency alone
Even perfect transparency wouldn’t solve algorithmic harms. Knowing how an algorithm works doesn’t stop it from:
- Optimizing for engagement in ways that promote misinformation
- Perpetuating historical biases present in training data
- Making errors that harm individuals
- Creating filter bubbles and echo chambers
These problems require regulation, not just transparency. Algorithms might need to be prohibited from certain uses, required to meet accuracy standards, or banned from using certain inputs—regardless of how transparent the process is.
Privacy vs. transparency tension
Full algorithmic transparency conflicts with privacy. To verify that an algorithm works correctly, auditors need access to the same data the algorithm uses—which often includes sensitive personal information.
Publishing detailed algorithm explanations can reveal private information about training data. If an algorithm makes decisions based on your purchase history, social connections, and browsing behavior, explaining its decision about you potentially reveals that private data.
Balancing transparency and privacy requires careful design: allowing audits without exposing individual data, providing explanations without revealing all inputs.
Regulatory approaches that work better
Instead of demanding full transparency, regulation should focus on outcomes:
Performance standards: Algorithms must meet minimum accuracy requirements. Loan algorithms can’t have false rejection rates above X%. Hiring algorithms must not discriminate.
Impact assessments: Before deploying consequential algorithms, companies must assess potential disparate impacts and take corrective measures.
Appeals and corrections: People affected by algorithmic decisions must have meaningful ways to challenge errors and get human review.
Prohibited uses: Some applications (real-time facial recognition surveillance, social credit scoring) might be banned regardless of transparency.
Auditing regimes: Instead of public transparency, require independent expert audits with results reported to regulators.
The false promise of “sunlight as disinfectant”
The metaphor “sunlight is the best disinfectant” works for traditional governance. Public scrutiny of government decisions creates accountability.
But algorithms aren’t laws or regulations. They’re mathematical systems too complex for public scrutiny to meaningfully constrain. Sunlight doesn’t disinfect what sunlight can’t illuminate.
Effective accountability requires expert oversight, clear standards, meaningful explanations for affected individuals, and consequences for harmful outcomes—not just public disclosure of incomprehensible code.
What to demand instead
Rather than generic “algorithmic transparency,” advocate for:
- Explainability: People affected by decisions deserve understandable reasons
- Auditability: Independent experts must be able to test for bias and errors
- Accountability: Companies must face consequences for algorithmic harms
- Standards: Minimum performance requirements for accuracy and fairness
- Redress: Meaningful ways to challenge algorithmic decisions
Transparency is a tool, not a solution. Used correctly—targeted, expert-driven, focused on accountability rather than publicity—it can help. But transparency alone won’t make algorithms fair, accurate, or aligned with public interest.
The goal isn’t to understand every line of code. It’s to ensure algorithms serve people rather than exploit them, regardless of whether the public can comprehend their inner workings.
Sources
- NIST: AI Risk Management Framework (AI RMF 1.0). https://www.nist.gov/itl/ai-risk-management-framework
- NIST (PDF): AI RMF 1.0. https://nvlpubs.nist.gov/nistpubs/ai/nist.ai.100-1.pdf
- FTC: AI Companies, uphold your privacy and confidentiality commitments (Jan 2024). https://www.ftc.gov/policy/advocacy-research/tech-at-ftc/2024/01/ai-companies-uphold-your-privacy-confidentiality-commitments
- FTC: FTC announces crackdown on deceptive AI claims and schemes (Sep 2024). https://www.ftc.gov/news-events/news/press-releases/2024/09/ftc-announces-crackdown-deceptive-ai-claims-schemes
- EU: AI Act policy page (high-level overview). https://digital-strategy.ec.europa.eu/en/policies/regulatory-framework-ai
- EU AI Act (commentary mirror of Article 13, transparency for high-risk systems). https://artificialintelligenceact.eu/article/13/
- arXiv: Explainable Artificial Intelligence: A Survey of Needs, Techniques, Applications, and Future Direction (2024). https://arxiv.org/abs/2409.00265
Discover more from The Rational Moderate
Subscribe to get the latest posts sent to your email.
