Ethics & limits
Epistemic, social, and environmental considerations
- Name and reason about the major ethical tensions in AI use for science.
- Identify the epistemic risks of outsourcing judgment to a model.
- Make grounded data-handling decisions using your institution’s policies.
Epistemic risks
- Homogenisation: everyone using the same model converges on the same framings, blind spots, and errors. The default LLM answer becomes the default scientific framing.
- Deskilling: offloading judgement tasks can erode the skills needed to catch the model when it’s wrong. The risk is highest for trainees who use AI before developing baseline competence.
- Calibration failure: LLMs present confident, fluent answers regardless of underlying certainty. Used uncritically, this corrodes the habits of calibrated scientific writing.
- Epistemic injustice: Kay, Kasirzadeh, and Mohamed (2024) argue that generative AI can amplify testimonial injustice (whose voices are heard) and hermeneutical injustice (whose conceptual frameworks are usable), particularly outside dominant languages and traditions.
Data, privacy, and consent
Read the syllabus data policy before any exercise with real data. A few clarifications:
- A “zero-retention” API tier is not the same as private. Most vendor “no-retention” tiers still retain data briefly for abuse monitoring and incident response. They are better than the consumer chat product, not equivalent to on-device.
- “De-identified” is not the same as unrestricted. De-identified clinical data may still be subject to HIPAA, IRB, and BAA terms, and to your institution’s contractual restrictions.
- For sensitive workloads, the safer options are institutional or on-prem deployments, vendor enterprise tiers with appropriate contracts (a signed BAA where applicable), or fully local inference (a local quantised model).
- Consent frameworks predate LLMs. Donor consent that allowed “research use” did not anticipate transmission to a third-party model vendor. When in doubt, ask the data steward, not the model.
Environmental and labour costs
Honestly stated:
- Pretraining a frontier model uses substantial energy. Per-query inference is small. Aggregate inference is not.
- Annotation, preference labelling, and content moderation depend on human labour, often outsourced under difficult conditions.
- Neither cost invalidates the technology. Both should shape how often and how thoughtlessly you reach for it.
Limits worth naming
- LLMs are not reliable sources of truth about the world.
- They are not reliable arbiters of quality, novelty, or significance.
- They do not “understand” causality, statistics, or biology in any robust sense, though they can produce outputs that look like they do, especially in the dominant subfields of their training distribution.
- They are not your peer reviewer. They are not your IRB. They are not your statistical consultant.
- A collaborator wants to list ChatGPT as a co-author on a manuscript because it “drafted half the introduction”. What is the convergent publisher position, and what is the underlying reason?
- Your institution has a “zero-retention” API tier with the model vendor. Can you safely paste de-identified clinical data into it? What questions would you ask first?
- Name one epistemic risk and one labour or environmental cost of frontier-model use, and explain why the two require different responses.
Answers: 1. AI cannot be listed as an author. Authorship requires accountability: the ability to respond to post-publication queries, accept responsibility for errors, and issue corrections. An AI system cannot be accountable. Disclosure of the AI’s contribution in the methods or acknowledgments is the right path, not authorship. 2. Probably not without further checks. “Zero-retention” tiers still typically retain briefly for abuse monitoring and are not equivalent to on-prem. De-identified clinical data may still be subject to HIPAA, IRB protocols, BAAs, and the original donor consent. Ask: does my institution have a signed BAA with this vendor? Does the IRB protocol cover transmission to a third-party model? What did the donor consent permit? 3. Epistemic: homogenisation. Everyone using the same model converges on the same framings and blind spots, eroding scientific diversity. Response: name AI use, vary tools, keep your own framing. Labour and environmental: annotation work performed under difficult outsourced conditions, and aggregate inference energy that is not trivial. Response: use AI deliberately, not thoughtlessly, and favour vendors with stronger labour and disclosure practices.
Further reading
- Kay, J., Kasirzadeh, A., & Mohamed, S. (2024). Epistemic injustice in generative AI. AAAI/ACM AIES.
- ICMJE Recommendations, current online version, on AI in authorship.
- Nature editorial policies on AI, current online version.
- Science family editorial AI guidelines, current online version.