Abstract
Large Language Models (LLMs) are predictive probabilistic models capable of passing several professional tests at a level comparable to humans. However, these capabilities come with ethical concerns. Ethical oversights in several LLM-based products include: (i) a lack of content or source attribution, and (ii) a lack of transparency in what was used to train the model. This paper identifies four touchpoints where ethical safeguards can be applied to realize a more responsible AI in LLMs. The key finding is that applying safeguards before the training occurs aligns with established engineering practices of addressing issues at the source. However, this approach is currently shunned. Finally, historical parallels are drawn with the US automobile industry, which initially resisted safety regulations but later embraced them once consumer attitudes evolved.
| Original language | English |
|---|---|
| Pages (from-to) | 1-1 |
| Journal | IEEE Transactions on Technology and Society |
| DOIs | |
| Publication status | Published - May 21 2024 |
Fingerprint
Dive into the research topics of 'How to Regulate Large Language Models for Responsible AI'. Together they form a unique fingerprint.Cite this
- APA
- Standard
- Harvard
- Vancouver
- Author
- BIBTEX
- RIS