Abstract
As large language models (LLMs) shape AI development, ensuring ethical prompt recommendations is crucial. LLMs offer innovation but risk bias, fairness issues, and accountability concerns. Traditional oversight methods struggle with scalability, necessitating dynamic solutions. This paper proposes using collaborative filtering, a technique from recommendation systems, to enhance ethical prompt selection. By leveraging user interactions, it promotes ethical guidelines while reducing bias. Contributions include a synthetic dataset for prompt recommendations and the application of collaborative filtering. The work also tackles challenges in ethical AI, such as bias mitigation, transparency, and preventing unethical prompt engineering.
| Original language | English |
|---|---|
| Number of pages | 12 |
| Journal | International Journal of Parallel, Emergent and Distributed Systems |
| DOIs | |
| Publication status | Published - 14 Oct 2025 |
Bibliographical note
Publisher Copyright:© 2025 The Author(s). Published by Informa UK Limited, trading as Taylor & Francis Group.
Keywords
- large language models
- prompts
- recommender systems
- prompt recommendations
- ethical AI