Guardrails are built-in safeguards and content moderation techniques designed to filter out harmful, biased, or inappropriate content. Developers implement them to ensure that the AI model adheres to ethical and safety standards when responding.
Tools like ChatGPT rely on user feedback to improve. If you find harmful or incorrect content in a response, you can click the “thumbs down” icon and provide additional information. OpenAI will review these responses and add or adjust guardrails as needed.
Adapted from "FAQs about generative AI" by Nicole Hennig, University of Arizona Libraries. Licensed under CC BY 4.0.