Q. What are guardrails (in tools like ChatGPT)?

  Jan 07, 2025

Guardrails are built-in safeguards and content moderation techniques designed to filter out harmful, biased, or inappropriate content. Developers implement them to ensure that the AI model adheres to ethical and safety standards when responding.

Tools like ChatGPT rely on user feedback to improve. If you find harmful or incorrect content in a response, you can click the “thumbs down” icon and provide additional information. OpenAI will review these responses and add or adjust guardrails as needed.


Adapted from "FAQs about generative AI" by Nicole Hennig, University of Arizona Libraries. Licensed under CC BY 4.0.


View All Topics

VIEW ALL FAQs chevron_right

Contact Us

Contact Us

email

Email

Email us your research questions and we’ll respond within 24 hours

question_answer

Chat

Talk online to a research librarian 7 days / week

smartphone

Text

Send us your questions at 617-431-2427

call

Call

Call for info or research assistance at 617-353-2700

people

Meet

Make an appointment with a subject specialist librarian over Zoom