AI Guidelines
Artificial Intelligence Guidelines
Purpose
Generative artificial intelligence (AI) describes algorithms that can be used to create new content, including audio, code, images, text, simulations, and videos. The rise in capabilities and ease of access of ChatGPT and other AI services provides exciting potential as a tool for research, programming, data processing and other applications. However, when using publicly available AI services, as with any technology, Clemson University employees, students and affiliates are obligated to protect and preserve the data we use at Clemson University every day.
This page is provided as a resource on University policies and responsibilities for University employees, students and affiliates as they explore these new technologies and their capabilities. This page will be regularly updated as additional guidance and regulations are developed.
AI Risks
- Entered data is retained and used to train models which can lead to an exposure of personal information or a data breach.
- Generative AI tools may produce erroneous responses that seem credible, sometimes referred to as “hallucinations”.
- Generative AI systems could be trained on copyrighted, proprietary, or sensitive data, without the owner’s or subject’s knowledge or consent.
AI Best Practices
- Enter only public data into an AI system unless you are certain it is a closed system that is approved for University use.
- Opt out of sharing data for AI learning whenever possible. Approved technology solutions may integrate new AI tools and features without your consent. If you become aware of these integrations, contact CCIT at ithelp@clemson.edu.
- Verify any results through authoritative sources.
- Consider legal, regulatory, ethical obligations and Clemson’s data classifications when using AI.
- Be transparent in disclosing and citing the use of AI tools.
- Be aware of the use of AI in virtual meetings or chats in which you engage, so as to avoid sharing personal or institutional information with outside parties.
AI Note Takers
- Your Clemson Zoom account (AI Companion) is approved for Public, Internal Use, Confidential and Restricted data, such as PHI and FERPA. However, third-party apps are NOT covered and should NEVER be used with sensitive data.
- If you have installed a third-party app or bot and intend to join a Zoom meeting where sensitive data will be discussed, remove the app from your account prior to joining.
- To prevent bots from joining your meeting:
Enable the Zoom Waiting Room feature for all meetings you host. Enabling the Waiting Room allows you to view each participant who attempts to join the meeting and only admit those who should be there. Bots will join the Waiting Room just as any other participant, and you can choose to remove them. The bot will present similar to “Mary’s Otter.ai Bot” or “John’s Firefly.ai Bot” in the waiting room.
- If your meeting is exclusively or mostly Clemson users, it’s recommended that you require participants to authenticate in order to join. Because the third-party bots cannot authenticate through a Clemson Zoom account, they will be blocked from joining.
- When attending a Zoom meeting:
Upon entering a Zoom meeting, ask the host if the meeting is being recorded or if live transcription is enabled. If the subject of the meeting is sensitive or includes other than Public information, request that the meeting not be recorded or leave the meeting.
Related University Policies and Guidance
Clemson University has multiple policies that help protect University data. University employees, students and affiliates must not enter Internal Use, Confidential, or Restricted institutional data into publicly available generative AI tools. This includes details like student information, personnel records, confidential University information from contracts or grants, and any proprietary or non-public intellectual property. Make sure that the information you submit is Public and doesn’t contain any personally identifiable or sensitive data.
Below are some of the relevant policies and standards that can help ensure that privacy and security are maintained, and guide decision making.
Policy and Guidance | Key points relevant to AI |
---|---|
Acceptable Use of IT Resources Policy |
|
Data Classification Policy |
|
IT Vendor Management Policy |
|
FERPA |
|
Academic Catalog |
|
Research Misconduct Policy |
|
Information Security Policy |
|
Need help?
- Submit an assessment to the IT Vendor Management Program Team to evaluate the AI service.
- Consult your Data Steward.
- Report information security incidents at ithelp@clemson.edu.
Other Resources
Some of these resources are provided for informational use only and may not reflect Clemson University’s regulations or recommendations.
Title | Key points relevant to AI |
---|---|
Clemson University Libraries – Artificial Intelligence in the Classroom | This guide caters to teaching faculty and librarians, offering insights into AI’s core concepts, practical applications, and ethical considerations in education. |
Clemson University Marketing & Communications – AI Guidelines | The University’s philosophy on the application of AI for Marketing and Communications |
Educause – Artificial Intelligence Resources | A compilation of AI resources offered by Educause. |
GenAI glossary | A glossary of generative AI terms |
NIST Trustworthy and Responsible AI Resource Center | A framework articulating characteristics of trustworthy AI and approaches for addressing them. |
Massachusetts Institute of Technology AI Risk Repository | A comprehensive living database of over 1000 AI risks categorized by their cause and risk domain. |
MalwareBytes AI Security Risks | A review of AI-related definitions and risks. |
A fit for purpose and borderless European Artificial Intelligence Regulation | An overview of the European Union Intelligence Act Risk levels. |
Future of Privacy Forum – Minding Mindful Machines: AI Agents and Data Protection Considerations | Characteristics of the newest AI agents and the data protection considerations to be mindful of when designing and deploying these systems. |
New York University Center for Responsible AI – The Algorithmic Transparency Playbook, Center for Responsible AI | This course explains algorithmic transparency and how to move towards having more open and accountable systems. The course includes a case study game exploring the tension between different key stakeholders vying for and against algorithmic transparency. |
OWASP Top 10 – LLM Applications Cybersecurity and Governance Checklist v1.1 | This checklist is for leaders across executive, tech, cybersecurity, privacy, compliance, and legal areas, as well as DevSecOps, MLSecOps, and Cybersecurity teams and defenders. |
OWASP Top 10 – LLM and Generative AI Security Center of Excellence Guide | Document for CISO security teams and cross-functional leadership to gain an understanding of the best practices framework to assist in educating teams on implementing their center of excellence for LLM and Generative AI Application security and adoption. |
Responsible Division
CCIT and University Compliance & Ethics
Reviewed Date
March 10, 2025