1 of 117

GenAI & Ethics:

Investigating ChatGPT, Gemini, & Copilot

Torrey Trust

Professor of Learning Technology �University of Massachusetts Amherst

torrey@umass.edu | www.torreytrust.com

2 of 117

Using This Slide Deck

This slide deck is licensed under CC BY NC 4.0, meaning that you can freely use, remix, and share it as long as you give attribution and do not use it for commercial purposes.

This means that you can do the following without needing permission:

  • Share these slides with others.
  • Show this entire slideshow in your class/workshop.
  • Select some of these slides to include in your own presentation.
  • Remix the slides.

As long as you give credit and do not use these slides to make money (e.g., including the slides in a presentation in which you are getting paid).

To give credit, use the following Creative Commons attribution:

"AI & Ethics" slide deck by Torrey Trust, Ph.D. is licensed under CC BY NC 4.0.

3 of 117

Table of Contents

4 of 117

GenAI Chatbots

5 of 117

GenAI Chatbots: ChatGPT (by OpenAI)

6 of 117

ChatGPT, a large language model developed by OpenAI, is a machine learning model that is able to generate human-like text based on the input provided.

The free version of ChatGPT runs on GPT 3.5. This version does not have access to the Internet.

The paid version ($20+/month) of ChatGPT runs on GPT 4.0 and allows for more advanced functionality like designing GPTs or using GPTs created by others. This version has access to current data on the Internet.

7 of 117

ChatGPT was launched in November 2022, and reached 100 million users by the start of 2024.

8 of 117

ChatGPT has already been integrated into many different fields and careers.

9 of 117

Research has shown that users of ChatGPT complete more tasks, finish tasks more quickly, and have higher-quality results than non-users.

10 of 117

11 of 117

GenAI Chatbots: Copilot (by Microsoft)

12 of 117

Copilot, a large language model developed by Microsoft, is a machine learning model that is able to generate human-like text based on the input provided.

While Copilot does not have direct Internet access, its responses include links to Internet-based resources to verify the accuracy and credibility of the information provided.

13 of 117

Script on screen:

“They say I will never open my own business. Or get my degree. They say I will never make my movie. Or build something. They say I’m too old to learn something new. Too young to change the world. But I say, Watch Me.”

Writing on MS Copilot then says: “Quiz me in organic chemistry.” MS Copilot then generates a question about an organic molecular formula, providing multiple choice options. Commercial ends with MS Copilot being asked “Can you help me” and it responds “Yes, I can help.” Screen script then says “Copilot, your everyday AI companion. Anyone. Anywhere. Any device.”

14 of 117

GenAI Chatbots: Gemini (by Google)

15 of 117

Gemini, a large language model developed by Google, is a machine learning model that is able to generate human-like text based on the input provided.

Gemini has access to a massive dataset of text and code that is constantly being updated, which allows it to stay current on information. Its responses often include links to Internet-based resources.

Because Gemini is a Google tool, it can be used to summarize YouTube (owned by Google) videos.

16 of 117

17 of 117

18 of 117

Data & Privacy

19 of 117

OpenAI Requires ChatGPT Users to be 13 Years or Older

The use of ChatGPT by individuals under 13 years old would violate the Children’s Online Privacy Protection Act (COPPA), since OpenAI collects a lot of user data!

20 of 117

Use of ChatGPT by 13-18 year olds requires parental permission

21 of 117

OpenAI collects a LOT of user data, including user’s time zone, country, dates and times of access, type of computer/device you’re using, computer connection!

Here’s an example of the type of data it might collect from a user: https://webkay.robinlinus.com/

22 of 117

OpenAI collects any information you input as data, so if you write a prompt including any personally identifiable information about your students, it keeps that data; and is a possible FERPA violation.

Likewise, if you ask a student to use ChatGPT to revise a college admissions essay that includes information about a trauma they experienced, OpenAI collects and keeps that data!

23 of 117

Quite simply, they use your data to make more money (e.g., improve their products)!

You can opt out of having your data used to improve the way they train their model!

24 of 117

Way down at the bottom of their Privacy Policy, they also note that they are collecting Geolocation data!

25 of 117

Want to learn more (and quite possibly be scared about) the collection of geolocation data?

Check out this New York Times Interactive: “Your Apps Know Where You Were Last Night, and They’re Not Keeping It Secret

And, read “Google tracked his bike ride past a burglarized home. That made him a suspect.

26 of 117

27 of 117

Gemini Requires Gemini Users to be 13 Years or Older

Google Workspace for Education account holders under 18 years old will not be able to use Gemini.

Students older than 13 would need to use their personal Gmail account to access Gemini.

28 of 117

Google collects a LOT of user data, including user’s “conversations” with the chatbot, usage data, location data, and feedback.

29 of 117

If you are 18 years or older, Google stores your activity (e.g., any “conversations” you have with Gemini) for up to 18 months. They also collect your location data, IP address, and home/work address.

30 of 117

Google collects any information you input as data, so if you write a prompt including any personally identifiable information about your students, it keeps that data; and is a possible FERPA violation.

Likewise, if you ask a student to use Gemini to revise a college admissions essay that includes information about a trauma they experienced, Google collects and keeps that data!

31 of 117

Quite simply, Google uses your data to make more money (e.g., improve their products).

You can change your location permissions for Google.

32 of 117

Microsoft Requires Copilot Users to be 13 Years or Older

33 of 117

Microsoft seems to have more data and privacy protections in place for children and young people.

34 of 117

Copilot in Bing has data retention and deletion policies…

That means you can better control your data!

35 of 117

Microsoft does not seem to collection location data from users.

36 of 117

Any prompts that you input into Copilot or anything you create with Copilot is immediately owned by Microsoft.

They can use your prompts and creations (without paying you) however they see fit (aka to make more money!)

So, if your students come up with a super amazing prompt that turns Copilot into a tutor for your class…Microsoft will own that prompt and could use/sell/share it!

37 of 117

Privacy & Data Overview

  • ChatGPT requires parental permission for 13-18 year old users, Gemini and Copilot do not.
  • ChatGPT and Gemini can give away any data collected to “affiliates,” including, if requested, to federal authorities.
  • Microsoft & Google have more data privacy protections for users (thank you GDPR!)
  • Google tracks user location, OpenAI collects IP address, Microsoft CoPilot doesn’t seem to collect any location data.
  • Don’t let students put in any sensitive or identifying information into any of these tools!
  • Don’t put any sensitive information in these tools (e.g., asking ChatGPT to write an email to a student about their grade - this is FERPA violation).
  • Any information input into these tools (e.g., any prompts they write) is data that can be used by the companies that made the tools.

38 of 117

How to Protect Student Data & Privacy

  • Use Gemini or CoPilot instead of ChatGPT, since Gemini and Copilot have stronger data protections due to the GDPR.
  • Ask students to use only one tool (the more tools they use, the more data is collected about them).
  • Use the AI tool only on a teacher computer/account.
    • Note: Sharing your login with students so they can access ChatGPT is a violation of OpenAI’s terms of use (“You may not share your account credentials or make your account available to anyone else and are responsible for all activities that occur under your account.”)
  • Ask students to only use the AI tools during class time (e.g., this protects their location data; compared to using these tools at home for homework).
  • Teach students about the privacy policies and terms of use of these tools (they may not know that what they type into a prompt is collected and stored).

39 of 117

Bias

40 of 117

41 of 117

42 of 117

43 of 117

44 of 117

45 of 117

This article highlights multiple types of bias, including machine/algorithmic bias, availability bias, representation bias, historical bias, selection bias, group attribution bias, contextual bias, linguistic bias, anchoring bias, automation bias, and confirmation bias.

46 of 117

GenAI tools are often trained on English-only data scraped from the Internet; therefore, their output is biased toward presenting American-centric and Westernized views.

47 of 117

48 of 117

49 of 117

50 of 117

51 of 117

OpenAI acknowledges that the biases in ChatGPT can negatively impact students, especially, for example, when using the tool to provide feedback on work by English language learners.

52 of 117

Considerations for Educators

Engage students in investigating how generative AI tools are designed (e.g., What data they are trained on? Why that data was selected? How will that data produce biased output?).

Encourage students to reflect upon how biased AI output can shape thinking, learning, education, and society.

Bonus: Ask students to design a code of ethics for AI developers in order to reduce the harms done by biased AI output.

Resources:

53 of 117

Hallucinations

54 of 117

55 of 117

OpenAI states that ChatGPT can give incorrect and misleading information. It can also make up things!

56 of 117

OpenAI’s Terms of Use states that when you use ChatGPT you understand and agree that the output may not always be accurate and that it should not be relied on as a sole source of truth.

57 of 117

58 of 117

59 of 117

Google acknowledges that “Gemini will make mistakes.”

Gemini has a “double-check feature”...but it too can make mistakes.

60 of 117

61 of 117

Google provides these disclaimers in its “Generative AI Additional Terms of Service.”

62 of 117

Microsoft downplayed the fact that Copilot can be wrong.

63 of 117

Copilot often (but not always) provides in-text links to sources to verify information.

64 of 117

Considerations for Educators

Teach students how to critically evaluate the output of generative AI chatbots; and not to take what these tools produce at face value!

Resources:

Readings:

65 of 117

Academic Integrity

66 of 117

With the ability to generate human-like text, generative AI chatbots have raised alarms regarding cheating and academic integrity

67 of 117

This recent study found that…

68 of 117

While another recent study found that…

69 of 117

Interestingly…

70 of 117

Even still…students need to learn when it is okay to use generative AI chatbots and when it is not okay, or else they might end up like…

71 of 117

Did you know that…

Representing output from ChatGPT as human-generated (when it was not) is not only an academic integrity issue, it is a violation of OpenAI’s Terms of Use.

72 of 117

Considerations for Educators

Middle and high school students might not have ever read their school’s or district’s Academic Honesty policies.

College students often gloss over the boilerplate “academic integrity” statement in a syllabus.

Potential Steps to Take:

  • Update/add to your course academic integrity policy in your syllabus to include what role AI technologies should and should not play and then ask students to collaboratively annotate the policy and offer their suggestions.
  • Invite students to co-design the academic integrity policy for your course (maybe they want to use AI chatbots for helping with their writing…Or, maybe they don’t want their peers to use AI chatbots because that provides an advantage to those who use the tools!).
  • Provide time in class for students to discuss the academic integrity policy.

73 of 117

Tips for (Re)designing Your Academic Integrity Syllabus Policy

  • Define what you mean by AI (e.g., Grammarly? ChatGPT? Google Docs Autocomplete?)
  • Be specific about when students can and cannot use AI:
    • When is the use of AI allowed? (e.g., for brainstorming? For a specific assignment? For improving writing quality?)
    • When is it not allowed? (e.g., for doing students’ work for them)
    • Does informing the instructor about the use of AI make its use allowable?
    • NOTE: If you ban AI for your entire course or certain assignments, consider who that might privilege, and who that might negatively impact (e.g., English language learners, Students with communication disabilities, and others who rely on these tools to support their writing).
  • Explain why the use of AI is not allowed (e.g., “writing helps improve and deepen thinking,” “writing makes your thinking visible to me,” “writing is an important 21st century skill”; see Terada, 2021)
  • Be transparent about how you plan to identify AI-generated texts:
    • Will you be using an AI text detector? (If so, read this first!)
    • What will happen if one of these tools flags student work as AI-generated?

74 of 117

Reflect - This author (winner of a prestigious writing award) used ChatGPT to write 5% of her book… would you let your students submit a paper where 5% of it was written by AI?

75 of 117

Resources for Educators

76 of 117

Copyright & Intellectual Property

77 of 117

78 of 117

Several authors are suing OpenAI for using their copyrighted works to train ChatGPT.

79 of 117

The New York Times is suing OpenAI and Microsoft for using its articles to train their AI tools.

80 of 117

Was it legal for OpenAI to scrape public, and often copyrighted, data from the Internet for free to train their tool?

Also, who owns the copyright of AI-generated work. If AI generates a new idea for a life-saving invention, does the person who wrote the prompt get the copyright/patent? Or OpenAI?

81 of 117

Considerations for Educators

Many academic integrity policies state that it is okay for students to use text generated from AI, as “long as they cite it.”

But, should students really be citing AI-generated text, when AI tools were designed by stealing copyrighted text from the Internet? Or, should students go to the original source and cite that?

This might be a conversation worth having with your students!

Resources:

82 of 117

Human Labor

83 of 117

OpenAI can use any data it collects from you to improve its services; thus helping it make more money (aka you are providing free labor!).

84 of 117

OpenAI states that you will not be given any compensation for providing feedback on the quality of ChatGPT’s output (aka you are providing free labor!).

85 of 117

Google can use any data it collects from you to improve its services; thus helping it make more money (aka you are providing free labor!).

86 of 117

Google states that it benefits from your feedback and data (aka you are providing free labor!).

87 of 117

Any prompts that you input into Copilot or anything you create with Copilot is immediately owned by Microsoft.

They can use your prompts and creations (without paying you) however they see fit (aka you are providing free labor!).

88 of 117

Many companies, including OpenAI, exploit human workers to review and train data for their AI technologies.

89 of 117

90 of 117

Considerations for Educators

Engage students in a conversation about whether they feel it is ethical for companies to use their data to make more money.

Encourage students to investigate the exploitation of data and human labor to improve AI technologies and make AI companies more money.

Resources:

91 of 117

Environmental Impact

92 of 117

93 of 117

94 of 117

95 of 117

96 of 117

97 of 117

98 of 117

Considerations for Educators

Encourage students to investigate the environmental cost of the design and use of generative AI chatbots.

Bonus: Ask them to identify ways to reduce the environmental impact of these technologies.

Resources:

99 of 117

Spreading Misinformation

100 of 117

This article examines how AI has made it easy for anyone to rapidly generate misinformation; and this can be very problematic leading up to the 2024 elections.

101 of 117

“In just 65 minutes and with basic prompting, ChatGPT produced 102 blog articles containing more than 17,000 words of disinformation” (DePeau-Wilson, 2023, para. 2).

102 of 117

NewsGuard is tracking AI-generated news and information websites that spread misinformation…to date, they’ve already found 725!

103 of 117

NOTE: “Bard” is now “Gemini.”

104 of 117

AI-generated misinformation is the world’s biggest short term threat!

105 of 117

Using Gemini to produce false or misleading information is not allowed, per the “Generative AI Prohibited Use Policy.”

106 of 117

Using ChatGPT to produce false or misleading information is not allowed, per the OpenAI Usage Policies.

107 of 117

Considerations for Educators

Help your students learn how to identify misinformation and combat the spread of misinformation

Because, the ability “to discern what is and is not A.I.-generated will be one of the most important skills we learn in the 21st century” (Marie, 2024, para.3).

Resources:

Readings:

108 of 117

The AI Digital Divide

109 of 117

The Digital Divide

There’s a major gap between people who can access and use digital technology and those who can’t. This is called the digital divide, and it’s getting worse as 3.7 billion people across the globe remain unconnected” (Connecting the Unconnected, 2024 para. 1).

There are different types of divides:

  • Access Divide – This refers to the difference between those who have access to technology and those who do not.
    • For example, students who have high-speed Internet access at home can more easily use AI tools than those who have limited or no Internet access at home. Students who can afford upgraded versions of AI tools (e.g., ChatGPT Plus) will have access to better features and functionality than those who cannot.
  • Usage Divide – This refers to the difference between those who know how to use technology and those who do not.
    • For example, let’s say that all students are given a laptop at school. The students who have family members and teachers who can show them how to use laptops to access generative AI tools for for thinking, communication, and learning will be at more of an advantage than those who do not.

110 of 117

This article highlights a third type of gap - quality of use!

111 of 117

112 of 117

Usage divide

113 of 117

Usage varies depending on ethnicity and gender!

114 of 117

Usage Divide by academic performance level.

115 of 117

While there are more than 7,000 languages spoken worldwide, generative AI large language models are often trained on just a few “standard” languages.

This creates a quality of use divide between those who speak the languages the AI tools were trained on and those who don’t.

116 of 117

This article focuses on the access divide.

117 of 117

Considerations for Educators

How might the digital divide affect your students?

  • Do they all have access to high-speed reliable Internet, and high quality devices, at home?
  • Do they have money to afford upgraded versions of AI?
  • Do they have family members who can teach them how to use AI?

How might you work to close the digital divide for your students?

  • Could you provide them with learning activities that incorporate the use of AI to help your students develop their AI literacy?
  • Could you incorporate learning activities that encourage a critical interrogation of AI (e.g., exploring the topics in these slides) so that all your students can learn how to make informed decisions about its use in their futures?

How might your students work on closing the digital divide in their school? Community? State? Country?

Resources: