ChatGPT
By our AI Review Team
.
Last updated October 13, 2023
A powerful, at times risky chatbot for people 13+ that is best used for creativity, not facts
What is it?
ChatGPT, which stands for Chat Generative Pre-Trained Transformer (say that 10 times fast…) is a generative AI chatbot that generates text in response to a wide range of prompts or questions. For example, it can respond to a user in a way that feels like a conversation, or come up with an outline for an essay on the history of television.
ChatGPT was created by the company OpenAI and launched for the public in November 2022. ChatGPT is on its fourth-generation LLM, otherwise known as GPT-4. The free version of ChatGPT currently uses an earlier version (GPT-3.5). To access the more powerful GPT-4, users must upgrade to Plus, which costs $20/month. Users can access ChatGPT on Android and iPhone mobile apps, through a desktop web browser (including Google Chrome, Mozilla Firefox, Safari, Microsoft Edge, or Opera) and through a mobile web browser.
How it works
ChatGPT is a form of generative AI, which is an emerging field of artificial intelligence. Generative AI is defined by the ability of an AI system to create ("generate") content that is complex and coherent and original. For example, a generative AI model can create sophisticated writing or images. ChatGPT is a chatbot interface that essentially sits on top of a large language model (LLM). This underlying system is what makes ChatGPT so powerful and able to respond to many kinds of human input.
Large language models are sophisticated computer programs that are designed to generate human-like text. Essentially, when a human user inputs a prompt or question, an LLM quickly analyzes patterns from its training data to guess which words are most likely to come next. For example, when a user inputs "It was a dark and stormy," an LLM is very likely to generate the word "night" but not "algebra." LLMs are able to generate responses to a wide range of questions and prompts because they are trained on massive amounts of information scraped from the internet. In other words, a chatbot powered by an LLM is able to generate responses for many kinds of requests and topics because the LLM has likely seen things like that before. Importantly, LLMs cannot reason, think, feel, or problem-solve, and do not have an inherent sense of right, wrong, or truth.
Highlights
- It's best for fiction and creativity. While this is an oversimplification, you can think of ChatGPT like a giant auto-complete system—it is simply predicting the words that will most likely come next. An LLM has been trained on a massive amount of text, so that "auto-complete" has a lot to work with. When a generative AI chatbot is factually correct, that's because those responses are generated from accurate information commonly found on the internet. Because of the above, and just like with all generative AI chatbots, ChatGPT performs best with fiction, not facts. It can be fun for creative use cases, but should not be relied on for anything that depends on factual accuracy.
- It's up to date. Paid Plus and Enterprise users can use ChatGPT's browsing feature to access the internet in real time. This feature is not yet available to free users, and generative AI chatbots like ChatGPT should not be used as a search tool.
Harms and Ethical Risks
- LLMs can, and do, create harms, and use of them is inherently risky. ChatGPT can be an amazing tool when used responsibly. Knowing why it is so risky can help determine how best to use it. This starts with ChatGPT's training data. Any text that can be scraped from the internet could be included in this model. While the details on which corners of the internet have been scraped are unclear, OpenAI has shared that GPT-4 was developed using data that is publicly available on the internet, information that OpenAI licenses from third parties, and data provided by human trainers and user inputs. OpenAI also shares that it filtered this pre-training data to reduce "inappropriate erotic text content." But the internet also includes a vast range of racist and sexist writing, conspiracy theories, misinformation and disinformation, toxic language, insults, and stereotypes about other people. As it predicts words, a generative AI chatbot can repeat this language unless a company stops it from doing so. Importantly, these attempts to limit objectionable material are like Band-Aids: They don't address the root causes, they don't change the underlying training data, and they can only limit harmful content that's already known. We don't know what they don't cover until it surfaces, and there are no standard requirements for what they do cover. And like bandages, they aren't comprehensive and are easily breakable.
- ChatGPT's false information can shape our worldview. ChatGPT can generate or enable false information in a few ways: from "hallucinations"—an informal term used to describe the false content or claims that are often output by generative AI tools; by reproducing misinformation and disinformation; and by reinforcing unfair biases. Because OpenAI's attempts to limit these are brittle, false information is being generated at an alarming speed. As these AI systems grow, it may become increasingly difficult to separate fact from fiction. ChatGPT also adds users' inputs to its already skewed training data. While this helps ChatGPT improve, it also likely increases those skews. This is because today's ChatGPT users are an early-adopter subset of the internet-connected population, which as a whole overrepresents people in wealthier nations, as well as views from people who are wealthier, younger, and male. Combined, these forces carry an even greater risk—one that OpenAI expects to happen if not enough is done to address it—of AI systems to "reinforce entire ideologies, worldviews, truths and untruths, and to cement them or lock them in, foreclosing future contestation, reflection, and improvement." We need much stronger oversight and governance of AI to prevent this from happening.
Review team note: We cannot address the full scope of the risks of ChatGPT that OpenAI has publicly discussed. That is not a reflection on whether those risks matter.
Limitations
- Because ChatGPT isn't factually accurate by design, it can and does get things wrong. In OpenAI's own words, GPT-4 has a "tendency to make up facts, to double-down on incorrect information, and to perform tasks incorrectly." For example, when tested against a benchmark that is widely used to assess a model's factual accuracy, GPT-4's responses were 60% accurate and GPT-3.5's responses were ~48% accurate. Unfortunately, ChatGPT's inaccuracies can be hard to detect, as the model's responses can sound correct, even if they aren't. Importantly, OpenAI found that GPT-4 actually generates responses "in ways that are more convincing and believable than earlier GPT models." Any seemingly factual output needs to be checked—and this absolutely goes for any links, references, or citations too.
- Parental permission is required, but this isn't obvious. Educators who are using ChatGPT in their classrooms need to know that children must be age 13, and anyone under 18 must have a parent's or legal guardian's permission to use ChatGPT. OpenAI does not check if permission is in place.
- ChatGPT performs best in English.
Misuses
- OpenAI details misuses of all of its models, including ChatGPT, in a comprehensive Usage Policy.
- OpenAI's terms of service do not allow its use for children under age 13.
- Teens aged 13–17 are required to have parental permission to use it.
Common Sense AI Principles Assessment
Our assessment of how well this product aligns with each AI Principle .
Additional Resources