Stable Diffusion
By our AI Review Team
.
Last updated November 5, 2023
Powerful image generator can unleash creativity, but is wildly unsafe and perpetuates harm
We do not consider this a safe tool, so we won't link directly to it in this review. Why this matters.
What is it?
Stable Diffusion is a generative AI product created by Stability AI. It can create realistic images and art from a text-based description that can combine concepts, attributes, and styles. Stability AI's full suite of image editing tools offers users a sophisticated range of options: extending generated images beyond the original frame (outpainting), making authentic modifications to existing user-uploaded or AI-generated pictures, and incorporating or eliminating components while considering shadows, reflections, and textures (inpainting). Once users achieve the generated image they want, they can download and use it.
Stability AI released Stable Diffusion to the public in November 2022. It is powered by a massive data set of image-text pairs scraped from the internet. The data set includes a subset of 2.32 billion images that contain English text. It was created by LAION, which stands for "Large-scale Artificial Intelligence Open Network." LAION is a nonprofit organization that is funded in part by Stability AI.
Stable Diffusion can be accessed in three separate places hosted by Stability AI:
- Clipdrop, Stability AI's text-to-image editor, which has three pricing tiers: free, pro ($9/month), and API pricing, in which users purchase credits that are used to pay for the computing cost of each request.
- Dreamstudio, another image editor from Stability AI that extends beyond text-to-image prompting with inpainting, outpainting, and image-to-image generation, which requires users to purchase credits that are used to pay for the computing cost of each request. Currently, $10 equals 1,000 credits, which Stability AI notes is ~5,000 images.
- Stability.ai's developer platform, which—like Dreamstudio—requires users to purchase credits that are used to pay for the computing cost of each request. Currently, $10 equals 1,000 credits, which Stability AI notes is ~5,000 images
In addition, Stable Diffusion has made all model weights and code available. Anyone is able to access, download, and use the full model.
How it works
Stable Diffusion is a form of generative AI, which is an emerging field of artificial intelligence. Generative AI is defined by the ability of an AI system to create ("generate") content that is complex and coherent and original. For example, a generative AI model can create sophisticated writing or images.
Stable Diffusion uses a particular type of generative AI called "diffusion models," named for the process of diffusion to generate new content. Diffusion is a natural phenomenon you've likely experienced before. A good example of diffusion happens if you drop some food coloring into a glass of water. No matter where that food coloring starts, eventually it will spread throughout the entire glass and color the water in a uniform way. In the case of computer pixels, random motion of those pixels will always lead to "TV static." That is the image equivalent of food coloring creating a uniform color in a glass of water. A machine-learning diffusion model works by, oddly enough, destroying its training data by successively adding "TV static," and then reversing this to generate something new. They are capable of generating high-quality images with fine details and realistic textures.
Stable Diffusion combines a diffusion model with a text-to-image model. A text-to-image model is a machine learning algorithm that uses natural language processing (NLP), a field of AI that allows computers to understand and process human language. Stable Diffusion takes in a natural language input and produces an image that attempts to match the description.
Highlights
- Stable Diffusion has the potential to enable creativity and artistic expression, allow for visualization of new ideas, and create new concepts and campaigns.
- Stability AI suggests that the best uses of Stable Diffusion include: generation of artworks and use in design and other artistic processes; applications in educational or creative tools; research on generative models; safe deployment of models that have the potential to generate harmful content; and probing and understanding the limitations and biases of generative models.
Harms and Ethical Risks
- Stable Diffusion's "view" of the world can shape impressionable minds, and with little accountability. Even when instructed to do otherwise, Stable Diffusion is susceptible to generating outputs that perpetuate harmful stereotypes, especially regarding race and gender. We confirmed this repeatedly with our own testing. These behaviors reflect both the way in which the model was trained and—critically—the choice of the data set used to train it. LAION 5B, the data set that powers Stable Diffusion, is uncurated. This means that it contains every image found in the Common Crawl repository that has one or more text labels that would be usable for the image-text pairs that the machine learning model needs to match a user's input to images it can use to generate the result. While some filters have been applied, LAION notes that because Stable Diffusion is uncurated, the links that make up the data set "may lead to strongly discomforting and disturbing content for a human viewer." Based on LAION's own measurement, 2.9% of the 2.3 billion image-text pairs used by Stable Diffusion are "unsafe"—that is, roughly 68 million unsafe images. All of the technical documentation clearly states that this data set should be used only for research purposes. But Stable Diffusion is accessible to anyone, and Stability AI has made the model that powers it available for anyone to download and use for their own purposes. These propensities towards harm are frighteningly powerful. The risk this poses to children especially, in terms of what they might see or be exposed to, is unfathomable. What happens to our children when they are exposed to the worldview of a biased algorithm repeatedly and over time? What view of the world will they assume is "correct," and how will this inform their interactions with real people and society? Who is accountable for allowing this to happen?
- Inappropriate sexualized representations of women and girls harm all users. Despite many public failings, Stable Diffusion continues to easily produce inappropriately sexualized representations of women and girls, even with prompts seeking images of women professionals. This perpetuates harmful stereotypes, unfair bias, unrealistic ideals of women's beauty and "sexiness," and incorrect beliefs around intimacy for humans of all genders. Numerous studies have shown that greater exposure to images that promote the objectification of women adversely affects the mental and physical health of girls and women. Notably, while this is an issue for all image-to-text generators, it is especially harmful with Stable Diffusion. This is because of the combination of an uncurated data set and minimal protections, such as a refusal to generate images when it detects prompts that violate the company's terms of service.
- Stable Diffusion consistently and easily reinforces harmful stereotypes. While Stable Diffusion's July 2023 update aimed to prevent it from generating some of the most objectionable content, this remains a significant risk. Recent findings show continued reinforcement of harmful stereotypes, and the manner in which Stability AI has open-sourced the model allows anyone to remove those protections in new applications. A great resource for exploring this problem further can be found at Stable Bias. Relevant articles:
- Tiku, N., Schaul, K., & Chen, S.Y. (2023, Nov. 1). How AI is crafting a world where our worst stereotypes are realized. Washington Post.
- Crawford, A., & Smith, T. (2023, June 28). Illegal trade in AI child sex abuse images exposed. BBC.
- Harlan, E., & Brunner, K. (2023, June 7). We are all raw material for AI. BR24.
- Nicoletti, L., & Bass, D. (June 2023). Humans are biased. Generative AI is even worse. Bloomberg.
- Vincent, J. (2023, Jan. 16). I art tools Stable Diffusion and Midjourney targeted with copyright lawsuit. The Verge.
- Edwards, B. (2022, Sept. 21). Artist finds private medical record photos in popular AI training data set. Ars Technica.
- Wiggers, K. (2022, Aug. 24). Deepfakes for all: Uncensored AI art model prompts ethics questions. TechCrunch.
- Wiggers, K. (2022, Aug. 12). This startup is setting a DALL-E 2-like AI free, consequences be damned. TechCrunch. - Stable Diffusion's advanced inpainting and outpainting features present new risks. While innovative and useful in many contexts, the high degree of freedom to alter images means they can be used to perpetuate harms and falsehoods. Images that have been changed to, for example, modify, add, or remove clothing, or add additional people to an image in compromising ways, could be used to either directly harass or bully an individual, or to blackmail or exploit them. These features can also be used to create images that intentionally mislead and misinform others. For example, disinformation campaigns can remove objects or people from images or create images that stage false events.
- Tools like Stable Diffusion pave the path to misinformation and disinformation. As with all generative AI tools, Stable Diffusion can easily generate or enable false and harmful content, both by reinforcing unfair biases, and by generating images that intentionally mislead or misinform others. Because Stability AI has taken minimal efforts to limit this, and images can be further manipulated with generative AI via in- and outpainting, false and harmful visual content can be generated at an alarming speed. We have already seen this in action. As OpenAI has noted in the context of DALL-E, as image generation matures, it "leaves fewer traces and indicators that outputs are AI-generated, making it easier to mistake generated images for authentic ones and vice versa." In other words, as these AI systems grow, it may become increasingly difficult to separate fact from fiction. This "Liar's Dividend" could erode trust to the point where democracy or civic institutions are unable to function.
- The Stable Diffusion model is intended for research only, but Stability AI has made it available to everyone. Deep in its technical model card, Stability AI notes that Stable Diffusion is intended for research purposes only, and that "while the capabilities of image generation models are impressive, they can also reinforce or exacerbate social biases." Unfortunately, this information is currently nowhere to be found on Clipdrop or Dreamstudio, where Stable Diffusion is accessible to anyone.
Limitations
- We did not receive participatory disclosures from Stability AI for Stable Diffusion. This assessment is based on publicly available information, our own testing and our review process.
- Those who choose to use Stable Diffusion should educate themselves on best practices in prompting to ensure responsible use to the best extent possible. Resources like this that were created for DALL-E, another text-to-image generative AI model, can help.
Misuses
Stable Diffusion does have legal terms, but protections for children are unclear. One reason for this lack of clarity stems from the fact that Stable Diffusion can be accessed in three separate places hosted by Stability AI:
- Clipdrop, Stability AI's text-to-image editor, a simple interface more accessible to consumers. Clipdrop states in its Terms of Use that users are prohibited from downloading or producing content that, among other prohibited uses, infringes on "public order and morality." Children's rights are not specifically addressed in Clipdrop's terms. Clipdrop's terms state that minors must have permission from their legal representative to use this product.
- Dreamstudio, another image editor from Stability AI that extends beyond text-to-image prompting with inpainting, outpainting, and image-to-image generation. Dreamstudio's Terms of Service contain an expanded list of prohibited uses and introduce Community Guidelines, which note that "contributions must be safe, legal, and in accordance with these Terms." Dreamstudio's terms state that minors are prohibited from using this product.
- Stability.ai's developer platform, which has its own, more exhaustive Acceptable Use Policy. These terms specifically prohibit use of Stability Technology for, among other prohibited uses, "Exploitation or harm to children, including the solicitation, creation, acquisition, or dissemination of child exploitative content." Stability AI's terms state that minors are prohibited from using this product.
Because Stable Diffusion can be accessed from each of these tools, it is unclear which set of terms may be enforced, why there is a discrepancy between these terms, how these terms might be enforced, and by whom.
Common Sense AI Principles Assessment
Our assessment of how well this product aligns with each AI Principle .
Additional Resources
Edtech Ratings
Apps and websites for making posters and collages
Free Lessons
AI Literacy for Grades 6–12