Ren, S. & Wierman, A. (2024, July 15). The uneven distribution of AI's environmental impacts. Harvard Business Review. The Uneven Distribution of AI’s Environmental Impacts (hbr.org)
Microsoft Bing. (2024). CoPilot (Oct 21 version) [Large language model]. https://www.bing.com
AI Hallucinations occur when generative AI produces wrong responses to a prompt or query. This could involve:
IBM describes an AI hallucination as when a large language model (LLM) - often a chatbot - perceives patterns or objects that are nonexistent or imperceptible to human observers, creating outputs that are nonsensical or altogether inaccurate. Image by Andrew Martin from Pixabay
Generative AI is great at pattern detection and mimicry. It is trained to make predictions and then generate a reponse to a prompt, based off its training data. Since the internet is filled with people making unfounded and outrageous claims, confidently stating falsehoods, and making things up, it is no surprise that AI trained on the internet would do the same things (Morris, 2024).
IBM. (n.d.). AI hallucinations. https://www.ibm.com/topics/ai-hallucinations
Morris, S. (2024). Exploring AI with critical information literacy. ALA Learning.
Weise, K., Metz, C., Lohr, S., & Grant, N. (2023, May 8). When Chatbots “Hallucinate.” New York Times, 172(59782), B4.
The main concerns about AI and copyright revolve around issues of ownership, infringement, and the use of copyrighted material to train AI models. Here are some key concerns:
1. Use of Copyrighted Material for AI Training: Many AI models, especially generative ones, are trained on vast amounts of data that may include copyrighted works. The unauthorized use of such material in training can raise legal and ethical questions. Creators may argue that AI companies are benefiting from their work without permission or compensation.
Image by Mohamed Hassan from Pixabay
2. Ownership of AI-Generated Content: AI-generated content blurs the lines of authorship. Since copyright typically protects works created by human authors, there are questions about whether AI-generated works can be copyrighted, and if so, who would own the copyright—the user, the AI developer, or someone else.
3. Potential for Infringement: AI tools can generate content that closely mimics or copies existing works, leading to potential copyright infringement. If an AI model recreates parts of a copyrighted work without significant transformation, it may violate copyright law.
4. Fair Use Doctrine: There are debates about whether the use of copyrighted works to train AI falls under "fair use." While some argue that training an AI model on large datasets for research or transformative purposes could qualify as fair use, others contend that large-scale commercial AI projects do not meet this standard.
5. Right to Protect and Monetize Creations: Artists, writers, and creators are concerned that AI-generated content could undermine their ability to monetize their own work, as AI can produce similar works more quickly and cheaply. This may impact creative industries economically.
6. Data Mining and Licensing: AI companies often rely on data scraping and mining to gather vast amounts of content for training purposes, which may include copyrighted material. Some argue that content creators should be compensated or have the option to license their works for AI training purposes.
Chaudhary, A., & Zhao, J. (2023, April 19). Generative AI has an intellectual property problem. Harvard Business Review. https://hbr.org/2023/04/generative-ai-has-an-intellectual-property-problem
Crawford, L., & Schultz, J. (2023, October 11). Generative AI: Challenges for copyright law. Issues in Science and Technology. https://issues.org/generative-ai-copyright-law-crawford-schultz/
OpenAI. (2024). ChatGPT (Oct 22 version). [Large language model].
World Economic Forum. (2024, January). Cracking the code: Generative AI and intellectual property. https://www.weforum.org/agenda/2024/01/cracking-the-code-generative-ai-and-intellectual-property/