Skip to Main Content

Artificial Intelligence Literacy

Hallucinations and Misinformation

AI Hallucinations occur when generative AI produces wrong responses to a prompt or query.  This could involve:                                   Image by Andrew Martin from Pixabay

  • Mixing true and false statements
  • Misstating a detail
  • Making entirely false claims and fabricating information (Morris, 2024)

IBM describes an AI hallucination as when a large language model (LLM) - often a chatbot - perceives patterns or objects that are nonexistent or imperceptible to human observers, creating outputs that are nonsensical or altogether inaccurate.                                                                                                                      Image by Andrew Martin from Pixabay

  • The way AI is trained can have a lot to do with it hallucinating
    • AI that is trained on incredibly large language models without "correct" answers can end up hallucinating.
    • Generative AI operates using a sophisticated algorithm that examines how humans arrange words online. It doesn't determine what is true or false. 
    • By identifying patterns in the data, a large language model (LLM) learns to predict the next word in a sequence, functioning similarly to an advanced autocomplete tool.
    • Since the internet contains false information, the model can also pick up and replicate those inaccuracies. Additionally, chatbots sometimes fabricate content, creating new text by blending billions of patterns in unpredictable ways. As a result, even if the model were trained solely on accurate data, it might still generate incorrect information (Weise et al., 2023).

Generative AI is great at pattern detection and mimicry.  It is trained to make predictions and then generate a reponse to a prompt, based off its training data. Since the internet is filled with people making unfounded and outrageous claims, confidently stating falsehoods, and making things up, it is no surprise that AI trained on the internet would do the same things  (Morris, 2024). 

 

IBM. (n.d.). AI hallucinations. https://www.ibm.com/topics/ai-hallucinations

Morris, S. (2024). Exploring AI with critical information literacy. ALA Learning.  

Weise, K., Metz, C., Lohr, S., & Grant, N. (2023, May 8). When Chatbots “Hallucinate.” New York Times172(59782), B4.

Copyright

The main concerns about AI and copyright revolve around issues of ownership, infringement, and the use of copyrighted material to train AI models. Here are some key concerns:

 

1. Use of Copyrighted Material for AI Training: Many AI models, especially  generative ones, are trained on vast amounts of data that may include copyrighted works. The unauthorized use of such material in training can raise legal and ethical questions. Creators may argue that AI companies are benefiting from their work without permission or compensation.

Image by Mohamed Hassan from Pixabay

2.  Ownership of AI-Generated Content: AI-generated content blurs the lines of authorship. Since copyright typically protects works created by human authors, there are questions about whether AI-generated works can be copyrighted, and if so, who would own the copyright—the user, the AI developer, or someone else.

3.  Potential for Infringement: AI tools can generate content that closely mimics or copies existing works, leading to potential copyright infringement. If an AI model recreates parts of a copyrighted work without significant transformation, it may violate copyright law.

4.  Fair Use Doctrine: There are debates about whether the use of copyrighted works to train AI falls under "fair use." While some argue that training an AI model on large datasets for research or transformative purposes could qualify as fair use, others contend that large-scale commercial AI projects do not meet this standard.

5.  Right to Protect and Monetize Creations: Artists, writers, and creators are concerned that AI-generated content could undermine their ability to monetize their own work, as AI can produce similar works more quickly and cheaply. This may impact creative industries economically.

6.  Data Mining and Licensing: AI companies often rely on data scraping and mining to gather vast amounts of content for training purposes, which may include copyrighted material. Some argue that content creators should be compensated or have the option to license their works for AI training purposes.

Chaudhary, A., & Zhao, J. (2023, April 19). Generative AI has an intellectual property problem. Harvard Business Review. https://hbr.org/2023/04/generative-ai-has-an-intellectual-property-problem

Crawford, L., & Schultz, J. (2023, October 11). Generative AI: Challenges for copyright law. Issues in Science and Technology. https://issues.org/generative-ai-copyright-law-crawford-schultz/

OpenAI. (2024). ChatGPT (Oct 22 version). [Large language model].  

World Economic Forum. (2024, January). Cracking the code: Generative AI and intellectual property. https://www.weforum.org/agenda/2024/01/cracking-the-code-generative-ai-and-intellectual-property/