Skip to Main Content

Generative AI

Bias

Bias

Generative AI, despite its impressive capabilities, is inherently prone to biases that can have far-reaching implications. These biases come from from various sources, including the data used to train the models and the algorithms themselves. AI models learn from vast datasets that often contain historical and societal biases. If these datasets include prejudiced information, the AI systems can replicate and even amplify these biases in their outputs. For instance, if a generative AI is trained on text that primarily reflects the perspectives of a particular demographic, it may produce content that marginalizes or misrepresents other groups.

One of the major challenges in addressing bias in generative AI is the opaque nature of these systems. The algorithms used in deep learning, which underpin many generative AI models, are often considered "black boxes" because it is difficult to understand how they make decisions. This lack of transparency makes it challenging to identify and correct biases. Even when biases are detected, mitigating them is not straightforward. Simply removing biased data can result in incomplete models that lack context or fail to capture the complexity of human language and behavior. Therefore, developing strategies to balance and de-bias training datasets without compromising the performance of AI models remains a critical and ongoing area of research.

The consequences of biased generative AI extend beyond unfair or inaccurate outputs. These biases can reinforce stereotypes, promote misinformation, and contribute to social inequalities. For example, biased AI-generated content in areas like journalism, education, or entertainment can shape public perception and influence societal norms in negative ways. Additionally, in sensitive applications such as hiring processes or legal judgments, biased AI systems can lead to discriminatory outcomes, perpetuating systemic injustices. As AI becomes more integrated into decision-making processes, the impact of its inherent biases becomes more pronounced and problematic.

Resources:

Copyright Infringement & Plagiarism

What is Copyright Infringement?

"Copyright infringement occurs when a copyrighted work is reproduced, distributed, performed, publicly displayed, or made into a derivative work without the permission of the copyright owner." Definition provided by The U.S. Copyright Office

Generative AI & Copyright Law

Copyright infringement concerns when using copyrighted materials to train AI are a central issue in the ongoing debate about the legal implications of generative AI. Training AI models often requires vast amounts of data, which can include text, images, music, and other content that is protected by copyright. The use of such copyrighted materials without proper authorization or licensing raises significant legal and ethical questions.

The core of the infringement issue lies in whether the use of copyrighted materials for training AI models constitutes a violation of copyright law. Copyright law typically grants the copyright holder exclusive rights to reproduce, distribute, perform, display, or create derivative works from their copyrighted material. When AI developers use copyrighted content as part of their training data, they might be reproducing or creating derivative works based on that content, which could potentially infringe on the copyright holder's exclusive rights.

One of the main defenses AI developers might invoke is the doctrine of fair use, particularly in jurisdictions like the United States. Fair use allows for limited use of copyrighted material without permission under certain circumstances, such as for purposes of criticism, comment, news reporting, teaching, scholarship, or research. To determine whether the use of copyrighted materials for training AI constitutes fair use, courts typically consider factors like the purpose and character of the use, the nature of the copyrighted work, the amount and substantiality of the portion used, and the effect of the use on the potential market for or value of the copyrighted work.

However, the application of fair use to AI training is not straightforward. While some argue that using copyrighted materials for training purposes is transformative and falls under fair use, others contend that it can harm the market for the original works, especially if the AI-generated outputs compete with or replicate the original content. The lack of clear legal precedents and guidelines makes it challenging to predict how courts will rule on these cases.

Recent high-profile lawsuits and settlements have highlighted the contentious nature of this issue. For example, artists, authors, and other content creators have sued AI companies for using their copyrighted works without permission to train AI models, claiming infringement. These cases are likely to shape the future of copyright law as it relates to AI.

As AI technology continues to evolve, it is imperative for policymakers, legal experts, and industry stakeholders to establish clear guidelines and regulations. These should balance the need to protect intellectual property rights with the benefits of fostering innovation and creativity in AI development. Developing a consensus on the fair and ethical use of copyrighted materials in AI training will be crucial for the sustainable growth of the AI industry.

Resources:

 

 

Environmental Impact

Environmental Impact

The environmental effects of generative AI are becoming a significant concern as the technology continues to advance. One of the primary environmental impacts stems from the substantial computational power required to train and run AI models. Training state-of-the-art generative AI models involves processing vast amounts of data through complex algorithms, which requires significant energy consumption. This energy is often derived from non-renewable sources, leading to increased carbon emissions and contributing to climate change. For instance, training a single large AI model can generate as much carbon dioxide as five cars over their entire lifetimes, highlighting the substantial environmental footprint of these technologies.

In addition to the energy demands of training AI models, the deployment and use of these models also contribute to their environmental impact. Running AI models in real-time applications, such as chatbots, content generation tools, or recommendation systems, requires ongoing computational resources. As these AI applications become more integrated into everyday life, the cumulative energy consumption continues to rise. This is compounded by the fact that many data centers, where these computations are performed, rely on fossil fuels for their energy needs, further exacerbating the environmental burden.

FERPA

What is FERPA?

"FERPA" stands for the Family Educational Rights & Privacy Act.

Below is an explanation of FERPA as described by the UW-Parkside FERPA website

"The University of Wisconsin-Parkside governs the distribution of student information based on the Family Educational Rights and Privacy Act of 1974 or FERPA. This Act, as amended, established the requirements governing the privacy of student educational records in regards to the release of those records and access to those records. This Act is also known as the Buckley Amendment. The Act gives four basic rights to students:

      • the right to review their education records;
      • the right to seek to amend their education records;
      • the right to limit disclosure of personally identifiable information (directory information);
      • the right to notify the Department of Education concerning an academic institution's failure to comply with FERPA regulations.

FERPA provides for confidentiality of student records; however, it also provides for basic identification of people at the University of Wisconsin-Parkside without the consent of the individual. Release of information to third parties includes directory information, such as contained in the online web-based people directory and in sports brochures. Students are notified of their FERPA rights and the procedures for limiting disclosure of directory information at Orientation, and in an annual email."

How does generative AI effect FERPA?

FERPA is a federal law that protects the privacy of student education records, granting parents and eligible students the right to access and amend their records while also limiting the disclosure of personally identifiable information without consent. With the increasing use of generative AI in educational settings, concerns arise about how these systems handle student data and the potential for inadvertent breaches of FERPA regulations.

One major concern is the data input used by generative AI models. These models often require vast amounts of data to function effectively, and if student information is included in these datasets without proper anonymization, there is a risk of violating FERPA's privacy requirements. Additionally, the output generated by AI systems could inadvertently reveal sensitive student information if not properly managed. For instance, an AI-generated report or analysis might include identifiable student data, leading to unauthorized disclosures. Therefore, educational institutions must ensure that generative AI tools comply with FERPA guidelines by implementing robust data anonymization techniques, securing data storage and transmission, and providing transparency about how AI systems are used and what data they process. 

Human Rights

Human Rights

Generative AI relies heavily on vast datasets for training, many of which are curated and labeled by human workers. Often, this labor is outsourced to workers in developing countries who perform tasks such as data labeling, content moderation, and quality assessment. These workers are frequently underpaid and work under poor conditions, highlighting significant ethical issues. The demand for cheap labor to process enormous amounts of data perpetuates exploitative practices, where workers are paid minimal wages and lack job security, benefits, or adequate working conditions. This exploitation is compounded by the fact that these workers often operate under strict time constraints and high-pressure environments, which can lead to stress and burnout.

The lack of transparency and recognition for these workers raises additional ethical concerns. Many companies utilizing generative AI do not disclose the extent to which human labor is involved in training and refining their models. This invisibility of labor means that the contributions of these workers are often unacknowledged, depriving them of the recognition and respect they deserve. The data they help to curate is crucial for the performance of generative AI systems, yet their role remains largely hidden from the end-users and the broader public. This lack of visibility can also lead to inadequate monitoring of labor conditions, making it difficult to address and rectify exploitative practices.

Another aspect of inappropriate use of human labor in generative AI is the ethical dilemma surrounding the content that these workers are required to handle. Data labeling and content moderation often involve exposure to harmful or disturbing content, including violence, abuse, and explicit material. Workers tasked with moderating this content can suffer from psychological trauma and emotional distress, yet they are often provided with minimal mental health support. This exposure not only impacts their well-being but also raises questions about the ethical implications of subjecting workers to such conditions without adequate protection or compensation.

The imbalance in power dynamics between tech companies and the labor force highlights systemic issues in the industry. Workers engaged in data labeling and content moderation are usually in precarious employment situations, with limited opportunities for advancement or skill development. This exploitation of labor reflects broader inequalities and underscores the need for more ethical practices in the AI industry. Addressing these issues requires a commitment to fair wages, improved working conditions, greater transparency, and mental health support, ensuring that the human labor underpinning generative AI is respected and fairly compensated.

Loss of Individual Voice

Loss of Individual Voice

Generative AI, with its advanced capabilities to create content, is significantly influencing students' individual voices in both positive and negative ways. On the positive side, generative AI tools can serve as powerful aids in enhancing students' writing skills. By providing examples of well-structured essays, creative stories, or even poetry, these tools offer students a reference point to improve their own work. They can learn about different writing styles, expand their vocabulary, and gain a deeper understanding of how to construct compelling narratives. This exposure can help students develop a more sophisticated and polished writing voice, ultimately enhancing their ability to express themselves effectively.

However, there are concerns that reliance on generative AI might stifle the development of a truly unique individual voice. When students use AI-generated content as a crutch, they might become overly dependent on these tools, leading to a homogenization of their writing. The distinctive nuances, personal insights, and creative quirks that define an individual’s voice could be overshadowed by the more formulaic and generic outputs of AI. This could result in a loss of originality and a decrease in the diversity of perspectives within student writing.

While generative AI has the potential to enhance students' writing abilities and expose them to diverse styles, it also poses significant risks to the development of individual voices. Balancing the use of these tools with a strong emphasis on originality and personal effort is crucial. 

University of Wisconsin-Parkside Library   |   Contact Us
900 Wood Road Kenosha, WI 53141 | (262) 595-3432