Skip to Main Content

Artificial Intelligence (AI) Research Tools

Generative AI information and tools
Last Updated: Aug 29, 2024 12:20 PM

Bias in AI

A potentially significant limitation of AI is the bias that can be embedded in the products it generates. Large language models (LLMs) are fed immense amounts of data and text available on the internet and trained to simply predict the most likely sequence of words in response to a given prompt. Therefore, the LLM will reflect and perpetuate the biases inherent in the inputted internet information. An additional source of bias lies in the fact that some generative AI (GAI) tools utilize reinforcement learning with human feedback (RLHF), with the caveat that the human testers used to provide this feedback are themselves non-neutral. Accordingly, GAI tools, like ChatGPT, are documented to have provided output that is socio-politically biased, occasionally even containing sexist, racist, or otherwise offensive information.

Selected Readings 

Plagiarism & Academic Integrity

Generative AI (GAI) tools have introduced new challenges in academic integrity, particularly related to plagiarism.

Plagiarism is typically defined as presenting someone else's work or ideas as one's own. While a generative AI tool might not qualify as a "someone," using text generated from an AI tool without citing is still considered plagiarism. University at Buffalo instructors have the academic freedom to determine what tools students can and cannot use in pursuit of meeting course learning objectives. See Artificial Intelligence Guidance from the UB's Office of Academic Integrity. Individual policies for using and crediting GAI tools may vary from class to class. GAI tools, such as ChatGPT, have been known to generate false citations, and even if the citations represent actual papers, the cited content in ChatGPT might still be inaccurate.

Related Recommendations

  • If GAI tools are permitted to be used for topic development, in the early stages of research, you might not need to cite them at all, but it's still important to check with your instructor first.
  • If you are providing commentary or analysis on the text generated by a chatbot and are either paraphrasing its results or quoting it directly, a citation is always required. You can find more information on citing GAI tools on this guide's Citing Generative AI page.
  • If you are a researcher planning to publish in a journal, it is best to review that journal's policies on the permitted use of GAI tools. (See 'Selected Readings' below for a couple of examples of journal policies.)
  • It's important to always look up citations and check to make sure they are accurate. If you're citing information from a GAI source, try to cite the original source, rather than the GAI.

Selected Readings

Privacy and AI

There are currently also multiple privacy concerns associated with the use of generative AI (GAI) tools. The most prominent issues revolve around the possibility of a breach of personal/sensitive data and re-identification. More specifically, most AI-powered language models, including ChatGPT, require users to input large amounts of data to be trained and generate new information products effectively. This translates into personal or sensitive user-submitted data becoming an integral part of the collection of material used to further train the AI without the explicit consent of the user. Moreover, certain GAI policies even permit AI developers to profit off of this personal/sensitive information by selling it to third parties. Even in cases when clear identifying personal information is not entered by AI user, the utilization of the system carries a risk of re-identification as the submitted dataset may contain patterns allowing for the generated information to be linked back to the individual or entity.  

Related Recommendations

  • Avoid sharing any personal or sensitive information via the AI-powered tools. 
  • Always review the privacy policy of the generative AI tools before utilizing them. Be cautious about policies that permit for the inputted data to be freely distributed to third-party vendors and/or other users. 

Selected Readings

Extractive Aspects of Generative AI

When discussing the ethics of AI, it is important to consider the impact on the environment and human labor. In their article for Ars Technica, Sasha Luccioni uses the following graphic to visualize the numerous costs of generative AI (GAI).

Selected Readings

Luccioni, S. (2023, April 12). The mounting human and environmental costs of Generative AI. Ars Technica. https://arstechnica.com/gadgets/2023/04/generative-ai-is-cool-but-lets-not-forget-its-human-and-environmental-costs/