In today's digital era, the advancement of artificial intelligence (AI) is revolutionizing numerous trades, including academic research and scholarly writing. Generative artificial intelligence (AI), for example, uses specialized AI systems to create new content, including images, text, and code.
Like countless other people, researchers and scholars around the world are exploring the value and utility of generative AI. When applied to academic writing, this set of AI tools has the potential to support researchers in all aspects of their writing, from brainstorming topics and formulating hypotheses to drafting abstracts and editing manuscripts.
There is debate, however, surrounding the ethics of using generative AI in research and academic writing. Questions like, “How will it affect the quality and originality of my work?” and “How can I avoid plagiarism and make the best use of this tool?” are top of everyone’s mind.
Here, we will explore these questions and reveal some agreed upon best practices for the responsible use of generative AI in research.
The short answer is ‘yes,’ always.
To ensure transparency and reproducibility, researchers must treat anything produced by generative AI just as they would any other source of information to avoid scrutiny and malfeasance.
How exactly to accomplish this is the tricky part.
Initially, there were instances of researchers listing a specific AI tool as co-author on their manuscripts. The new practice sparked immediate backlash and rebuttal in both the science and publishing industries for good reason. AI, machine learning, and algorithmic tools in general do not meet the criteria of authorship. Most importantly, it cannot take responsibility nor be held accountable for the content.
There was also discussion about implementing a disclosure system. The system involved adding a disclaimer to written works that would help readers quickly identify who or what contributed information. Robert J. Gates proposed a simple model:
The current standard of practice for acknowledging the use of generative AI in research projects is outlined in the authorship and publishing guidelines of all the major scholarly journals, such as:
How to cite the use of AI is still evolving. Understanding when it is necessary can be confusing, but it is always best to err on the side of caution. Follow your institution’s and intended journal’s guidelines explicitly. Never assume ‘they won’t find out.’
AI detectors exist. They’re growing in number and becoming more accurate everyday. They are used extensively by professors, peers, editors, and anyone else who wants to know how a piece was written. Many of these tools detect AI generated text and images and determine what percentage of the work was done by AI by highlighting the parts.
Again, the answer is a definitive ‘yes’. AI is a technology, a tool; it has no sense of morality or reason. It simply formulates answers to prompts based on a finite dataset without any true understanding of its informational inputs or outputs.
Because generative AI has no means to interpret the data it works with, the answers are often incorrect. AI can only combine and recombine answers in ways that potentially satisfy a prompt. The liability of accuracy falls squarely on the users’ shoulders.
For this reason, it is of the utmost importance that researchers are diligent in reviewing, verifying, and correcting all AI generated content. Some of the most common mistakes are found in:
When using generative AI, the extent of one’s expectations needs to directly correspond with the degree of their efforts. By carefully structuring and restructuring the prompts they input, researchers ensure the best possible outputs.
Try this formula and examples:
Acting as [ROLE] perform [TASK] in [FORMAT]: insert unique data
Example 1: Acting as an expert science writer write an abstract in PDF: (paste your manuscript here)
Example 2: Acting as a PhD student in Chemistry draft references in APA citation and alphabetize. (paste a list of your sources here)
Use this formula as a jumping board for identifying the key facets of your prompt. Then, continue reworking it and adding details until satisfied with the results.
When an author chooses to learn and understand the limited capabilities of generative AI, it can be a valuable asset. Think of this tool as an assistant for supporting various aspects of the writing process:
It is important to note the variety of AI choices in these examples. Each tool is trained on a unique set of data that determines its effectiveness in performing specialized tasks. Pinpointing and selecting the one that is most appropriate for your needs is essential for getting the most out of these tools.
Like any new tool or technology, generative artificial intelligence is surrounded by a buzz of excitement and a host of concerns. It has the potential to both revolutionize academic research and scholarly writing and sabotage its ethical foundations.
Cultivating, following, and sharing responsible best practices for these tools is imperative to upholding the integrity of the research process. By embracing transparency, providing human oversight, and acknowledging the limitations of generative AI, researchers will remain at the forefront of ethical innovation.