Skip Navigation
Skip to contents

Science Editing : Science Editing

OPEN ACCESS
SEARCH
Search

Articles

Page Path
HOME > Sci Ed > Volume 11(2); 2024 > Article
Review
Research ethics and issues regarding the use of ChatGPT-like artificial intelligence platforms by authors and reviewers: a narrative review
Sang-Jun Kimorcid
Science Editing 2024;11(2):96-106.
DOI: https://doi.org/10.6087/kcse.343
Published online: August 20, 2024

Korea Research Institute of Bioscience and Biotechnology, Daejeon, Korea

Correspondence to Sang-Jun Kim sjkim@kribb.re.kr
• Received: July 15, 2024   • Accepted: July 24, 2024

Copyright © 2024 Korean Council of Science Editors

This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

prev next
  • 2,911 Views
  • 256 Download
  • 1 Web of Science
  • 1 Crossref
  • 1 Scopus
  • While generative artificial intelligence (AI) technology has become increasingly competitive since OpenAI introduced ChatGPT, its widespread use poses significant ethical challenges in research. Excessive reliance on tools like ChatGPT may intensify ethical concerns in scholarly articles. Therefore, this article aims to provide a comprehensive narrative review of the ethical issues associated with using AI in academic writing and to inform researchers of current trends. Our methodology involved a detailed examination of literature on ChatGPT and related research trends. We conducted searches in major databases to identify additional relevant articles and cited literature, from which we collected and analyzed papers. We identified major issues from the literature, categorized into problems faced by authors using nonacademic AI platforms in writing and challenges related to the detection and acceptance of AI-generated content by reviewers and editors. We explored eight specific ethical problems highlighted by authors and reviewers and conducted a thorough review of five key topics in research ethics. Given that nonacademic AI platforms like ChatGPT often do not disclose their training data sources, there is a substantial risk of unattributed content and plagiarism. Therefore, researchers must verify the accuracy and authenticity of AI-generated content before incorporating it into their article, ensuring adherence to principles of research integrity and ethics, including avoidance of fabrication, falsification, and plagiarism.
Background

The emergence of ChatGPT and fierce technological competition

ChatGPT, a generative artificial intelligence (AI) platform developed by OpenAI, was launched in December 2022 based on generative pretrained transformer 3.5 (GPT-3.5). It is reported that the early release of GPT-4, a multimodal GPT capable of responding to text questions with images, has reduced instances of hallucinations [1]. While ChatGPT is engineered to minimize false responses, it can still produce inaccurate, absurd, or unethical answers. OpenAI’s recent development, Sora, a text-to-video model, generates videos that preserve the quality of the physical world and adhere to user prompts. These AIs, capable of generating sophisticated text, images, and videos, have attracted a large user base, highlighting the rapid advancements in AI technology. Consequently, a race among information companies to lead in generative AI has intensified. Examples include Google’s Bard and its multimodal AI Gemini, Meta’s LLaMA 2, Baidu’s Ernie Bot, and Naver’s HyperCLOVA X. Additionally, Coscientist, an AI dedicated to designing, planning, and conducting chemical experiments, has been developed [2]. The rapid expansion of AI technology has led the Science and Nature journals to recognize AI as one of the top achievements and contributors of 2023. AI technology is increasingly infiltrating everyday life, exemplified by Samsung’s S24 mobile phone, which offers on-device AI interpretation and translation [3].

Changes in the research and publication landscape in response to ChatGPT

As the academic use of general-purpose generative AI has attracted attention, a review article has been published that examined the publication and research trends of ChatGPT-related articles from the perspective of their use, rather than their technical aspects [1]. That review identified a surge of publications, primarily consisting of rapid opinions and suggestions, during the initial phase of ChatGPT’s introduction. One year later, the highest number of ChatGPT-related articles was recorded in Scopus, with PubMed in the medical field and Web of Science across all subjects showing similar figures. Despite a decrease in the proportion of articles in the medical field among ChatGPT-related articles with identifiable subject areas in major databases—from 71.4% of 935 articles in 7 months [1] to 55.7% of 3,222 articles in 12 months [3], the medical community’s interest in this topic has persisted. The community is actively exploring the use of ChatGPT, although its practical application in medicine is still rare. Meanwhile, the education sector has raised numerous concerns about the educational implications, including the evaluation of assignments. The legal community has faced lawsuits concerning AI’s intellectual ownership and unauthorized learning. Additionally, the journal publishing community has begun to establish early guidelines for AI usage in article writing [1]. Therefore, the necessity to clearly define the appropriate use of ChatGPT in scholarly writing has become evident. Generative AI’s exceptional capability to assist with time-intensive tasks, such as reviewing prior research and analyzing data, has led to its increased adoption driven by the pressures of the publish-or-perish environment. The widespread use of ChatGPT-like AI platforms in scholarly articles has posed challenges to academic integrity, raising concerns about the reliability of peer review and the integrity of journal editing. Consequently, the excessive reliance on generative AI could further complicate issues related to research ethics in scholarly publications.

Appearance of academic AI trained on articles

Due to the impact of ChatGPT, as a nonacademic platform, Elsevier launched Scopus AI, which utilizes abstracts from articles published after 2013 in Scopus. This service, available since 2024, offers natural language summaries, visualized results, and suggested follow-up questions. Similarly, Elicit, which also relies on article abstracts, is accessible for a fee, while a private AI trial using Dimensions has been announced. Clarivate has initiated the development of an AI that is trained on Web of Science articles [3]. Unlike ChatGPT, which primarily sources its training data from publicly available internet content, these new types of academic generative AI, trained on bibliographic databases containing articles, aims to assist researchers in drafting introductions and literature reviews. Given that 62.2% of researchers consider the use of ChatGPT in article writing unethical [4], the ethical implications of AI-driven articles are a significant concern. Although academic AI requires either library subscriptions or personal expenses and has been slower to spread—having been introduced a year after ChatGPT—it represents a more controlled and potentially ethical approach. In contrast, nonacademic ChatGPT is both affordable and widely accessible.

Need for a review on the use of AI in article writing

Unlike academic AI, the use of general-purpose and nonacademic generative AI platforms such as ChatGPT raises several research ethics concerns. These include the quality of training data, bias in learning outcomes, the accuracy of generated content, and potential issues with unquoted material and plagiarism [5]. Therefore, researchers must strictly adhere to research ethics when using nonacademic AI tools to enhance their productivity in article writing. The ease of using AI tools like ChatGPT in articles, without proper citations and references, could exacerbate research ethics issues due to fierce competition. While information technology (IT) tools such as Grammarly (Grammarly Inc) can improve the quality of writing without raising research ethics concerns, AI that instantly produces sophisticated text, images, and videos in response to user prompts presents various ethical issues in the learning and creation processes. According to a survey published in Nature, more than half of the 1,600 researchers surveyed expected AI tools to be very important or essential in their fields [6]. Nonetheless, about 68% believed these tools would make plagiarism easier, 55% were concerned that they could facilitate fraud, and 53% indicated that careless use might lead to irreproducible research. Consequently, there is an urgent need to provide researchers with clear guidelines on using AI for writing articles. These guidelines should cover the use of nonacademic generative AI, the detection and allowance of AI use during peer review, and the adherence to research ethics by both authors and reviewers.
Objectives
This article aims to comprehensively review the research ethics issues raised by using ChatGPT-like AIs in article writing and to share trends with researchers interested in utilizing AIs. For this purpose, we utilized the author’s previous research on the academic usage trends of ChatGPT [1,3] as a foundation to identify additional papers on AI usage reported in major databases up to March 2024. We then conducted searches in PubMed for similar articles with citations in the literature and expanded our review through supplementary searches. The content of the final collection of articles was primarily examined for this purpose. The research ethics issues in articles using AI were categorized from the perspectives of authors, reviewers, and editors. These issues were thoroughly investigated and are discussed in this article from the standpoint of research ethics compliance. In this review, AI refers to nonacademically generated AI, such as ChatGPT, unless specified otherwise; it does not refer to academically generated AI, such as Scopus AI.
Ethics statement
This review was not a study with human subjects, so neither institutional review board approval nor informed consent was required.
What aspects of writing articles can be AI-assisted?
In writing articles, AI can be utilized for accurate translation, grammatical correctness, and idea generation, as well as for summarizing content, and crafting conclusions [1]. Journal reviewers have found AI particularly useful for generating research ideas and summarizing data [7]. AI has also proven effective in generating ideas for systematic reviews in plastic surgery, addressing complex problems with accuracy and simplicity [8]. Surgical students in the United States have reported that they perceived articles generated by ChatGPT as more straightforward and better organized than those from evidence-based sources [9]. Although ChatGPT can produce more readable abstracts and respond concisely to critical information, the quality of these abstracts was generally rated as lower than that of the original texts [10]. A study comparing ChatGPT-generated abstracts with those from an obstetrics journal indicated that the median score—based on Grammarly scores, writing issues, and correctness errors—was lower than that of the original abstracts, although there were no grammatical inaccuracies [11]. In expert assessments of clinical reviews, AI-generated summaries were found to be more extractive and comprehensive, though human-generated summaries often included valuable interpretations that were missing from AI summaries [12]. ChatGPT has been noted for producing more refined sentences more quickly than traditional English proofreading services, making it a valuable tool for language editing [13]. AI offers numerous advantages throughout the various stages of article writing [14]. Upon reviewing the literature, the anticipated benefits of AI in article writing can be summarized as follows: researchers can use AI to discover, translate, and summarize articles and research trends, identify experimental methods and scientific knowledge, and compile results and statistics. For authors, AI can assist in drafting abstracts, keywords, introductions, literature reviews, data analyses, conclusions, and references, and it can help identify similarities for anti-AI checks. Reviewers can benefit from AI for similarity and anti-AI checks, locating peer reviewers, and drafting review results. Readers may find AI helpful for translating and summarizing content, discovering complementary knowledge, and organizing related information.
How many poor references and hallucinations are due to AI learning problems?
The evaluation scores assigned by reviewers to generated articles were ranked in the following order: abstract, literature review, research idea, methodology, citations, and references [15]. In that study, only 8% of the references were retrievable via Google Scholar and Mendeley. In other studies of article writing, the accuracy of references was low, with only 7% being correct [16] and 70% of the cited references being inaccurate [17], highlighting the need for careful reference checking when using AI in article writing. ChatGPT correctly answered questions from radiologists 67% of the time, and only 36.2% of the references it provided contained accurate information [18]. In medical articles, 56% of references did not have a digital object identifier (DOI) or were not searchable on Google [19]. Many articles written by AI contained significant inaccuracies and fictitious references, potentially appearing credible to nonexpert readers [20]. The issue with such unreliable referencing is its plausibility, which can mislead uninformed authors or readers into accepting it as factual. Although AI-related errors and hallucinations are problematic, they are not quantifiable, making it challenging to locate prior research on these issues.
Can AI write both review articles and case reports?
The results of a trial on AI-only writing of scientific review articles were published in the journal Current Osteoporosis Reports. Human authors completed their tasks more quickly than AI, which required additional time for fact-checking and editing [21]. When AIs assist in review writing, they produce drafts more rapidly but necessitate extra time for content verification due to concerns about plagiarism and high similarity scores [17], or they require more editing because of hallucinations and plagiarism [22]. Similarly, while AI can independently write scientific review articles, the author must subsequently verify the scope and length of edits, the accuracy of claims, and the potential for plagiarism [23]. In a hybrid narrative review case involving collaboration between humans and ChatGPT, the results highlighted both the effectiveness and the concerns associated with ChatGPT [24]. Meanwhile, in evaluations of dermatology case reports by experts, AI-derived text received higher scores for quality and readability, yet human reviewers only correctly identified an AI-generated case report 35% of the time [25]. AI is also capable of simplifying text to a layperson’s level, as demonstrated in the drafting of a fracture case report [26]. Although the case report had some errors, it was also accurate and complete [27]. However, a medical case report entirely generated by ChatGPT was found to have inadequate content, lacked references, and was deemed unsuitable for producing scientific articles [28]. Exploring ChatGPT’s potential for creating clinical vignettes could lead to misinformation [29]. Ultimately, entrusting the complete creation of scientific reviews or case reports to AI is not advisable due to concerns over content accuracy and quality reliability. This necessitates additional time for authors to post-check and suggests that AI should be used primarily to design drafts or specific sections of these articles.
How often is AI used in articles?
In a ZeroGPT (ZeroGPT) test of each 200 Orthopaedics & Traumatology: Surgery & Research (OTSR) journal articles before and after the advent of ChatGPT, AI-generated text was detected in 10.1% of abstracts and in 5.6% of the main texts [30]. Although the plagiarism rate remained unchanged, the incidence of AI-generated content was approximately 5% higher than previously observed, suggesting an association between the use of translation software and increased AI detection rates. A comparison of OTSR journal articles from 2022 and 2023 using ZeroGPT revealed that the average percentage of articles with suspected AI use was 11%± 6%, indicating a rise in the utilization of AI [31]. An examination of acknowledgments that mentioned AI use in articles showed that medicine had the highest number of publications utilizing ChatGPT. Specifically, 24% of these articles acknowledged AI use in content generation, 37% recognized the contributions of OpenAI and ChatGPT, 41% mentioned their use for support and collaboration, and 53% cited AI use in their references [32].
Should AI authorship and attribution of AI be allowed?
After the release of ChatGPT, the journal publishing community quickly addressed the issue of AI authorship and the acceptability of AI-assisted articles [33]. The Committee on Publication Ethics (COPE) and the top five publishers [34]—Elsevier, Springer, MDPI, Wiley, and Taylor & Francis—took a similar stance, stating that if an AI platform is used, it is the responsibility of the author to acknowledge the use, but AI cannot be the author [35]. However, the journal Science has adopted a stricter policy, stating that using AI without permission is considered scientific misconduct [36]. The guidelines for AI use in most journals, except Science, are summarized as follows [1]: AI cannot be credited with authorship; human authors must assume full responsibility for the accuracy of the results; and there must be clear disclosure of the AI tools utilized. In 2021, the sharing rate of scientific articles from these top five publishers exceeded 62% [34]. Following the introduction of ChatGPT, many journals have adopted these AI authorship policies, which are now upheld by reviewers. Therefore, most articles, particularly those from the top five publishers, adhere to these policies. However, some journals still lack specific policies regarding AI authorship.
Can reviewers and editors determine whether an article is AI-assisted?
A previous study noted that human-written articles were specific, varied, and informative, whereas ChatGPT tended to use general terms, prioritizing fluency and logical coherence over contextual accuracy [37]. When orthopedic surgeons compared AI-generated abstracts to their original counterparts, they found that the overall quality of human-written abstracts was superior, although the differences were not statistically significant [38]. In tests where human reviewers compared original and ChatGPT-generated abstracts, they correctly identified 68% of the ChatGPT abstracts but mistakenly labeled 14% of the original abstracts as AI-generated [39]. When only titles and abstracts were used for comparison, reviewers accurately identified 30% of ChatGPT abstracts and 43.7% of human abstracts, yet they struggled to distinguish between the two precisely [40]. Orthopedic residents and professors were able to correctly identify the authors of the abstracts 34.9% and 31.7% of the time, respectively, while two AI detectors achieved higher accuracy rates of 42.9% and 66.6% [41]. When reviewers were given AI-generated abstracts along with the full text of the journal, they achieved a 93% accuracy rate in identifying ChatGPT abstracts. The quality of these abstracts was significantly lower in unstructured formats, but no quality differences were observed in structured abstracts with four subheadings [42]. The quality of the generated abstracts did not meet the threshold for acceptance in peer-reviewed journals, and AI-generated text was more likely to be recognized by experienced reviewers [15]. Therefore, articles written by AI are often not detected by less experienced reviewers.
How accurate are tools for detecting AI-assisted articles?
Given the distinct characteristics of text written by humans and AI, algorithms can effectively detect AI-generated content [37]. AI writing in chemistry journals can be accurately detected, as demonstrated in article introductions [43]. Although this detector developed just 1 month after GPT-4 showed 100% accuracy in chemistry journals, its performance in other journals with a more liberal experimental process remains unknown. AI abstracts generated from the titles of medical journals were detected with over 99.9% accuracy by the GPT-2 Output Detector [39]. In a study, ZeroGPT identified AI-written texts with an accuracy of only about 35% to 65%, and OpenAI’s text classifier achieved around 10% to 55% accuracy [44]. In evaluating the AI detection performance of tools like GPTZero (GPTZero Inc), ZeroGPT, Writer AI content detector (Writer Inc), and Originality (Originality.AI Inc), Originality excelled in distinguishing between AI-generated and human texts [45]. In one study, GPTZero showed a low classification rate for human text as machine-generated and a high rate for classifying machine-generated text as human [46]. However, concerns have been raised that paraphrasing may reduce the detection capabilities of GPT detectors [47]. Prevention is deemed more important than detection, as approximately 50% of AI-generated texts using obfuscation technology are misattributed to humans [48]. The accuracy rates were 96% for human-written texts, 76% for machine-translated to English, 74% for machine-generated without modifications, 42% for machine-generated with subsequent human editing, and only 26% for machine-generated with subsequent machine paraphrasing [48]. Therefore, more accurate AI detection tools will be invaluable in peer review, content correction, and syntax improvement across various journals. These AI detection tools are not always correct or reliable, and content obfuscation techniques further degrade their performance. Testing the performance of AI detection tools with machine paraphrasers, translators, and patch writers is essential. It is challenging to perfectly detect the behavior of paraphrasing text in an article, as current technology struggles to identify different expressions of the same meaning used to evade detection. This situation resembles the proverbial competition between spears and shields, reflecting the ongoing battle between generative AIs attempting to evade detection and machine tools striving to detect them more effectively. While article reviewers and editors rely on the support and robustness of these AI detection tools, their practical use is limited by technical challenges in detecting AI text that has been edited and paraphrased perfectly.
Should AI be used to support reviewers and editors in the review process?
When using ChatGPT, journal reviewers were only able to distinguish between AI-generated and human-written text 38.9% of the time, with 25.9% expressing disagreement, even though many editors considered AI acceptable for use [49]. Therefore, numerous journal editors deem AI-generated abstracts suitable for summarizing and editing text. AI detection tools used by reviewers struggle to identify AI-assisted content, underscoring the importance of weighing the risks and benefits of AI in the peer review process. In evaluations of MEDLINE abstracts, originality demonstrated 100% sensitivity, 95% specificity, and 97.6% accuracy in differentiating texts, although the likelihood of AI-generated text rose from 21.7% to 36.7% [50]. Quality assessments by human reviewers and ChatGPT for accepted articles were generally comparable. However, there were more pronounced differences in the scores for rejected articles, indicating less consistency in ChatGPT’s evaluations [51]. Thus, AI can expedite the review process and support peer reviewers by enhancing efficiency. Nevertheless, the overarching conclusion from previous studies is that AI cannot fully replace human reviewers.
The use of ChatGPT in article writing is likely to become more widespread, as it can be used for various purposes. Therefore, ChatGPT poses potential ethical challenges, including threats to academic integrity and data accuracy [52]. To address these issues, a collaborative effort from all stakeholders in academia is essential. This includes improving peer review systems with technology designed to detect AI-generated plagiarism, advocating for higher ethical standards across the academic community, and providing clear guidelines for the responsible use of AI tools [53]. Therefore, the main ethical concerns and research issues that authors and reviewers should consider when incorporating AI into their articles are summarized in Fig. 1, focusing on representative issues.
High similarity scores and copyright violation concerns in AI learning
Generative AI primarily relies on publicly available data for learning, which means it may not access the most professional or up-to-date information. If inaccuracies from AI-generated content are incorporated into articles, issues such as data bias may arise. During its training, AI probabilistically generates text, often producing content that closely resembles existing research or sources. This similarity raises the potential for copyright infringement due to unattributed use of learning data and plagiarism. There have been instances of copyright litigation involving media data used in AI training, as well as the creation of new images from training images obtained without permission [1]. The European Commission has proposed a new legal framework for the intellectual property rights of AI-generated works, including a copyright exception for “text and data mining” [54]. Furthermore, in the United States, AI-generated works may qualify for copyright registration under certain conditions as outlined in the rulemaking guidelines. Following these developments, the European Parliament passed the Artificial Intelligence Act in March 2024. Consequently, if AI-generated content is used in articles without proper verification, the author could face copyright infringement charges, necessitating a thorough follow-up review.
Concerns about low quality due to AI’s poor citations and inadequate references
Ensuring the reliability and accuracy of article content is crucial for scientific research. Generating misleading or inaccurate content compromises scientific integrity. The excessive use of AI-generated content without proper citation can lead to inadvertent plagiarism. Plagiarism, whether it involves stealing words or ideas, can undermine research integrity [55]. While the plagiarism of ideas has traditionally been unacceptable, attitudes toward text plagiarism often vary depending on cultural perspectives [46]. Additionally, double plagiarism poses a significant issue when an author reuses plagiarized text without acknowledging the original source, which may have been learned by AI. However, it is challenging to determine definitively whether AI-generated content is original. Copyright infringements through plagiarism can lead to potential criminal consequences and civil liabilities. In the field of nephrology, ChatGPT has a citation accuracy rate of only 38% [56], which, although better than other AIs, still indicates a significant issue with referencing. Poor referencing by AIs can result in not only low-quality articles but also in the perpetuation of inaccuracies if authors do not verify the information. Nevertheless, we believe that the issues of missing citations and inadequate references in ChatGPT can be addressed by academic AIs like Scopus AI, which has been trained on scholarly articles.
Anxiety related to fabrication, falsification, plagiarism, and research integrity
Research integrity is fundamentally linked to the honesty and reliability of the research process. The use of AI to fabricate, falsify, or plagiarize content in research data or results represents a severe breach of this integrity [49]. Concerns arise when image or video-generating AI tools such as DALL-E 2 (OpenAI), Midjourney (Midjourney Inc), Stable Diffusion (Stability AI), or Sora are used to create images from tables and figures for inclusion in articles without proper justification, leading to potential fabrication and falsification of research data. Additionally, if AI generates fabricated texts, transcripts of interviews, or responses to questions, it could lead to the spread of misinformation, dissemination of bias or discrimination, decreased transparency, privacy violations, and diminished researcher responsibility and accountability. The capability of reviewers and AI detection tools to identify AI-generated content is therefore a critical issue that journal publishers must urgently address [53]. As AI-generated content becomes more prevalent, it will be interesting to observe whether mandatory labeling of AI content effectively addresses issues of fabrication, falsification, and data security. Moreover, if AI randomly generates biased, discriminatory, and untruthful texts that are used in articles, issues of plagiarism and research integrity may emerge since these texts are characterized by falsehoods and distortions. Human authors must assume full responsibility for ensuring the originality, transparency, integrity, and validity of their AI-supported manuscripts. In the context of AI-assisted articles, we can anticipate the development of AI-supported reviewers that assist editors in ensuring the fairness and efficiency of the peer review process.
The need for uniform standards for AI attribution and inappropriate authorship
To date, most countries have maintained that for an AI creation to be eligible for copyright, it must reflect the thoughts and feelings of a human and exhibit a minimum level of creativity. If AI is incorrectly identified as the author, it is not recognized as such under copyright law due to its nonhuman status. This misattribution could also constitute a violation of copyright law. Although it is technically feasible to submit an article listing AI as an author, it is unlikely to be accepted. Many journals have policies that prevent AI from being credited as an author, insisting that the actual author must assume full responsibility for the content’s integrity and adherence to research ethics [1]. An investigation of the AI authorship policies of 300 major journals found that 58.7% had an AI authorship policy, with 96.6% allowing AI to improve manuscript quality and 98.9% stating that AI cannot be an author [57]. However, a database search for articles that named ChatGPT as an author yielded 16 unique entries after one year [3]. These articles predominantly appear in journals, repositories, or archives that lack established AI authorship policies. Looking ahead, if the EU’s proposal to grant copyrights to AI products is adopted, the way authors and papers utilizing AI products are cited and labeled might change. Nonetheless, there is a pressing need for the development of author policies and guidelines concerning AI use. These should be crafted in collaboration with the academic community to establish clear, common standards for both authors and reviewers. Recommendations from research ethics committees could further guide these discussions on AI utilization [58].
Disrupting subsequent research with numerous retractions due to the overuse of AI
The use of AI to rapidly generate low-quality manuscripts and the proliferation of predatory journals that rely heavily on AI for rapid publication—as exemplified by the publication speed of MDPI [34]—pose significant concerns for publishers, editors, reviewers, readers, and other stakeholders in the publishing industry. This trend leads to wasted energy and time as these parties strive to verify the authenticity of the content. It is crucial to differentiate between human-generated and AI-assisted content, given that the imperfections in AI detection tools make accurate classification both time-consuming and expensive. If technological measures are not implemented early in the review process to assess AI-assisted text and images, there is a risk of producing spurious or low-quality articles, potentially leading to an increased rate of retractions.
Science list in 2023 belonged to Hindawi, a publisher acquired by Wiley. Along with removing these journals, the number of articles produced by “paper mills” was estimated to be in the hundreds of thousands [59]. In 2023, Wiley-Hindawi retracted over 8,000 articles from a total of approximately 10,000 due to issues such as incoherent text and irrelevant references. Many of these articles were from special issues overseen by guest editors. This led to Wiley’s decision to discontinue the Hindawi brand name and a loss of US $35 million in annual revenue [59]. When an article is retracted and removed from bibliographic databases due to suspicions of inadequate peer review or excessive AI assistance, the credibility of subsequent studies or reviews that cite or discuss it is compromised, thereby disrupting the knowledge ecosystem and societal trust. ChatGPT can significantly enhance research by aiding in idea generation, content creation, and experience direction, provided it is used within the guidelines and limitations set by journals. If properly implemented, ChatGPT can improve the quality of articles and research ethics. However, AI cannot completely replace the critical roles of human authors and reviewers [60].
With the emergence of ChatGPT, competition in the field of generative AI technology is growing, and the application of general-purpose, nonacademic generative AIs like ChatGPT in scholarly articles is becoming more common. Utilizing ChatGPT in scientific research poses potential risks to the integrity of research, sparking concerns about the accuracy and originality of published articles. In this study, we initially reviewed existing research to pinpoint issues that both authors and reviewers need to be aware of when incorporating AI in the processes of article writing and peer review. We then explored the ethical challenges faced by authors and reviewers using ChatGPT-like AIs, including issues such as high similarity in AI outputs, copyright law violations, compromised quality due to inadequate citations and unquoted sources, concerns over research integrity (including fabrication, falsification, and plagiarism), the necessity for standardized AI authorship guidelines, inappropriate attribution of authorship, and the potential disruption of future research due to numerous retractions [6163] stemming from AI overuse.
While academic generative AIs are significantly less prone to hallucinations and illusions associated with poor referencing, nonacademic generative AIs such as ChatGPT are inherently more susceptible to unquote and plagiarism due to the nondisclosure of trained sources. This review reveals that in many articles, AI is not permitted to be listed as an author in principle, and human authors must assume responsibility for all fact-checking and research integrity. Additionally, the identification of AI-generated content remains imperfect for reviewers and AI detection tools, necessitating their use in the submission and review process. It is also essential to differentiate between the legal issues in the learning process and research ethics when utilizing content. Moreover, a unified AI-related policy could provide guidance and transparency for authors, readers, reviewers, editors, and publishers. Addressing the research ethics issues associated with AI use will require education on writing articles with research integrity, which prevents fabrication, falsification, and plagiarism. Consequently, ChatGPT should be considered an additional AI tool to expedite the production of high-quality articles. However, this review is based on the assumption that all results from previous studies are valid, and the accuracy and truthfulness of this review are limited to the reliability of these cited studies, reflecting the author’s lack of expertise in all of the discussed domains.

Conflict of Interest

No potential conflict of interest relevant to this article was reported.

Funding

The author received no financial support for this article.

Data Availability

Data sharing is not applicable to this article as no new data were created or analyzed in this study.

The author did not provide any supplementary materials for this article.
Fig. 1.
The research framework of this review. AI, artificial intelligence.
kcse-343f1.jpg

Figure & Data

References

    Citations

    Citations to this article as recorded by  
    • How is ChatGPT acknowledged in academic publications?
      Kayvan Kousha
      Scientometrics.2024;[Epub]     CrossRef

    Figure
    • 0
    Research ethics and issues regarding the use of ChatGPT-like artificial intelligence platforms by authors and reviewers: a narrative review
    Image
    Fig. 1. The research framework of this review. AI, artificial intelligence.
    Research ethics and issues regarding the use of ChatGPT-like artificial intelligence platforms by authors and reviewers: a narrative review

    Science Editing : Science Editing
    TOP