Artificial Intelligence Can Generate Fraudulent but Authentic-Looking Scientific Medical Articles: Pandora’s Box Has Been Opened

BackgroundArtificial intelligence (AI) has advanced substantially in recent years, transforming many industries and improving the way people live and work. In scientific research, AI can enhance the quality and efficiency of data analysis and publication. However, AI has also...

Full description

Bibliographic Details
Main Authors: Martin Májovský, Martin Černý, Matěj Kasal, Martin Komarc, David Netuka
Format: Article
Language:English
Published: JMIR Publications 2023-05-01
Series:Journal of Medical Internet Research
Online Access:https://www.jmir.org/2023/1/e46924
_version_ 1797734113999847424
author Martin Májovský
Martin Černý
Matěj Kasal
Martin Komarc
David Netuka
author_facet Martin Májovský
Martin Černý
Matěj Kasal
Martin Komarc
David Netuka
author_sort Martin Májovský
collection DOAJ
description BackgroundArtificial intelligence (AI) has advanced substantially in recent years, transforming many industries and improving the way people live and work. In scientific research, AI can enhance the quality and efficiency of data analysis and publication. However, AI has also opened up the possibility of generating high-quality fraudulent papers that are difficult to detect, raising important questions about the integrity of scientific research and the trustworthiness of published papers. ObjectiveThe aim of this study was to investigate the capabilities of current AI language models in generating high-quality fraudulent medical articles. We hypothesized that modern AI models can create highly convincing fraudulent papers that can easily deceive readers and even experienced researchers. MethodsThis proof-of-concept study used ChatGPT (Chat Generative Pre-trained Transformer) powered by the GPT-3 (Generative Pre-trained Transformer 3) language model to generate a fraudulent scientific article related to neurosurgery. GPT-3 is a large language model developed by OpenAI that uses deep learning algorithms to generate human-like text in response to prompts given by users. The model was trained on a massive corpus of text from the internet and is capable of generating high-quality text in a variety of languages and on various topics. The authors posed questions and prompts to the model and refined them iteratively as the model generated the responses. The goal was to create a completely fabricated article including the abstract, introduction, material and methods, discussion, references, charts, etc. Once the article was generated, it was reviewed for accuracy and coherence by experts in the fields of neurosurgery, psychiatry, and statistics and compared to existing similar articles. ResultsThe study found that the AI language model can create a highly convincing fraudulent article that resembled a genuine scientific paper in terms of word usage, sentence structure, and overall composition. The AI-generated article included standard sections such as introduction, material and methods, results, and discussion, as well a data sheet. It consisted of 1992 words and 17 citations, and the whole process of article creation took approximately 1 hour without any special training of the human user. However, there were some concerns and specific mistakes identified in the generated article, specifically in the references. ConclusionsThe study demonstrates the potential of current AI language models to generate completely fabricated scientific articles. Although the papers look sophisticated and seemingly flawless, expert readers may identify semantic inaccuracies and errors upon closer inspection. We highlight the need for increased vigilance and better detection methods to combat the potential misuse of AI in scientific research. At the same time, it is important to recognize the potential benefits of using AI language models in genuine scientific writing and research, such as manuscript preparation and language editing.
first_indexed 2024-03-12T12:38:36Z
format Article
id doaj.art-a55ae9e2c00544088c68192a94e4528e
institution Directory Open Access Journal
issn 1438-8871
language English
last_indexed 2024-03-12T12:38:36Z
publishDate 2023-05-01
publisher JMIR Publications
record_format Article
series Journal of Medical Internet Research
spelling doaj.art-a55ae9e2c00544088c68192a94e4528e2023-08-28T23:59:58ZengJMIR PublicationsJournal of Medical Internet Research1438-88712023-05-0125e4692410.2196/46924Artificial Intelligence Can Generate Fraudulent but Authentic-Looking Scientific Medical Articles: Pandora’s Box Has Been OpenedMartin Májovskýhttps://orcid.org/0000-0001-7725-5181Martin Černýhttps://orcid.org/0000-0002-8601-0554Matěj Kasalhttps://orcid.org/0000-0001-6445-8983Martin Komarchttps://orcid.org/0000-0003-4106-5217David Netukahttps://orcid.org/0000-0001-8609-4789 BackgroundArtificial intelligence (AI) has advanced substantially in recent years, transforming many industries and improving the way people live and work. In scientific research, AI can enhance the quality and efficiency of data analysis and publication. However, AI has also opened up the possibility of generating high-quality fraudulent papers that are difficult to detect, raising important questions about the integrity of scientific research and the trustworthiness of published papers. ObjectiveThe aim of this study was to investigate the capabilities of current AI language models in generating high-quality fraudulent medical articles. We hypothesized that modern AI models can create highly convincing fraudulent papers that can easily deceive readers and even experienced researchers. MethodsThis proof-of-concept study used ChatGPT (Chat Generative Pre-trained Transformer) powered by the GPT-3 (Generative Pre-trained Transformer 3) language model to generate a fraudulent scientific article related to neurosurgery. GPT-3 is a large language model developed by OpenAI that uses deep learning algorithms to generate human-like text in response to prompts given by users. The model was trained on a massive corpus of text from the internet and is capable of generating high-quality text in a variety of languages and on various topics. The authors posed questions and prompts to the model and refined them iteratively as the model generated the responses. The goal was to create a completely fabricated article including the abstract, introduction, material and methods, discussion, references, charts, etc. Once the article was generated, it was reviewed for accuracy and coherence by experts in the fields of neurosurgery, psychiatry, and statistics and compared to existing similar articles. ResultsThe study found that the AI language model can create a highly convincing fraudulent article that resembled a genuine scientific paper in terms of word usage, sentence structure, and overall composition. The AI-generated article included standard sections such as introduction, material and methods, results, and discussion, as well a data sheet. It consisted of 1992 words and 17 citations, and the whole process of article creation took approximately 1 hour without any special training of the human user. However, there were some concerns and specific mistakes identified in the generated article, specifically in the references. ConclusionsThe study demonstrates the potential of current AI language models to generate completely fabricated scientific articles. Although the papers look sophisticated and seemingly flawless, expert readers may identify semantic inaccuracies and errors upon closer inspection. We highlight the need for increased vigilance and better detection methods to combat the potential misuse of AI in scientific research. At the same time, it is important to recognize the potential benefits of using AI language models in genuine scientific writing and research, such as manuscript preparation and language editing.https://www.jmir.org/2023/1/e46924
spellingShingle Martin Májovský
Martin Černý
Matěj Kasal
Martin Komarc
David Netuka
Artificial Intelligence Can Generate Fraudulent but Authentic-Looking Scientific Medical Articles: Pandora’s Box Has Been Opened
Journal of Medical Internet Research
title Artificial Intelligence Can Generate Fraudulent but Authentic-Looking Scientific Medical Articles: Pandora’s Box Has Been Opened
title_full Artificial Intelligence Can Generate Fraudulent but Authentic-Looking Scientific Medical Articles: Pandora’s Box Has Been Opened
title_fullStr Artificial Intelligence Can Generate Fraudulent but Authentic-Looking Scientific Medical Articles: Pandora’s Box Has Been Opened
title_full_unstemmed Artificial Intelligence Can Generate Fraudulent but Authentic-Looking Scientific Medical Articles: Pandora’s Box Has Been Opened
title_short Artificial Intelligence Can Generate Fraudulent but Authentic-Looking Scientific Medical Articles: Pandora’s Box Has Been Opened
title_sort artificial intelligence can generate fraudulent but authentic looking scientific medical articles pandora s box has been opened
url https://www.jmir.org/2023/1/e46924
work_keys_str_mv AT martinmajovsky artificialintelligencecangeneratefraudulentbutauthenticlookingscientificmedicalarticlespandorasboxhasbeenopened
AT martincerny artificialintelligencecangeneratefraudulentbutauthenticlookingscientificmedicalarticlespandorasboxhasbeenopened
AT matejkasal artificialintelligencecangeneratefraudulentbutauthenticlookingscientificmedicalarticlespandorasboxhasbeenopened
AT martinkomarc artificialintelligencecangeneratefraudulentbutauthenticlookingscientificmedicalarticlespandorasboxhasbeenopened
AT davidnetuka artificialintelligencecangeneratefraudulentbutauthenticlookingscientificmedicalarticlespandorasboxhasbeenopened