GPT-fabricated scientific papers on Google Scholar : Key features, spread, and implications for preempting evidence manipulation
(2024) In Misinformation Review- Abstract
- Academic journals, archives, and repositories are seeing an increasing number of questionable research papers clearly produced using generative AI. They are often created with widely available, general-purpose AI applications, most likely ChatGPT, and mimic scientific writing. Google Scholar easily locates and lists these questionable papers alongside reputable, quality-controlled research. Our analysis of a selection of questionable GPT-fabricated scientific papers found in Google Scholar shows that many are about applied, often controversial topics susceptible to disinformation: the environment, health, and computing. The resulting enhanced potential for malicious manipulation of society’s evidence base, particularly in politically... (More)
- Academic journals, archives, and repositories are seeing an increasing number of questionable research papers clearly produced using generative AI. They are often created with widely available, general-purpose AI applications, most likely ChatGPT, and mimic scientific writing. Google Scholar easily locates and lists these questionable papers alongside reputable, quality-controlled research. Our analysis of a selection of questionable GPT-fabricated scientific papers found in Google Scholar shows that many are about applied, often controversial topics susceptible to disinformation: the environment, health, and computing. The resulting enhanced potential for malicious manipulation of society’s evidence base, particularly in politically divisive domains, is a growing concern. (Less)
Please use this url to cite or link to this publication:
https://lup.lub.lu.se/record/6eadbf59-d5f7-4f62-b997-98a652d91e93
- author
- Haider, Jutta LU ; Söderström, Kristofer Rolf LU ; Ekström, Björn and Rödl, Malte
- organization
- publishing date
- 2024-09-03
- type
- Contribution to journal
- publication status
- published
- subject
- in
- Misinformation Review
- ISSN
- 2766-1652
- DOI
- 10.37016/mr-2020-156
- project
- Why do we choose the Internet instead of the doctor next door? The Internet as a site for medicines in grey zones.
- language
- English
- LU publication?
- yes
- id
- 6eadbf59-d5f7-4f62-b997-98a652d91e93
- date added to LUP
- 2024-09-04 10:37:16
- date last changed
- 2024-09-13 09:55:53
@article{6eadbf59-d5f7-4f62-b997-98a652d91e93, abstract = {{Academic journals, archives, and repositories are seeing an increasing number of questionable research papers clearly produced using generative AI. They are often created with widely available, general-purpose AI applications, most likely ChatGPT, and mimic scientific writing. Google Scholar easily locates and lists these questionable papers alongside reputable, quality-controlled research. Our analysis of a selection of questionable GPT-fabricated scientific papers found in Google Scholar shows that many are about applied, often controversial topics susceptible to disinformation: the environment, health, and computing. The resulting enhanced potential for malicious manipulation of society’s evidence base, particularly in politically divisive domains, is a growing concern.}}, author = {{Haider, Jutta and Söderström, Kristofer Rolf and Ekström, Björn and Rödl, Malte}}, issn = {{2766-1652}}, language = {{eng}}, month = {{09}}, series = {{Misinformation Review}}, title = {{GPT-fabricated scientific papers on Google Scholar : Key features, spread, and implications for preempting evidence manipulation}}, url = {{http://dx.doi.org/10.37016/mr-2020-156}}, doi = {{10.37016/mr-2020-156}}, year = {{2024}}, }