The impact of widely used algorithms of large language models on information retrieval and the associated potential risks

Autor

  • Jakub Lichosik Łukasiewicz Research Network – Institute of Aviation Aleja Krakowska 110/114, 02-256 Warszawa, Polan
  • Iwona Zborowska Internist nursing facility, Faculty of Health Sciences, Medical University of Wrocław ul. Bartla 5, 51-618 Wrocław, Poland https://orcid.org/0000-0001-9495-9957
  • Anna Dąbek Internist nursing facility, Faculty of Health Sciences, Medical University of Wrocław ul. Bartla 5, 51-618 Wrocław, Poland https://orcid.org/0000-0002-2179-2559
  • Angelika Dudek Institute of Law, Administration, Criminology and Security, Warszawa Collegium Humanum – Warsaw Management University ul. Aleje Jerozolimskie 133A, 02-304 Warszawa, Poland

DOI:

https://doi.org/10.15503/emet2023.93.101

Słowa kluczowe:

Artificial intelligence, Learning, Psychology, Manipulation

Abstrakt

Thesis. The aim of this work is to perform an analysis concerning the subject of arti- ficial intelligence integration and its application in modern knowledge acquisition. The article will be an overview of scientific research and papers to raise a discussion of the ad- vantages and potential hazards associated with the widespread implementation of large language models.

Concept. The main part of the article will focus on describing the AI enhanced tools available on the market and common ways of utilising them. This knowledge will be jux- taposed with psychological research on human consciousness in the context of decision- making, creativity, and held beliefs to list potential threads related.

Results. The analysis of tools that have recently been enhanced with algorithms un- equivocally indicates the dominance of the sector providing informational and creative support. These tools possess a measurably high level of reliability among the public, re- sulting in a reduced number of verification actions. A review of existing literature and research on human psychology shows a very strong correlation between the influence of previous social authorities on decision-making behaviour and an uncritical approach to information obtained through AI. Studies have shown that access to these tools, without proper controlling actions of results, exposes users to bias, manipulation, or susceptibility to marketing activities. Alternatively, with proper and rational use of these implemented algorithms, humans are able to obtain extraordinarily precise and knowledge-based sup- port in decision-making. Similar polarisations of results have been observed in the con- text of creativity, innovation, and inquisitiveness.

Originality. The article addresses topics that relate to the psychological nature of hu- mans in the context of new and expanded AI augmented tools. The results of the explora- tory study clearly indicate that programmes supported by algorithms show potential in terms of both scientific and social development. Unfortunately, they also present many potential dangers for both, which, discussed in the foundation, provide the groundwork for more in-depth research.

##plugins.generic.usageStats.downloads##

##plugins.generic.usageStats.noStats##

Biogram autora

Iwona Zborowska - Internist nursing facility, Faculty of Health Sciences, Medical University of Wrocław ul. Bartla 5, 51-618 Wrocław, Poland

 

Bibliografia

Ashoori, M., & Weisz, J. D. (2019). In AI we trust? Factors that influence trustworthiness of AI-infused decision-making processes.

Cherniak, K. (2024). Chatbot statistics: What businesses need to know about digital assistants. Master of code global.

Kaddour, J., Harris, J., Mozes, M., Bradley, H., Raileanu, R., & McHardy, R. (2023). Chal- lenges and applications of large language models.

Martin, H.-J. (1994). The History and Power of Writing.

Morrison, A. B., Conway, A. R. A., & Chein, J. M. (2014). Primacy and recency effects as

indices of the focus of attention. Front Hum Neurosci.

owen, J. M. (2007). The scientific article in the age of digitization.

Reddy, S., Allan, S., Coghlan, S., & Cooper, P. (2019). A governance model for the application

of AI in health care.

Rong, Q., Lian, Q., & Tang, T. (2022). Research on the influence of AI and VR technology for

students’ concentration and creativity.

Rotter, J. B. (1967). A new scale for the measurement of interpersonal trust. Journal of

Personality, 35(4), 651–665.

Simpson, J. A., & Vieth, G. (2022). Trust and psychology. In the neurobiology of trust (chapter

.

Suzuki, M., & Yamamoto, Y. (2021). Characterizing the influence of confirmation bias on

web search behavior. Frontiers in psychology.

Wagner, P. (2020). Cookies: Privacy risks, attacks, and recommendations.

Wang, D., Weisz, J. D., Muller, M., Ram, P., Geyer, W., Dugan, C., Tausczik, Y., Samulow-

itz, H., & Gray, A. (2019). Human AI collaboration in data science: Exploring data scientists’

perceptions of automated AI. Computer supported cooperative work.

Xu, Y., Liu, X., Cao, X., Huang, C., Liu, e., Qian, S., ... & Zhang, J. (2021). Artificial intel-

ligence: A powerful paradigm for scientific research.

Zawacki-Richter, o., Marín, V. I., Bond, M., & Gouverneur, F. (2019). Systematic review of

research on artificial intelligence applications in higher education – where are the educators?

Opublikowane

2024-10-27

Jak cytować

Lichosik, J., Zborowska, I., Dąbek, A., & Dudek, A. (2024). The impact of widely used algorithms of large language models on information retrieval and the associated potential risks. E-Methodology, 10(10), 93–101. https://doi.org/10.15503/emet2023.93.101

Numer

Dział

“On the Internet” – Research