Please use this identifier to cite or link to this item: https://dspace.ncfu.ru/handle/123456789/30519
Title: Machine Learning Research Methods for Identifying Inaccurate Content
Authors: Багаутдинова, А. Р.
Lapina, M. A.
Лапина, М. А.
Lapin, V. A.
Лапин, В. А.
Bagautdinova, A. R.
Keywords: Artificial intelligence;Social networks;Authenticity;Data analysis;Deception recognition;Deep learning;Facial expression;Lie detection;Fake news;Machine learning;Neural networks
Issue Date: 2025
Publisher: Springer Science and Business Media Deutschland GmbH
Citation: Lapina M., Anita M., Bagautdinova A., Lapin V., Rudenko M. Machine Learning Research Methods for Identifying Inaccurate Content // Lecture Notes in Networks and Systems. - 2025. - 1295. - pp. 193 - 201. - DOI: 10.1007/978-981-96-3311-1_16
Series/Report no.: Lecture Notes in Networks and Systems
Abstract: Social media, especially when disseminating news, is a valuable information resource. The paper presents methods for detecting fake news, comparing their effectiveness, identifying existing problems, and describes the vectors of further development of this research area. The paper begins with a description of the relevance of the Fake News problem, which clearly describes the negative impact of false news on all spheres of human life. The following is a description of methods for detecting false news, starting from the usual rules of text analysis and ending with complex ML algorithms. In this paper, a comparative analysis of detection methods is carried out, which is based on criteria of efficiency and accuracy. The author identifies the main problems of existing methods related to data quality, changing Fake News formats and the difficulties of automatically determining the reliability of information.
URI: https://dspace.ncfu.ru/handle/123456789/30519
Appears in Collections:Статьи, проиндексированные в SCOPUS, WOS

Files in This Item:
File SizeFormat 
scopusresults 3586.pdf
  Restricted Access
127.11 kBAdobe PDFView/Open


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.