Please use this identifier to cite or link to this item: https://dspace.ncfu.ru/handle/123456789/29360
Title: Development of an Approach to Confidential Learning with Errors in the Design of Neural Networks
Authors: Bezuglova, E. S.
Безуглова, Е. С.
Shiriaev, E. M.
Ширяев, Е. М.
Keywords: Cryptography;Matrix transposition;Machine learning;Neural networks
Issue Date: 2024
Publisher: Springer Science and Business Media Deutschland GmbH
Citation: Bezuglova, E., Shiriaev, E. Development of an Approach to Confidential Learning with Errors in the Design of Neural Networks // Lecture Notes in Networks and Systems. - 2024. - 1207 LNNS. - pp. 24-30. - DOI: 10.1007/978-3-031-77229-0_4
Series/Report no.: Lecture Notes in Networks and Systems
Abstract: In this paper, the presented method, applied to transpose densified weight matrices, involves tenth machine learning when moving from feedforward to backpropagation. The algorithm is based on the diagonal matrix construction method. The process of manufacturing change will significantly increase the number of operations using technology, especially in the ten linear transformations. The number of multiplications required for linear transformations depends on the dimensionality of the input and output data, but these differences are taken into account during the training process, which includes both forward and back propagation. The proposed method leads to improved training efficiency and computational efficiency in contextual machine learning.
URI: https://dspace.ncfu.ru/handle/123456789/29360
Appears in Collections:Статьи, проиндексированные в SCOPUS, WOS

Files in This Item:
File SizeFormat 
scopusresults 3363.pdf
  Restricted Access
132.02 kBAdobe PDFView/Open


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.