×

You are using an outdated browser Internet Explorer. It does not support some functions of the site.

Recommend that you install one of the following browsers: Firefox, Opera or Chrome.

Contacts:

+7 961 270-60-01
ivdon3@bk.ru

  • Application of language neural network models for malware detection

    The growing popularity of large language models in various fields of scientific and industrial activity leads to the emergence of solutions using these technologies for completely different tasks. This article suggests using the BERT, GPT, and GPT-2 language models to detect malicious code. The neural network model, previously trained on natural texts, is further trained on a preprocessed dataset containing program files with malicious and harmless code. The preprocessing of the dataset consists in the fact that program files in the form of machine instructions are translated into a textual description in a formalized language. The model trained in this way is used for the task of classifying software based on the indication of the content of malicious code in it. The article provides information about the conducted experiment on the use of the proposed model. The quality of this approach is evaluated in comparison with existing antivirus technologies. Ways to improve the characteristics of the model are also suggested.

    Keywords: antivirus, neural network, language models, malicious code, machine learning, model training, fine tuning, BERT, GPT, GPT-2

  • Optimization of the dense matrix multiplication procedure for shared memory systems

    The study presents an extensive analysis of methods for low-level optimization of the matrix multiplication algorithm for computing systems with shared memory. Based on a comparison of various approaches, including block optimization, parallel execution with OpenMP, vectorization with AVX and the use of the Intel MKL library, significant improvements in the performance of the resulting software implementations are revealed. In particular, block optimization reduces the number of cache misses, parallelism effectively uses multicore, and vectorization and Intel MKL demonstrate maximum acceleration due to more efficient software optimizations. The obtained results emphasize the importance of careful selection of optimization methods and their compliance with the architecture of the computing system in order to achieve the required performance parameters of the designed software.

    Keywords: low-level optimization, block optimization, parallel execution, OpenMP, vectorization, AVX, Intel MKL, performance, benchmarking, matrix multiplication