Vol 21, No 5 (2022)

Artificial intelligence, knowledge and data engineering

Random Survival Forests Incorporated by the Nadaraya-Watson Regression

Utkin L.V., Konstantinov A.V.

Abstract

An attention-based random survival forest (Att-RSF) is presented in the paper. The first main idea behind this model is to adapt the Nadaraya-Watson kernel regression to the random survival forest so that the regression weights or kernels can be regarded as trainable attention weights under important condition that predictions of the random survival forest are represented in the form of functions, for example, the survival function and the cumulative hazard function. Each trainable weight assigned to a tree and a training or testing example is defined by two factors: by the ability of corresponding tree to predict and by the peculiarity of an example which falls into a leaf of the tree. The second main idea behind Att-RSF is to apply the Huber's contamination model to represent the attention weights as the linear function of the trainable attention parameters. The Harrell's C-index (concordance index) measuring the prediction quality of the random survival forest is used to form the loss function for training the attention weights. The C-index jointly with the contamination model lead to the standard quadratic optimization problem for computing the weights, which has many simple algorithms for its solution. Numerical experiments with real datasets containing survival data illustrate Att-RSF.

Informatics and Automation. 2022;21(5):851-880
pages 851-880 views

Approach to Software Integration of Heterogeneous Sources of Medical Data Based on Microservice Architecture

Yusupova N.I., Vorobeva G.R., Zulkarneev R.K.

Abstract

The task of processing medical information is currently being solved in our country and abroad by means of heterogeneous medical information systems, mainly at the local and regional levels. The ever-increasing volume and complexity of the accumulated information, along with the need to ensure transparency and continuity in the processing of medical data (in particular, for bronchopulmonary diseases) in various organizations, requires the development of a new approach to integrating their heterogeneous sources. At the same time, an important requirement for solving the problem is the possibility of web-oriented implementation, which will make the corresponding applications available to a wide range of users without high requirements for their hardware and software capabilities. The paper considers an approach to the integration of heterogeneous sources of medical information, which is based on the principles of building microservice web architectures. Each data processing module can be used independently of other program modules, providing a universal entry point and the resulting data set in accordance with the accepted data schema. Sequential execution of processing steps implies the transfer of control to the corresponding program modules in the background according to the Cron principle. The schema declares two types of data schemas - local (from medical information systems) and global (for a single storage system), between which the corresponding display parameters are provided according to the principle of constructing XSLT tables. An important distinguishing feature of the proposed approach is the modernization of the medical information storage system, which consists in creating mirror copies of the main server with periodic replication of the relevant information. At the same time, the interaction between clients and data storage servers is carried out according to the type of content delivery systems with the creation of a connection session between end points based on the principle of the nearest distance between them, calculated using the haversine formula. The computational experiments carried out on test data on bronchopulmonary diseases showed the effectiveness of the proposed approach both for loading data and for obtaining them by individual users and software systems. Overall, the reactivity score of the corresponding web-based applications was improved by 40% on a stable connection.
Informatics and Automation. 2022;21(5):881-915
pages 881-915 views

Opening the Black Box: Finding Osgood’s Semantic Factors in Word2vec Space

Surov I.A.

Abstract

State-of-the-art models of artificial intelligence are developed in the black-box paradigm, in which sensitive information is limited to input-output interfaces, while internal representations are not interpretable. The resulting algorithms lack explainability and transparency, requested for responsible application. This paper addresses the problem by a method for finding Osgood’s dimensions of affective meaning in multidimensional space of a pre-trained word2vec model of natural language. Three affective dimensions are found based on eight semantic prototypes, composed of individual words. Evaluation axis is found in 300-dimensional word2vec space as a difference between positive and negative prototypes. Potency and activity axes are defined from six process-semantic prototypes (perception, analysis, planning, action, progress, and evaluation), representing phases of a generalized circular process in that plane. All dimensions are found in simple analytical form, not requiring additional training. Dimensions are nearly orthogonal, as expected for independent semantic factors. Osgood’s semantics of any word2vec object is then retrieved by a simple projection of the corresponding vector to the identified dimensions. The developed approach opens the possibility for interpreting the inside of black box-type algorithms in natural affective-semantic categories, and provides insights into foundational principles of distributive vector models of natural language. In the reverse direction, the established mapping opens machine-learning models as rich sources of data for cognitive-behavioral research and technology.

Informatics and Automation. 2022;21(5):916-936
pages 916-936 views

Verification of Marine Oil Spills Using Aerial Images Based on Deep Learning Methods

Favorskaya M.N., Nishchhal N.

Abstract

The article solves the problem of verifying oil spills on the water surfaces of rivers, seas and oceans using optical aerial photographs, which are obtained from cameras of unmanned aerial vehicles, based on deep learning methods. The specificity of this problem is the presence of areas visually similar to oil spills on water surfaces caused by blooms of specific algae, substances that do not cause environmental damage (for example, palm oil), or glare when shooting (so-called look-alikes). Many studies in this area are based on the analysis of synthetic aperture radars (SAR) images, which do not provide accurate classification and segmentation. Follow-up verification contributes to reducing environmental and property damage, and oil spill size monitoring is used to make further response decisions. A new approach to the verification of optical images as a binary classification problem based on the Siamese network is proposed, when a fragment of the original image is repeatedly compared with representative examples from the class of marine oil slicks. The Siamese network is based on the lightweight VGG16 network. When the threshold value of the output function is exceeded, a decision is made about the presence of an oil spill. To train the networks, we collected and labeled our own dataset from open Internet resources. A significant problem is an imbalance of classes in the dataset, which required the use of augmentation methods based not only on geometric and color manipulations, but also on the application of a Generative Adversarial Network (GAN). Experiments have shown that the classification accuracy of oil spills and look-alikes on the test set reaches values of 0.91 and 0.834, respectively. Further, an additional problem of accurate semantic segmentation of an oil spill is solved using convolutional neural networks (CNN) of the encoder-decoder type. Three deep network architectures U-Net, SegNet, and Poly-YOLOv3 have been explored for segmentation. The Poly-YOLOv3 network demonstrated the best results, reaching an accuracy of 0.97 and an average image processing time of 385 s with the Google Colab web service. A database was also designed to store both original and verified images with problem areas.
Informatics and Automation. 2022;21(5):937-962
pages 937-962 views

Deep Transfer Learning of Satellite Imagery for Land Use and Land Cover Classification

Yifter T.T., Razoumny Y.N., Lobanov V.K.

Abstract

Deep learning has been instrumental in solving difficult problems by automatically learning, from sample data, the rules (algorithms) that map an input to its respective output. Purpose: Perform land use landcover (LULC) classification using the training data of satellite imagery for Moscow region and compare the accuracy attained from different models. Methods: The accuracy attained for LULC classification using deep learning algorithm and satellite imagery data is dependent on both the model and the training dataset used. We have used state-of-the-art deep learning models and transfer learning, together with dataset appropriate for the models. Different methods were applied to fine tuning the models with different parameters and preparing the right dataset for training, including using data augmentation. Results: Four models of deep learning from Residual Network (ResNet) and Visual Geometry Group (VGG) namely: ResNet50, ResNet152, VGG16 and VGG19 has been used with transfer learning. Further training of the models is performed with training data collected from Sentinel-2 for the Moscow region and it is found that ResNet50 has given the highest accuracy for LULC classification for this region. Practical relevance: We have developed code that train the 4 models and make classification of the input image patches into one of the 10 classes (Annual Crop, Forest, Herbaceous Vegetation, Highway, Industrial, Pasture, Permanent Crop, Residential, River, and Sea&Lake).

Informatics and Automation. 2022;21(5):963-982
pages 963-982 views

Digital information telecommunication technologies

Analysis of the Correlation Properties of the Wavelet Transform Coefficients of Typical Images

Dvornikov S.V., Dvornikov S.S., Ustinov A.A.

Abstract

The increasing flow of photo and video information transmitted through the channels of infocommunication systems and complexes stimulates the search for effective compression algorithms that can significantly reduce the volume of transmitted traffic, while maintaining its quality. In the general case, the compression algorithms are based on the operations of converting the correlated brightness values of the pixels of the image matrix into their uncorrelated parameters, followed by encoding the obtained conversion coefficients. Since the main known decorrelating transformations are quasi-optimal, the task of finding transformations that take into account changes in the statistical characteristics of compressed video data is still relevant. These circumstances determined the direction of the study, related to the analysis of the decorrelating properties of the generated wavelet coefficients obtained as a result of multi-scale image transformation. The main result of the study was to establish the fact that the wavelet coefficients of the multi-scale transformation have the structure of nested matrices defined as submatrices. Therefore, it is advisable to carry out the correlation analysis of the wavelet transformation coefficients separately for the elements of each submatrix at each level of decomposition (decomposition). The main theoretical result is the proof that the core of each subsequent level of the multi-scale transformation is a matrix consisting of the wavelet coefficients of the previous level of decomposition. It is this fact that makes it possible to draw a conclusion about the dependence of the corresponding elements of neighboring levels. In addition, it has been found that there is a linear relationship between the wavelet coefficients within the local area of ​​the image with a size of 8×8 pixels. In this case, the maximum correlation of submatrix elements is directly determined by the form of their representation, and is observed between neighboring elements located, respectively, in a row, column or diagonally, which is confirmed by the nature of the scattering. The obtained results were confirmed by the analysis of samples from more than two hundred typical images. At the same time, it is substantiated that between the low-frequency wavelet coefficients of the multi-scale transformation of the upper level of the expansion, approximately the same dependences are preserved uniformly in all directions. The practical significance of the study is determined by the fact that all the results obtained in the course of its implementation confirm the presence of characteristic dependencies between the wavelet transform coefficients at different levels of image decomposition. This fact indicates the possibility of achieving higher compression ratios of video data in the course of their encoding. The authors associate further research with the development of a mathematical model for adaptive arithmetic coding of video data and images, which takes into account the correlation properties of wavelet coefficients of a multi-scale transformation.
Informatics and Automation. 2022;21(5):983-1015
pages 983-1015 views

Discrete Time Sequence Reconstruction of a Signal Based on Local Approximation Using a Fourier Series by an Orthogonal System of Trigonometric Functions

Yakimov V.N.

Abstract

The article considers the development of mathematical and algorithmic support for the sample’s reconstruction in problem sections of a discrete sequence of a continuous signal. The work aimed to ensure the reconstruction of lost samples or sections of samples with a non-constant distorted time grid when sampling a signal with a uniform step and at the same time to reduce the computational complexity of digital reconstruction algorithms. The solution to the stated problem is obtained based on the local approximation method. The specific of this method application was the use of two subsequences of samples located symmetrically concerning the reconstructed section of the sequence. The approximating model is a Fourier series on an orthogonal system of trigonometric functions. The optimal solution to the approximation problem is based on the minimum square error criterion. Mathematical equations are obtained for this type of error. They allow us to estimate its value depending on the model order and the samples number in the subsequences used in the reconstruction process. The peculiarity of the mathematical equations obtained in this paper for signal reconstruction is that they do not require the preliminary calculation of the Fourier series coefficients. They provide a direct calculation of the values of reconstructed samples. At the same time, when the number of samples in the subsequences used for reconstruction will be even, it is not necessary to perform multiplication operations. All this made it possible to reduce the computational complexity of the developed algorithm for signal reconstruction. Experimental studies of the algorithm were carried out based on simulation modeling using a signal model that is an additive sum of harmonic components with a random initial phase. Numerical experiments have shown that the developed algorithm provides the reconstruction result of signal samples with a sufficiently low error. The algorithm is implemented as a software module. The operation of the module is carried out on the basis of asynchronous control of the sampling reconstruction process. It can be used as part of metrologically significant software for digital signal processing systems.
Informatics and Automation. 2022;21(5):1016-1043
pages 1016-1043 views

The Statistical Analysis of the Security for a Wireless Communication System with a Beaulieu-Xie Shadowed Fading Model Channel

Gvozdarev A.S., Artemova T.K., Patralov P.E., Murin D.M.

Abstract

The paper considers the problem of a wireless communication system’s physical level security for a multipath signal propagation channel and the presence of a wiretap channel. To generalize the propagation effects, the Beaulieu-Xie shadowed channel model was assumed. To describe the security of the information transfer process, such a metric as the secure outage probability of was considered. An analytical expression of the secure outage probability was obtained and analyzed depending on the characteristics of the channel and the communication system: the average value of the signal-to-noise ratio in the main channel and the wiretap channel, the effective path-loss exponent, the relative distance between the legitimate receiver and the wiretap and the threshold bandwidth, normalized to the bandwidth of a smooth Gaussian channel. The analysis considers the sets of parameters that cover all practically important scenarios for the functioning of a wireless communication systems: both deep fading (corresponding to the hyper-Rayleigh scenario) and small fading, both in the case of a significant line-of-sight component and a significant number of multipath clusters, and with significant shadowing of the dominant component and a small number of multipath waves, as well as all intermediate options. It is found out that the value of the energy requirements for guaranteed secure communication at a given speed is determined primarily by the power of multipath components, as well as the existence of an irreducible secure outage probability of a communication session with an increase for channels with strong overall shadowing of the signal components, which from a practical point of view is important to take into account when imposing requirements for the values of the signal-to-noise ratio and the data transfer rate in the direct channel, providing the desired degree of security of the wireless communication session.
Informatics and Automation. 2022;21(5):1044-1078
pages 1044-1078 views

Согласие на обработку персональных данных с помощью сервиса «Яндекс.Метрика»

1. Я (далее – «Пользователь» или «Субъект персональных данных»), осуществляя использование сайта https://journals.rcsi.science/ (далее – «Сайт»), подтверждая свою полную дееспособность даю согласие на обработку персональных данных с использованием средств автоматизации Оператору - федеральному государственному бюджетному учреждению «Российский центр научной информации» (РЦНИ), далее – «Оператор», расположенному по адресу: 119991, г. Москва, Ленинский просп., д.32А, со следующими условиями.

2. Категории обрабатываемых данных: файлы «cookies» (куки-файлы). Файлы «cookie» – это небольшой текстовый файл, который веб-сервер может хранить в браузере Пользователя. Данные файлы веб-сервер загружает на устройство Пользователя при посещении им Сайта. При каждом следующем посещении Пользователем Сайта «cookie» файлы отправляются на Сайт Оператора. Данные файлы позволяют Сайту распознавать устройство Пользователя. Содержимое такого файла может как относиться, так и не относиться к персональным данным, в зависимости от того, содержит ли такой файл персональные данные или содержит обезличенные технические данные.

3. Цель обработки персональных данных: анализ пользовательской активности с помощью сервиса «Яндекс.Метрика».

4. Категории субъектов персональных данных: все Пользователи Сайта, которые дали согласие на обработку файлов «cookie».

5. Способы обработки: сбор, запись, систематизация, накопление, хранение, уточнение (обновление, изменение), извлечение, использование, передача (доступ, предоставление), блокирование, удаление, уничтожение персональных данных.

6. Срок обработки и хранения: до получения от Субъекта персональных данных требования о прекращении обработки/отзыва согласия.

7. Способ отзыва: заявление об отзыве в письменном виде путём его направления на адрес электронной почты Оператора: info@rcsi.science или путем письменного обращения по юридическому адресу: 119991, г. Москва, Ленинский просп., д.32А

8. Субъект персональных данных вправе запретить своему оборудованию прием этих данных или ограничить прием этих данных. При отказе от получения таких данных или при ограничении приема данных некоторые функции Сайта могут работать некорректно. Субъект персональных данных обязуется сам настроить свое оборудование таким способом, чтобы оно обеспечивало адекватный его желаниям режим работы и уровень защиты данных файлов «cookie», Оператор не предоставляет технологических и правовых консультаций на темы подобного характера.

9. Порядок уничтожения персональных данных при достижении цели их обработки или при наступлении иных законных оснований определяется Оператором в соответствии с законодательством Российской Федерации.

10. Я согласен/согласна квалифицировать в качестве своей простой электронной подписи под настоящим Согласием и под Политикой обработки персональных данных выполнение мною следующего действия на сайте: https://journals.rcsi.science/ нажатие мною на интерфейсе с текстом: «Сайт использует сервис «Яндекс.Метрика» (который использует файлы «cookie») на элемент с текстом «Принять и продолжить».