Preview

Proceedings of the Institute for System Programming of the RAS (Proceedings of ISP RAS)

Advanced search
Vol 35, No 6 (2023)
View or download the full issue PDF (Russian)
7-28
Abstract

In this study, the relationship between the improvement of software requirement quality and the software product quality in use was explored and analyzed. Analysis was based on the design of software product quality-in-use and the measure of metrics from ISO/IEC 25010 standard in two software products. The results show that the validation activities introduced in the software requirements stage have a positive relationship with the quality in use of the software products analyzed. In the software studied, it can be said that the improvement of the quality of the requirements has contributed to the improvement of the quality in use of software products.

29-42
Abstract

Any software development project needs to estimate Non-Functional Requirements (NFR). Typically, software managers are forced to use expert judgment to estimate the NFR. Today, NFRs cannot be measured, as there is no standardized unit of measurement for them. Consequently, most estimation models focus on the Functional User Requirements (FUR) and do not consider the NFR in the estimation process because these terms are often subjective. The objective of this paper was to show how an NFR estimation model was created using fuzzy logic, and K-Nearest Neighbors classifier algorithm, aiming to consider the subjectivity embedded in NFR terms to solve a specific problem in a Mexican company. The proposed model was developed using a database with real projects from a Mexican company in the private sector.

43-74
Abstract

Software development can be a time-consuming and costly process that requires a significant amount of effort. Developers are often tasked with completing programming tasks or making modifications to existing code without increasing overall complexity. It is essential for them to understand the dependencies between the program components before implementing any changes. However, as code evolves, it becomes increasingly challenging for project managers to detect indirect coupling links between components. These hidden links can complicate the system, cause inaccurate effort estimates, and compromise the quality of the code. To address these challenges, this study aims to provide a set of measures that leverage measurement theory and hidden links between software components to expand the scope, effectiveness, and utility of accepted software metrics. The research focuses on two primary topics: (1) how indirect coupling measurements can aid developers with maintenance tasks and (2) how indirect coupling metrics can quantify software complexity and size, leveraging weighted differences across techniques. The study presents a comprehensive set of measures designed to assist developers and project managers with project management and maintenance activities. Using the power of indirect coupling measurements, these measures can enhance the quality and efficiency of software development and maintenance processes.

75-94
Abstract

Agile software development companies considered small entities (VSE) face a new reality of remote development. Remote communication has generated many videos derived from video calls recorded for later reference. The video calls recorded contains architectural knowledge from virtual meetings and is essential for companies facing the knowledge vaporization problem. However, only some proposals in the literature can potentially manage AK in videos. This article proposes a solution to recover this architectural knowledge contained in videos using an ontology as a classification scheme, following the architectural knowledge condensation concept. We validated our proposal to manage architectural knowledge following the Methontology guidelines. Implementing an ontology as a classification scheme represents a step forward to achieving the condensation of architectural knowledge in an agile development environment for VSE.

95-102
Abstract

Teaching students compilation and transformation technologies is in demand. The article discusses the criteria for choosing a programming language for practical work as a target language, gives a brief description of the Trivil language and examines the suitability of this language for using in accordance with these criteria.

103-120
Abstract

The paper describes a unified representation for an abstract syntax tree (AST) suitable for static analysis of several programming languages. The proposed analysis scheme consists of saving an intermediate representation in the form of a unified AST from compilers of the corresponding languages and subsequent analysis of the saved trees. We have implemented this described representation for Java, Kotlin and Python. The unified AST analyzer has 27 checkers. In the paper we present structure and entities of our unified AST, provide more details regarding language specifics that have to be reflected in the UAST representation. We give extensive experimental results that show UAST generation and analysis speed, analysis quality, and comparison with the old scheme of analyzing compiler ASTs where applicable. As a result, we see that we observe some degradation of analysis speed, but we pay it for the separation of AST construction and checkers’ implementation. This separation allows easier support of many languages in the analyzer, where one can just generate UAST and support the required checker once within the UAST infrastructure instead of implementing a checker once per language.

121-134
Abstract

The reason of undefined behavior is source code written in violation of the C language standard. Undefined behavior leads to vulnerabilities in software. One of the common sources of undefined behavior is an incorrect use of functions for nonlocal jumps (in particular setjmp and longjmp). This paper considers the means of detecting this type of undefined behavior which are implemented in the major modern compilers (GCC, Clang, MSVC). We conclude that these means either have significant disadvantages or are absent in some compilers. This paper presents the implementation of a new method of compiler warning of the considered undefined behavior. The described method is accurate enough for practical application on real projects. We consider the advantages of the proposed solution over similar existing ones.

135-156
Abstract

Data races are a class of concurrency errors where two threads access a shared memory location without proper synchronization. Data races are hard to reveal and debug. This paper presents RaceHunter - a dynamic data race detection technique which monitors executions of shared memory concurrent programs, discovers pairs of conflicting memory accesses and systematically verifies them for data races. RaceHunter does not report false data races when the target software exploits non-standard synchronization primitives or unknown synchronization protocols and can find data races missed by other techniques. Dynamic data race detectors can monitor continuous, e.g. real-life, program executions or they can verify relatively short program executions, e.g. organized by system tests. The latter is the primary use case scenario for RaceHunter.

157-166
Abstract

When highlighting the logical structure of documents, a number of properties are used, one of which is the bold style of text words. In documents, headings, defined words, and column names in tables are often highlighted in bold. This paper proposes a method for classifying text by boldness, which consists of a sequence of steps. The first step is binarization of the entire image. The purpose of this step is to separate the image pixels into text and background pixels. The second step is to evaluate each word. The result is returned a value characterizing the thickness of the main stroke of the character in the given word. At the last step, the ratings are clustered into two clusters: bold text and regular. The proposed method was implemented and tested on three data sets, and the source code was published in an open repository.

167-178
Abstract

The paper describes a tool for testing the security of cryptographic protocol implementations working on the basis of specifications written in a declarative interoperable domain-specific language implemented as EDSL (Embedded [in Haskell] DSL). The problem of forming high-quality reference test suites for testing the security of cryptoprotocol implementations is considered. A method of addressing this problem within the tool being developed is discussed.

179-188
Abstract

We propose a membership inference attack against the neural classification algorithm from the draft national standard developed by the Omsk State Technical University under the auspices of the Technical Committee on Standardization «Artificial Intelligence» (TC 164). The attack allows us to determine whether the data were used for neural network training, and aimed at violating the confidentiality property of the training set. The results show that the protection mechanism of neural network classifiers described by the draft national standard does not provide the declared properties. The results were previously announced at Ruscrypto’2023 conference.

189-198
Abstract

The open IC design flow OpenLane make available the RTL-to-GDS design flow to be implemented, but still some tasks remain unsolved. An example of such task is the correction of antenna rules violations. The detection tools for this procedure are the part of the OpenLane flow but it does not contain any tools for avoiding them. In this article the software is presented that has been developed and allows to avoid violations of antenna rules based on input data in the form of LEF file with information about the standard cell library and DEF file with information about the placement and routing of the design. A parallel implementation is also described; we show the results of more than 60% gain in time with using parallel version in comparison with sequential version. The article describes both the developed algorithm and the software built on its basis, the capability of embedding into an open OpenLane flow, and an example of embedding script code. The use of the developed software makes it possible to correct a significant part of antenna rules violations, thereby increasing the yield.

199-212
Abstract

The article discusses machine learning algorithms for predicting turbulent viscosity using the case of flow past the backward-facing step. The training data is obtained by calculations using the OpenFOAM software package and a  turbulence model. The significance of flow parameters, including velocity fluctuations, pressure and velocity gradients, strain rate tensor and their combinations and invariants are analyzed for predicting turbulent viscosity. Different machine learning algorithms are compared. It is found that the most optimal algorithm for predicting turbulent viscosity in this case is the Decision Tree Regressor. Using the chosen model, the distribution of turbulent viscosity in the computational domain is predicted for various Reynolds numbers.

213-234
Abstract

We consider rotation-based fast multipole method for the Laplace equation and its application in cases where particle interactions are governed by the Biot—Savart law. The paper presents the necessary formulas for algorithm implementation and addresses less frequently discussed topics, such as the normalization of spherical harmonics and the associated Wigner matrices normalization. The main focus of the paper is devoted to describing the details of the software implementation that significantly accelerate the performance of the code both on CPU and GPU (using CUDA technology). We provide a comprehensive explanation of proposed techniques and include code examples. We have implemented a C++ program using these methods and conducted a comparative analysis with open-source implementations of the fast multipole method, confirming the high efficiency of our approach.

235-246
Abstract

The work provides a comparative analysis of the results of numerical modeling and experimental studies of fluid flow in a heat exchanger (HE). The OpenFOAM package for solving CFD problems was used as a tool for numerical modeling. Experimental studies were carried out on the TMJ-2M hydraulic stand. The heat exchanger was connected directly to the modules of the hydraulic stand. Distilled water at room temperature was used as the working fluid. Verification of the experimental results was carried out on the basis of numerical modeling in the OpenFOAM package using the same characteristics of the media used in the field experiment. Analysis of the obtained data showed good convergence of the results of the numerical and experimental study of the multi-section HE.

247-264
Abstract

This paper proposes the implementation of a neural network training approach for object detection, using augmentation - style transfer. This method improves the generalization ability of the neural network to determine the location of objects in the image by improving the interaction with low-level features such as textures, different colors and small changes in shapes. The effectiveness of the method is experimentally proved and the numerical values of the object detection metrics are demonstrated on several datasets with different classes. The application of augmentation is proposed using an unused before neural network architecture capable of carrying an arbitrary number of styles. The peculiarity of the approach is also that the weights of the neural network for styling are frozen and it is added to the graph of the detection network, which allows augmentation speed.

265-282
Abstract

This article presents the results of an analysis of three dictionaries by G.F. Miller, which were created in the 18th century and are hosted on the LingvoDoc platform (lingvodoc.ispras.ru). Miller’s manuscript dictionaries from the collection of the Russian State Archive of Ancient Documents (RSAAD) are introduced into scholarly discourse for the first time. The analysis reveals: 1) in the 18th century, all consonant transitions in the Tatar language had already been completed; 2) at this time, some vowels (known as the Volga region vowels) had not yet completed their transition process, and for [*o] and [*u], this process had not even begun; 3) in the 18th century, dialectal zones were already identified in the Tatar language; 4) the 1791 period Cyrillic-script dictionary differs in innovations from modern Tatar language and Latin-script dictionaries; 5) Latin-script dictionaries are closer to the modern Tatar literary language.

283-292
Abstract

The purpose of the research is to study the peculiarities of the structure of the sound of Russian speech by Chinese and Ghanaian students at different levels of study, based on experimental phonetic programs. The subject of the study is the intonational structure of the main communicative intentions: greeting, address, completion, incompletion, a question without an interrogative word, a question with the Russian conjunction «а» (“and”), a positive evaluation and surprise. The study contributes to the creation of an ethno-oriented methodology for teaching Russian intonation. The scientific novelty is in the analysis of the sound of the Russian speech of foreign students at the syntagma level in comparison with that of a native Russian speaker using the Praat and Lingvodoc programs. The results of the study consist in the visual representation of the peculiarities of the structure of the main communicative intentions in the Russian speech of Chinese and Ghanaians to predict and correct errors at the level of intonation.

293-310
Abstract

The article presents the results of computer phonetic-etymological and glottochronological analysis of data from the dialect of the village of Bobya-Ucha, Malopurginsky District, Udmurt Republic. As far as we know, the dialectal peculiarities of the studied material have not previously attracted the attention of researchers. In the course of work under the RSF grant, audio recordings of two dictionaries of the colloquial language were collected. The vocabulary of which is still used in everyday life. At the present days, the audio materials of the dictionaries were transcribed and placed on the LingvoDoc platform (lingvodoc.ispras.ru), which made it possible to establish etymological links with the basic dictionaries of Udmurt dialects. Then, with the help of special programs, the divergence of the Bobya-Ucha dialect from the Udmurt literary language was analyzed. The phonetic peculiarities of the studied dialect were revealed and its dialectal affiliation was clarified. The comparative-historical analysis of phonetic peculiarities has shown that the consonant phonemes of the studied dialect have not undergone any changes and have no dialectal peculiarities. However, there is a specific sound ’ in the Bobya-Ucha dialect, the study of which requires further research. The vocalic system of the studied vernacular is determined by several changes in the acoustic and articulatory properties of individual phonemes: *ɪ̈ > ə, *ɐ > ə, *ɞ > ɜ. Thus, based on the etymological-phonetic and glottochronological analysis (96% coincidence), we can see that the dialect of the village Bobya-Ucha belongs to the central-southern dialect of the Udmurt language.

311-330
Abstract

The article analyzed the glossed corpus and concordance of the books “Sacred History ...” and Indication of the “Way to the Kingdom of Heaven ...”, which are publicly available on the LingvoDoc platform located at lingvodoc.ispras.ru. The books are samples of a weakly normalized young–written Shor language - a written version of the Kondom (Shor) dialect. The corpus revealed cases of differences in the language of the first books from the modern language that had not been previously described: a multivariate form of the present tense, combining analytical (kӧlep char-ï ‘bestows'), hyphenated, transitional to synthetic (pildir–char-ï ‘means’), and synthetic (aitchar or aitchar-ï ‘says’) verbal constructions with the same indicator (–)char(ï); elimination of the third person singular affix -ï in synthetic forms of the present tense on (–)char(ï); the processes of synthesizing this time form with an unstable spelling. Our task was to analyze the regularity of these phenomena, as well as the introduction into scientific circulation of materials of the little-studied young-written Shor language, which is now in danger of extinction, as well as the replenishment of the automatic Parser of the Shor language with the missing affixes of multivariate verb forms. It was found that there is a single form of present, with the affix –char, as a more complete form of the present tense affix -cha, formed from the verb chat– ‘to lie’; in analytical constructions, word forms with affix –ï predominate (83.3%), word forms without affix make up only 1.7%, in synthetic forms, the number of word forms without affix is already 45.4%. The rule of falling off -(I)p in the bases ending in a consonant has not yet been normalized and works sporadically; the spelling of the fused-separately multivariate form of the present tense on (–)char has not been normalized either.

331-336
Abstract

The Mongolian languages are a language family that includes several closely related languages of Mongolia, China, Russia and Afghanistan. According to lexicostatistics, they broke up around the 5th century AD. Dagur, Shira-Yugur, Dongxiang, Bao'an, Tu (Monguor) are commonly referred to as South Mongolian languages. The South Mongolian languages, in comparison with other groups of Mongolian languages, were influenced by Chinese and Turkic languages, various historical events had an undoubted influence on them.

337-346
Abstract

The solution to the problem of multilingualism in the international scientific and technical information space is connected with machine translation (MP) technologies. The translation process within the framework of the concept of phraseological conceptual translation of texts can be represented as the process of transmitting the semantic content of the source text by means of the output language. Within the framework of this concept, the translation of texts is provided by the conceptual analysis of the source text and the transformation of its semantic content into the target language. This approach is based on understanding the laws of the functioning of natural languages and theoretical ideas about the semantic structure of texts. The basis of the approach is the technology of automated formation of thematic dictionary bases that adequately reflect the conceptual composition of various thematic areas.



Creative Commons License
This work is licensed under a Creative Commons Attribution 4.0 License.


ISSN 2079-8156 (Print)
ISSN 2220-6426 (Online)