Preview

Proceedings of the Institute for System Programming of the RAS (Proceedings of ISP RAS)

Advanced search
Vol 31, No 2 (2019)
View or download the full issue PDF (Russian)
7-13
Abstract

This special issue contains selected papers that had been submitted to Proceedings of the Institute for System Programming of the Russian Academy of Sciences. Thirteen submissions from nine countries (England, Mexico, China, Uruguay, Spain, Pakistan, Cuba, Dominican Republic, and Russia) cover several important topics in rapidly expanding area of research and development related with Advanced Computing. Authors show a spectrum of approaches to solve complex problems such as: data-oriented scheduling, scientific workflows, cloud computing, evolutionary algorithms, content distribution networks, soft computing, parallel programming model for multicore machines, high performance computing, data mining, software birth-marking, anomaly detection, swarm robotics, neural networks, machine learning, security, secret-sharing schemes, heterogeneous distributed computing, and Internet of Things.

15-20
Abstract

This article presents the application of soft computing methods for solving the problem of designing and optimizing cloud-based Content Distribution Networks (CDN). A multi-objective approach is applied to solve the resource provisioning problem for building the infrastructure for the network, considering the objectives of minimizing the cost of the virtual machines, network, and storage, and the maximization of the quality-of-service provided to end-users. A specific brokering model is proposed to allow a single cloud-based CDN to be able to host multiple content providers applying a resource sharing strategy. Following the proposed brokering model, three multiobjective evolutionary approaches are studied for the offline optimization of resource provisioning and a greedy heuristic method is proposed for addressing the online routing of contents. The experimental evaluation of the proposed approach is performed over a set of realistic problem instances. The obtained experimental results indicate that the proposed approach is effective for designing and optimizing cloud-based Content Distribution Networks: total costs are reduced by up to 10.34% while maintaining high quality-of-service values.

21-31
Abstract

This article presents the application of Virtual Savant to solve resource allocation problems, a widely-studied area with several real-world applications. Virtual Savant is a novel soft computing method that uses machine learning techniques to compute solutions to a given optimization problem. Virtual Savant aims at learning how to solve a given problem from the solutions computed by a reference algorithm, and its design allows taking advantage of modern parallel computing infrastructures. The proposed approach is evaluated to solve the Knapsack Problem, which models different variant of resource allocation problems, considering a set of instances with varying size and difficulty. The experimental analysis is performed on an Intel Xeon Phi many-core server. Results indicate that Virtual Savant is able to compute accurate solutions while showing good scalability properties when increasing the number of computing resources used.

33-39
Abstract
Early warning during sky survey provides a crucial opportunity to detect low-mass, free-floating planets. In particular, to search short-timescale microlensing (ML) events from high-cadence and wide- field survey in real time, a hybrid method which combines ARIMA (Autoregressive Integrated Moving Average) with LSTM (Long-Short Time Memory) and GRU (Gated Recurrent Unit) recurrent neural networks (RNN) is presented to monitor all observed light curves and identify ML events at their early stages. Experimental results show that the hybrid models perform better in accuracy and less time consuming of adjusting parameters. ARIMA+LSTM and ARIMA+GRU can achieve improvement in accuracy by 14.5% and 13.2%, respectively. In the case of abnormal detection of light curves, GRU can achieve almost the same result as LSTM with less time by 8%. Same models are also applied to MIT-BIH Arrhythmia Databases ECG dataset with similar abnormal pattern and we yield from both sets that we can reduce up to 40% of time consuming for researchers to adjust the model to 90% accuracy.
41-52
Abstract

The article deals with the search for the global extremum in the training of artificial neural networks using the correlation index. The proposed method is based on a mathematical model of an artificial neural network, represented as an information transmission system. The efficiency of the proposed model is confirmed by its broad application in information transmission systems for analyzing and recovering the useful signal against the background of various interferences: Gaussian, concentrated, pulsed, etc. The analysis of the convergence of training and experimentally obtained sequences based on the correlation index is carried out. The possibility of estimating the convergence of the training and experimentally obtained sequences by the cross-correlation function as a measure of their energy similarity (difference) is confirmed. To evaluate the proposed method, a comparative analysis is carried out with the currently used target indicators. Possible sources of errors of the least squares method and the possibility of the proposed index to overcome them are investigated.

53-66
Abstract

802.11 wireless local area networks (WLANs) can support multiple data rates at physical layer by using adaptive modulation and coding (AMC) scheme. However, this differential data rate capability introduces a serious performance anomaly in WLANs. In a network comprising of several nodes with varying transmission rates, nodes with lower data rate (slow nodes) degrade the throughput of nodes with higher transmission rates (fast nodes). The primary source of this anomaly is the channel access mechanism of WLANs which ensures long term equal channel access probability to all nodes irrespective of their transmission rates. In this work, we investigate the use of adaptable width channelization to minimize the effect of this absurdity in performance. It has been observed that surplus channel-width due to lower transmission rate of slow nodes can be assigned to fast nodes connected to other access points (APs) which can substantially increase the overall throughput of the whole network. We propose a medium access control (MAC) layer independent anomaly prevention (MIAP) algorithm that assigns channel-width to nodes connected with different APs based on their transmission rate. We have modeled the effect of adaptable channelization and provide lower and upper bounds for throughput in various network scenarios. Our empirical results indicate a possible increase in network throughput by more than 20% on employing the proposed MIAP algorithm.

67-81
Abstract

Nowadays artificial intelligence and swarm robotics become wide spread and take their approach in civil tasks. The main purpose of the article is to show the influence of common knowledge about surroundings sharing in the robotic group navigation problem by implementing the data transferring within the group. Methodology provided in article reviews a set of tasks implementation of which improves the results of robotic group navigation. The main questions for the research are the problems of robotics vision, path planning, data storing and data exchange. Article describes the structure of real-time laser technical vision system as the main environment-sensing tool for robots. The vision system uses dynamic triangulation principle. Article provides examples of obtained data, distance-based methods for resolution and speed control. According to the data obtained by provided vision system were decided to use matrix-based approach for robots path planning, it inflows the tasks of surroundings discretization, and trajectory approximation. Two network structure types for data transferring are compared. Authors are proposing a methodology for dynamic network forming based on leader changing system. For the confirmation of theory were developed an application of robotic group modeling. Obtained results show that common knowledge sharing between robots in-group can significantly decrease individual trajectories length.

83-96
Abstract

We propose a new approach to solving important practical problems of complex debugging, joint testing, and analysis of the execution time of software module versions in a heterogeneous distributed computing environment that integrating Grid and cloud computing. These problems arise in the process of supporting the continuous integration of modules of distributed applied software packages. The study focuses on the packages that are used to conduct large-scale computational experiments. The scientific novelty of the proposed approach is to combine the methodology for creating the packages with modern software development practices based on its continuous integration using knowledge about the specifics of the problems being solved. Our contribution is multifold. We expanded the capabilities of continuous integration tools by developing new additional tools for the markup and transformation of data from poorly structured sources and predicting modules execution time. In addition, we developed a technological scheme of the joint applying our developed tools and external systems for continuous integration. Therefore, we provide a more large range of capabilities of continuous integration in relation to the processes of creating and using the packages in comparison with the well-known tools. The fundamental basis of their functioning is a new conceptual model of the packages. This model supports the specification, planning, and execution of software continuous integration processes taking into account the specific subject data and problems being solved. Applying the developed tools in practice leads to a decrease in the number of errors and failures of applied software in the development and use of the packages. In turn, such decrease significantly reduces the time for large-scale computational experiments and increases the efficiency of using resources of the environment. The results of practical experiments on the use of system prototype for continuous integration of applied software show their high efficiency.

97-120
Abstract
The Multi-Bulk Synchronous Parallel (Multi-BSP) model is a recently proposed parallel programming model for multicore machines that extends the classic Bulk Synchronous Parallel model. Multi-BSP aims to be a useful model to design algorithms and estimate their running time. This model heavily relies on the right computation of parameters that characterize the hardware. Of course, the hardware utilization also depends on the specific features of the problems and the algorithms applied to solve them. This article introduces a semi-automatic approach for solving problems applying parallel algorithms using the Multi-BSP model. First, the specific multicore computer to use is characterized by applying an automatic procedure. After that, the hardware architecture discovered in the previous step is considered in order to design a portable parallel algorithm. Finally, a fine tuning of parameters is performed to improve the overall efficiency. We propose a specific benchmark for measuring the parameters that characterize the communication and synchronization costs in a particular hardware. Our approach discovers the hierarchical structure of the multicore architecture and compute both parameters for each level that can share data and make synchronizations between computing units. A second contribution of our research is a proposal for a Multi-BSP engine. It allows designing algorithms by applying a recursive methodology over the hierarchical tree already built by the benchmark, focusing on three atomic functions based in a divide-and-conquer strategy. The validation of the proposed method is reported, by studying an algorithm implemented in a prototype of the Multi-BSP engine, testing different parameter configurations that best fit to each problem and using three different high-performance multicore computers.
121-135
Abstract

Cloud computing is one of the most prominent parallel and distributed computing paradigm. It is used for providing solution to a huge number of scientific and business applications. Large scale scientific applications which are structured as scientific workflows are evaluated through cloud computing. Scientific workflows are data-intensive applications, as a single scientific workflow may consist of hundred thousands of tasks. Task failures, deadline constraints, budget constraints and improper management of tasks can also instigate inconvenience. Therefore, provision of fault-tolerant techniques with data-oriented scheduling is an important approach for execution of scientific workflows in Cloud computing. Accordingly, we have presented enhanced data-oriented scheduling with Dynamic-clustering fault-tolerant technique (EDS-DC) for execution of scientific workflows in Cloud computing. We have presented data-oriented scheduling as a proposed scheduling technique. We have also equipped EDS-DC with Dynamic-clustering fault-tolerant technique. To know the effectiveness of EDS-DC, we compared its results with three well-known enhanced heuristic scheduling policies referred to as: (a) MCT-DC, (b) Max-min-DC, and (c) Min-min-DC. We considered scientific workflow of CyberShake as a case study, because it contains most of the characteristics of scientific workflows such as integration, disintegration, parallelism, and pipelining. The results show that EDS-DC reduced make-span of 10.9% as compared to MCT-DC, 13.7% as compared to Max-min-DC, and 6.4% as compared to Min-min-DC scheduling policies. Similarly, EDS-DC reduced the cost of 4% as compared to MCT-DC, 5.6% as compared to Max-min-DC, and 1.5% as compared to Min-min-DC scheduling policies. These results in respect of make-span and cost are highly significant for EDS-DC as compared with above referred three scheduling policies. The SLA is not violated for EDS-DC in respect of time and cost constraints, while it is violated number of times for MCT-DC, Max-min-DC, and Min-min-DC scheduling techniques.

137-152
Abstract

In this paper, we give an overview of the movement, foraging and feeding ecology as well as sensors technologies that could be embedded into an IoT-based platform for Precision Livestock Farming (PLF). A total of 43 peer-reviewed journal papers indexed by Web of Science were surveyed. Firstly, sensors technologies (e.g., RFID, GPS, or Accelerometer) used by the authors of each paper were identified. Then, papers were classified according to their applicability to ecological studies in the fields of foraging and feeding behavior.

153-169
Abstract

Mobile Ad-Hoc Networks (MANET) require special approaches to the design and selection of data transmission and security algorithms Nodes mobility and dynamic topology give rise to two key problems of MANET – the difficulty of ensuring confidentiality when transmitting data through a network and the complexity of organizing reliable data transfer. This paper proposes a new approach to organizing data transfer through MANET, based on node disjoint multipath routing and modular coding of data. Distributed modular coding allows the use of secret-sharing schemes to ensure confidentiality on the one hand and reliable coding on the other hand. In this paper, a Computationally Secure Secret Sharing Scheme based on the Residue Number System is used, which ensures the confidentiality of data and the reliability of their transmission. Such an approach also allows for balancing the network loading.

171-185
Abstract
The emergence of software artifacts greatly emphasizes the need for protecting intellectual property rights (IPR) hampered by software piracy requiring effective measures for software piracy control. Software birthmarking targets to counter ownership theft of software by identifying similarity of their origins. A novice birthmarking approach has been proposed in this paper that is based on hybrid of text-mining and graph-mining techniques. The code elements of a program and their relations with other elements have been identified through their properties (i.e code constructs) and transformed into Graph Manipulation Language (GML). The software birthmarks generated by exploiting the graph theoretic properties (through clustering coefficient) are used for the classifications of similarity or dissimilarity of two programs. The proposed technique has been evaluated over metrics of credibility, resilience, method theft, modified code detection and self-copy detection for programs asserting the effectiveness of proposed approach against software ownership theft. The comparative analysis of proposed approach with contemporary ones shows better results for having properties and relations of program nodes and for employing dynamic techniques of graph mining without adding any overhead (such as increased program size and processing cost).
187-201
Abstract
An important operation for data processing is a number comparison. In Residue Number System (RNS), it consists of two steps: the computation of the positional characteristic of the number in RNS representation and comparison of its positional characteristics in the positional number system. In this paper, we propose a new efficient method to compute the positional characteristic based on the approximate method. The approximate method as a tool to compare numbers does not require resource-consuming non-modular operations that are replaced by fast bit right shift operations and taking the least significant bits. We prove that in case when the dynamic range of RNS is an odd number, the size of the operands is reduced by the size of the module. If one of the RNS moduli is a power of two, then the size of the operands is less than the dynamic range. It makes our method efficient for hardware implementation of cryptographic primitives and digital signal processing.


Creative Commons License
This work is licensed under a Creative Commons Attribution 4.0 License.


ISSN 2079-8156 (Print)
ISSN 2220-6426 (Online)