• Main Navigation
  • Main Content
  • Sidebar

Russian Digital Libraries Journal

  • Home
  • About
    • About the Journal
    • Aims and Scopes
    • Themes
    • Editor-in-Chief
    • Editorial Team
    • Submissions
    • Open Access Statement
    • Privacy Statement
    • Contact
  • Current
  • Archives
  • Register
  • Login
  • Search
Published since 1998
ISSN 1562-5419
16+
Language
  • Русский
  • English

Search

Advanced filters

Search Results

Science Data Infrastructure for Access to Earth Observation Satellite Data

Е.Б. Кудашев
Abstract: Virtual research centre of digital preservation in Europe provides a natural basis for long-term consolidation of digital preservation research and expertise. Spatial Data Infrastructure will cover technical methods for preservation, access and most importantly re-use of data holdings over the whole lifecycle; legal and economic issues including costs and governance issues as well as digital rights; and outreach within and outside the consortium to help to create a discipline of data curators with appropriate qualifications. Main tasks of Spatial Data Infrastructure SDI development are building global infrastructure for IT and geodata; satellite information harmonization; usage of agreed upon set of standards; clear documentation describing the parts of the system; interoperability between independently created applications and databases; common standards within their interfaces, protocols and data formats; and finally support of a general data policy for data creation, access, and support of satellite information. Fundamental principle of Russian segment of SDI is providing interoperability – the ability of interaction for heterogeneous services and data catalogues within the bounds of a unified informational system. The Russian segment of distributed informational system has been built on the basis of EOLI-XML and SSE technologies.
Keywords: Science Data Infrastructure, e-Science, Earth Observation data, Scientific e-Infrastructure, Open Data Infrastructure, Data management.

Process Approach and Construction of the Database for Non-Core Asset Management in Credit Organizations

Marat Khaidarovich Shakirov
710-753
Abstract:

A method for building end-to-end management accounting in a division of the Bank’s subdevision specializing in working with non-core assets is proposed. Has been proposed the process approach, an algorithm for building a database for the formation of key performance and control indicators.


Has been described the key stages of the department's work, the attribute composition of entities (set) arriving, enriched and transmitted at each stage of the department's work. By modeling the process has been built a role model, access and editing rights for employees. Data sources (reference books) for optimization and unification of the process of filling the database (tuple) are proposed. A method of accessing the database in the Power Query Microsoft Excel add-in is proposed, which allows you to collect data from files of all basic data types, process and refine the received data. In the interactive programming environment Jupyter Notebook, mathematical and financial models for data analysis (logistic regression, decision tree and discounted cash flow method) were built based on data in order to predict costs, the timing of asset exposure and make a decision on the optimal cost of putting property on the Bank's balance sheet and selling price. Based on ready-made libraries (matpotlib, seaborn, plotly), options for data visualization for management are proposed. Using the example of the Bank's division, the author describes the positive effects and opportunities that open up to the management of different levels in solving day-to-day tasks and planning the activities of the division. A technical task was proposed for the development of a showcase for the sale of non-core assets on the Bank's website as an environment for the accumulation of external data for making flexible management decisions.

Keywords: non-core assets, process approach, database, Power Query, data visualization, mathematical and financial methods of data analysis, regression analysis, decision tree, discounted cash flow method.

Application of Synthetic Data to the Problem of Anomaly Detection in the Field of Information Security

Artem Igorevich Gurianov
187-200
Abstract:

Currently, synthetic data is highly relevant in machine learning. Modern synthetic data generation algorithms make it possible to generate data that is very similar in statistical properties to the original data. Synthetic data is used in practice in a wide range of tasks, including those related to data augmentation.


The author of the article proposes a data augmentation method that combines the approaches of increasing the sample size using synthetic data and synthetic anomaly generation. This method has been used to solve an information security problem of anomaly detection in server logs in order to detect attacks.


The model trained for the task shows high results. This demonstrates the effectiveness of using synthetic data to increase sample size and generate anomalies, as well as the ability to use these approaches together with high efficiency.

Keywords: synthetic data, anomaly detection, information security, anomaly generation, data augmentation, machine learning.

База данных RePEc и ее российский партнер Socionet

Т. Крихель, С. Паринов
Abstract: Онлайновая экономическая библиотека RePEc.org занимает второе место в мире (после arXiv.org) по количеству бесплатно предлагаемых научных материалов. У RePEc совсем другая модель организации пополнения базы данных по сравнению с arXiv и совсем другое содержание предоставляемой информации. Предлагаемая статья посвящена описанию этих особенностей.
Модель организации RePEc имеет архитектуру открытого типа. База данных открыта в двух смыслах: 1)для пополнения (организации могут помещать в нее свои материалы), а так же для развития (разработчики могут создаваться различные сервисы для пользователей). Библиотеки традиционного типа – включая многие электронные библиотеки – закрыты в обоих направлениях. В данной статье также обсуждается особенность функционального соединения с RePEc с системой Соционет.
Что касается содержания предоставляемой информации, база данных RePEc нацелена на создание реляционного набора данных о научных ресурсах, а также связанных с ними сведений. Это должно включать данные обо всех авторах, статьях и организаций, имеющих прямое отношение к исследованиям по экономике. Подобный амбициозный проект может быть реализован только в случае, если затраты на сбор данных будут децентрализованы и малы, а также если выгоды от использования данной информации будут достаточно велики.

Digital 3D-Objects Visualization in Forming Virtual Exhibitions

Nikolay Evgenvich Kalenov, Sergey Alexandrovich Kirillov, Irina Nikolaevna Sobolevskaya, Aleksandr Nikolaevich Sotnikov
418-432
Abstract: The paper is presents approaches to solving the problem of creating realistic interactive 3D web-collections of museum exhibits. The presentation of 3D-models of objects based on oriented polygonal structures is considered. The method of creating a virtual collection of 3D-models using interactive animation technology is described. It is also shown how a full-fledged 3D-model is constructed on the basis of individual exposure frames using photogrammetry methods. The paper assesses the computational complexity of constructing realistic 3D-models. For the creation of 3D-models in order to provide them to a wide range of users via the Internet, the so-called interactive animation technology is used. The paper presents the differences between the representations of full-fledged 3D-models and 3D-models presented in the form of interactive multiplication. The technology of creating 3D-models of objects from the funds of the State Biological Museum named K.A Timiryazev and the formation on their basis of the digital library “Scientific Heritage of Russia” of a virtual exhibition dedicated to the scientific activities of M.M. Gerasimov and his anthropological reconstructions, and vividly demonstrating the possibility of integrating information resources by means of an electronic library. The format of virtual exhibitions allows you to combine the resources of partners to provide a wide range of users with collections stored in museum, archival and library collections.
Keywords: photogrammetry, 3D-modeling, interactive animation, web-design, polygonal modeling.

Formalization of Processes for Forming User Collections in the Digital Space of Scientific Knowledge

Nikolay Evgenvich Kalenov, Irina Nikolaevna Sobolevskaya, Aleksandr Nikolaevich Sotnikov
433-450
Abstract: The task of forming a digital space of scientific knowledge (DSSK) is analyzed in the paper. The difference of this concept from the general concept of the information space is considered. DSSK is presented as a set containing objects verified by the world scientific community. The form of a structured representation of the digital knowledge space is a semantic network, the basic organization principle of which is based on the classification system of objects and the subsequent construction of their hierarchy, in particular, according to the principle of inheritance. The classification of the objects that make up the content of the DSSK is introduced. A model of the central data collection system is proposed as a collection of disjoint sets containing digital images of real objects and their characteristics, which ensure the selection and visualization of objects in accordance with multi-aspect user requests. The concept of a user collection is defined, and a hierarchical classification of types of user collections is proposed. The use of the concepts of set theory in the construction of DSSK allows you to break down information into levels of detail and formalize the algorithms for processing user queries, which is illustrated by specific examples.
Keywords: recursive link, knowledge cyberdomain, digital library, detail levels, data entries hierarchy.

Experience of TLS 1.3 Clients Verification

Aleksey Vyacheslavovich Nikeshin, Victor Zinovievich Shnitman
104-121
Abstract:

This paper presents the experience of verifying client implementations of the TLS cryptographic protocol version 1.3. TLS is a widely used cryptographic protocol today, designed to create secure data transmission channels. The protocol provides the necessary functionality for its tasks: confidentiality of transmitted data, data integrity, and authentication of the parties. In the new version 1.3 of the TLS architecture was significantly redesigned, eliminating a number of shortcomings of previous versions that were identified both during the development of implementations and during their operation. We used a new test suite for verifying client implementations of the TLS 1.3 for compliance with Internet specifications, developed on the basis of the RFC8446, using UniTESK technology and mutation testing methods. To test implementations for compliance with formal specifications, UniTESK technology is used, which provides testing automation tools based on the use of finite state machines. The states of the system under test define the states of the state machine, and the test effects are the transitions of this machine. When performing a transition, the specified impact is passed to the implementation under test, after which the implementation's reactions are recorded and a verdict is automatically made on the compliance of the observed behavior with the specification. Mutational testing methods are used to detect non-standard behavior of the system under test by transmitting incorrect data. Some changes are made to the protocol exchange flow created in accordance with the specification: either the values of the message fields formed on the basis of the developed protocol model are changed, or the order of messages in the exchange flow is changed. The protocol model allows one to make changes to the data flow at any stage of the network exchange, which allows the test scenario to pass through all the significant states of the protocol and in each such state to test the implementation in accordance with the specified program. The presented approach has proven effective in several of our projects when testing network protocols, providing detection of various deviations from the specification and other errors. The current work is part of the TLS 1.3 protocol verification project and covers TLS client implementations.

Keywords: security, TLS, TLSv1.3, protocols, testing, verification, evaluate robustness, Internet, standards, formal specifications.

Experience of Implementation of the Protocol TLS 1.3 Verification

Aleksey Vyacheslavovich Nikeshin, Victor Zinovievich Shnitman
902-922
Abstract:

This paper presents the experience of verifying server implementations of the TLS cryptographic protocol version 1.3. TLS is a widely used cryptographic protocol designed to create secure data transmission channels and provides the necessary functionality for this: confidentiality of the transmitted data, data integrity, and authentication of the parties. The new version 1.3 of the TLS protocol was introduced in August 2018 and has a number of significant differences compared to the previous version 1.2. A number of TLS developers have already included support for the latest version in their implementations. These circumstances make it relevant to do research in the field of verification and security of the new TLS protocol implementations. We used a new test suite for verifying implementations of the TLS 1.3 for compliance with Internet specifications, developed on the basis of the RFC8446, using UniTESK technology and mutation testing methods. The current work is part of the TLS 1.3 protocol verification project and covers some of the additional functionality and optional protocol extensions. To test implementations for compliance with formal specifications, UniTESK technology is used, which provides testing automation tools based on the use of finite state machines. The states of the system under test define the states of the state machine, and the test effects are the transitions of this machine. When performing a transition, the specified impact is passed to the implementation under test, after which the implementation's reactions are recorded and a verdict is automatically made on the compliance of the observed behavior with the specification. Mutational testing methods are used to detect non-standard behavior of the system under test by transmitting incorrect data. Some changes are made to the protocol exchange flow created in accordance with the specification: either the values of the message fields formed on the basis of the developed protocol model are changed, or the order of messages in the exchange flow is changed. The protocol model allows one to make changes to the data flow at any stage of the network exchange, which allows the test scenario to pass through all the significant states of the protocol and in each such state to test the implementation in accordance with the specified program. So far, several implementations have been found to deviate from the specification. The presented approach has proven effective in several of our projects when testing network protocols, providing detection of various deviations from the specification and other errors.

Keywords: security, TSL, TSLv1.3, protocols, testing, verification, evaluate robustness, Internet, standards, formal specifications.

Building Subject Domain Ontology on the Base of a Logical Data Mod

Alexander M. Gusenkov, Naille R. Bukharaev, Evgeny V. Biryaltsev
390-417
Abstract: The technology of automated construction of the subject domain ontology, based on information extracted from the comments of the TATNEFT oil company relational databases, is considered. The technology is based on building a converter (compiler) translating the logical data model of Epicenter Petrotechnical Open Software Corporation (POSC), presented in the form of ER diagrams and a set of the EXPRESS object-oriented language descriptions, into the OWL ontology description language, recommended by the W3C consortium. The basic syntactic and semantic aspects of the transformation are described.
Keywords: subject domain ontology, relational databases, POSC, OWL.

Stability Studies of a Coupled Model to Perturbation of Initial Data

Konstantin Pavlovich Belyaev, Gury Mikhaylovich Mikhaylov, Alexey Nikolaevich Salnikov, Natalia Pavlovna Tuchkova
615-633
Abstract: The stability problem is considered in terms of the classical Lyapunov definition. For this, a set of initial conditions is set, consisting of their preliminary calculations, and the spread of the trajectories obtained as a result of numerical simulation is analyzed. This procedure is implemented as a series of ensemble experiments with a joint MPI-ESM model of the Institute of Meteorology M. Planck (Germany). For numerical modeling, a series of different initial values of the characteristic fields was specified and the model was integrated, starting from each of these fields for different time periods. Extreme ocean level characteristics over a period of 30 years were studied. The statistical distribution was built, the parameters of this distribution were estimated, and the statistical forecast for 5 years in advance was studied. It is shown that the statistical forecast of the level corresponds to the calculated forecast obtained by the model. The localization of extreme level values was studied and an analysis of these results was carried out. Numerical calculations were performed on the Lomonosov-2 supercomputer of Lomonosov Moscow State University.
Keywords: non-linear circulation models, Ensemble numerical experiments, analysis of stability of the model trajectories.

Creating a comparison method for relational tables

Азат Шавкатович Якупов, Даниил Андреевич Клинов
173-183
Abstract: The article is devoted to creating a quick method of comparing a huge amount of data tables in relational database management systems. Creating an effective method for comparing relational systems is really relevant today. The study of existing solutions was conducted. The algorithm in this article was created using the probabilistic data structure «Countable Bloom filter» and the Monte Carlo Method. The proposed solution is unique in its direction, as it uses the least amount of temporary resources. A probabilistic model of the created algorithm is constructed, this algorithm can be used for parallelization.
Keywords: multiset, comparison of relational tables, heterogeneous system, Countable Bloom filter, Monte Carlo method, replication, Oracle, PostgreSQL, Probabilistic data structure.

Computed knowledge base for description of information resources of molecular spectroscopy. 2. Data model of quantitative spectroscopy.

Н.А. Лаврентьев, А.И. Привезенцев, А.З. Фазлиев
Abstract: The data model is needed for implementation of the information system in domain of quantitative spectroscopy is represented. The extended makeup language (XML) was used for description of the data model. The entities and relations of two domains “Substance” and “Quantitative molecular spectroscopy” are in the focus of our interest. Under the specifications of data semantics only the datatype properties were used.

Machine learning methods for determining the relationship between academic success and data of social network profile

Ильяс Раисович Ихсанов, Ирина Сергеевна Шахова
95-118
Abstract: The paper is aimed to propose the machine learning model for determining the relationship between data of social network profile and academic success of students and predicting the success using the data.
Keywords: machine learning, social networks, psychometrics, academic success, education, abiturient.

Modeling an Adaptive Interface using Semantic Ontology Relations

Olga Muratovna Ataeva, Vladimir Alekseevich Serebriakov, Natalia Pavlovna Tuchkova
2-17
Abstract:

The work is devoted to the problem of customizing the user interfaces of an information system that integrates data. An adaptive interface serves as one of the means of organizing the presentation of subject domain data. The issue of using the semantic relations of ontology to select data corresponding to the objectives of the study is investigated. A model of an adaptive interface is considered, which allows the most accurate reflection of the needs of a researcher within a particular subject domain. It is shown how the adaptive interface is formed by means of the semantic library model.

Keywords: ontology, adaptive interface, subject domain, data model.

Development of Methods and Software Tools for the Formation of a Digital Portrait of Students

Marat Albertovich Solntsev, Mikhail Mikhailovich Abramskiy
697-717
Abstract:

This paper considers the questions about the possibility of using data about the students presented in electronic form to build their digital portraits.  A set of characteristics necessary for its construction is proposed, a data model is designated.


Implemented tools for collecting data about students from social networks and other Internet resources. Algorithms for constructing a digital portrait are proposed. The application of machine learning algorithms for these tasks is illustrated. Examples of the use of digital portraits in education are given.

Keywords: social networks, data retrieval, personal portrait of user, education.

Use of heterogeneous geospatial data insolving the problems of sustainable development of territories

Е.Б. Кудашев, С.Ю. Марков, М.А. Попов
Abstract: We refer to the problems of integration of heterogeneous geospatial data to solve the problems of sustainable development of territories. The conceptual approaches to build up the e-Science Data infrastructure for permanent access to heterogeneous spatial data are proposed. The main challenges of integrating heterogeneous data are analyzed. A model of integration of geospatial data based on the architecture SOA are developed.
Keywords: geospatial data, heterogeneity, metadata.

Progress in Dvm-System

Valery Fedorovich Aleksahin, Vladimir Aleksandrovich Bakhtin, Olga Fedorovna Zhukova, Dmitry Aleksandrovich Zakharov, Victor Alekseevich Krukov, Nataliya Victorovna Podderyugina, Olga Antonievna Savitskaya
247-270
Abstract: DVM-system is designed for the development of parallel programs of scientific and technical calculations in the C-DVMH and Fortran-DVMH languages. These languages use a single DVMH-model of parallel programming model and are an extension of the standard C and Fortran languages with parallelism specifications in the form of compiler directives. The DVMH model makes it possible to create efficient parallel programs for heterogeneous computing clusters, in the nodes of which accelerators, graphic processors or Intel Xeon Phi coprocessors can be used as computing devices along with universal multi-core processors. The article presents new features of DVM-system that have been developed recently.
Keywords: automation of development of parallel programs, DVM-system, accelerator, GPU, Fortran, С, irregular grid, unstructured grid.

Steel Defects Analysis Using CNN (Convolutional Neural Networks)

Rodion Dmitrievich Gaskarov, Alexey Mikhailovich Biryukov, Alexey Fedorovich Nikonov, Daniil Vladislavovich Agniashvili, Danil Aydarovich Khayrislamov
1155-1171
Abstract:

Steel is one of the most important bulk materials these days. It is used almost everywhere - from medicine to industry. Detecting this material's defects is one of the most challenging problems for industries worldwide. This process is also manual and time-consuming. Through this study we tried to automate this process. A convolutional neural network model UNet was used for this task for more accurate segmentation with less training image data set for our model. The essence of this NN (neural network) is in step-by-step convolution of every image (encoding) and then stretching them to initial resolution, consequently getting a mask of an image with various classes on it. The foremost modification is changing an input image's size to 128x800 px resolution (original images in dataset are 256x1600 px) because of GPU memory size's limitation. Secondly, we used ResNet34 CNN (convolutional neural network) as encoder, which was pre-trained on ImageNet1000 dataset with modified output layer - it shows 4 layers instead of 34. After running tests of this model, we obtained 92.7% accuracy using images of hot-rolled steel sheets.

Keywords: CNN, neural networks, steel, machine learning, AI, Unet, ResNet, defects detection, segmentation, classification.

Debugging Parallel Programs in DVM-System

Vladimir Aleksandrovich Bakhtin, Dmitry Aleksandrovich Zakharov, Aleksandr Aleksandrovich Ermichev, Victor Alekseevich Krukov
866-886
Abstract: DVM-system is designed for the development of parallel programs of scientific and technical calculations in the C-DVMH and Fortran-DVMH languages. These languages use a single DVMH-model of parallel programming model and are an extension of the standard C and Fortran languages with parallelism specifications in the form of compiler directives. The DVMH model makes it possible to create efficient parallel programs for heterogeneous computing clusters, in the nodes of which accelerators, graphic processors or Intel Xeon Phi coprocessors can be used as computing devices along with universal multi-core processors. The article describes the method of debugging parallel programs in DVM-system, as well as new features of DVM-debugger.
Keywords: automation of development of parallel programs, automation of debugging of parallel programs, dynamic control, relative debugger, DVM-system, accelerator, GPU, Fortran, С.

Heterogeneous spatial data in gas and oil prospecting

М.А. Попов, С.А. Станкевич, С.Ю. Марков, А.В. Зайцев, Е.Б. Кудашев
Abstract: The paper is devoted to the problems of mutual use of heterogeneous geospatial information for gas and oil searching. The model of aerospace information and in-situ spectral data integration is assumed, data integration approach using FME software is offered, the results of experimental researches are presented.
Keywords: гетерогенная геопространственная информация, аэрокосмическая информация, наземные спектральные измерения, программный продукт FME.

New Possibilities of the Fourier Transformation: How to Describe an Arbitrary Frequency-Phase Modulated Signal?

Raoul Rashidovich Nigmatullin, Alexander Alekseevich Litvinov, Sergey Igorevich Osokin
378-397
Abstract:

In this paper, the authors found a transformation that is valid for any arbitrary signal. This transformation is strictly periodical and therefore it allows to apply the ordinary F-transformation for the fitting of the transformed signal. The most interesting application (in accordance with the author's opinion) is the fitting of the frequency-phase modulated signals that actually located inside the found transformation. This new transformation will be useful for application of the responses of different complex systems when an ordinary model is absent.


As an available data we consider meteo-data corresponding to measurements of methane concentration (CH4) in atmosphere during 4 weeks of its observation. For us it is important to consider the integral (cumulative) data and find their amplitude-frequency response (AFR). If one considers each column as frequency-phase modulated signal, then AFR can be evaluated with the help of F-transformation that has the period equals 2p that is valid for any analyzed random signal. This "universal" F-transformation allows to fit a wide set of random signals and compare them with each other in terms of their AFRs. Concluding the abstract one can say that these new possibilities of the traditional F-analysis will serve as a common tool in the armory of the methods used by researchers in data processing area.

Keywords: Fourier transform, random signal, frequency-phase modulated signal, amplitude-frequency response, complex systems, meteorological data, eddy covariance.

Unified Representation of the Common Digital Space of Scientific Knowledge Ontology

Nikolay Evgenievich Kalenov, Alexander Nikolaevch Sotnikov
80-103
Abstract:

The Common Digital Space of Scientific Knowledge (CDSSK) is a digital information environment aggregating heterogeneous information related to various aspects of scientific knowledge. One of the important functions of the CDSSK is to provide information for solving artificial intelligence problems, which makes it necessary to support data in a structure that complies with the rules of the semantic WEB. The features of the CDSSK are, on the one hand, the polythematics and heterogeneity of content elements, on the other hand, the high dynamics of the emergence of new types of objects and connections between them, which is due to the specifics of the development of science. At the same time, it should be possible to navigate through heterogeneous space resources using semantic links between them. The possibilities of the CDSSK are largely determined by the structure of the ontology of space, the model of which is proposed in this paper. Within the framework of the model, the hierarchical structuring of the CDSSK ontology is carried out; such elements as "subspace", "class of objects", "object", "attributes of an object", three types of pairwise relations of objects or attributes (universal, quasi-universal and specific) are distinguished and defined. The structure of each elements type is determined by a "reference book" of a unified type; specific values of attributes and relationships are contained in dictionaries of a unified structure. A class of "Formats" objects describing the rules for the formation of attributes and values of relationships is allocated. The formalization of CDSSK reference books and dictionaries representations is proposed. The proposed model allows you to simply add new types of objects, of their pairwise relationships and attributes to the space, as needed.

Keywords: digital space of scientific knowledge, ontologies, structuring, related data, data attributes, semantic WEB.

The Using of DVM-System for Developing of a Program for Calculations of the Problem of Radiation Magnetic Gas Dynamics and Research of Plasma Dynamics in the QSPA Channel

Vladimir Aleksandrovich Bakhtin, Dmitry Aleksandrovich Zakharov, Andrey Nikolaevich Kozlov, Veniamin Sergeevich Konovalov
594-614
Abstract: DVM-system is designed for the development of parallel programs of scientific and technical calculations in the C-DVMH and Fortran-DVMH languages. These languages use a single DVMH-model of parallel programming model and are an extension of the standard C and Fortran languages with parallelism specifications in the form of compiler directives. The DVMH model makes it possible to create efficient parallel programs for heterogeneous computing clusters, in the nodes of which accelerators, graphic processors or Intel Xeon Phi coprocessors can be used as computing devices along with universal multi-core processors. The article describes the experience of the successful using of DVM-system to develop a parallel software code for calculating the problem of radiation magnetic gas dynamics and for research of plasma dynamics in the QSPA channel.
Keywords: automation of development of parallel programs, DVM-system, plasma accelerator, radiation magnetic gas dynamics.

Evolution of Visualization Methods for Research Publication Collections

Alexander Ivanovich Legalov, Igor Alexandrovich Legalov, Ivan Vasilievich Matkovsky
788-807
Abstract: It is proposed to add a static system of types to the dataflow functional model of parallel computing and the dataflow functional parallel programming language developed on its basis. The use of static typing increases the possibility of transforming dataflow functional parallel programs into programs running on modern parallel computing systems. Language constructions are proposed. Their syntax and semantics are described. It is noted that the need to use the single assignment principle in the formation of data storages of a particular type. The features of instrumental support of the proposed approach are considered.
Keywords: visualization of document collections, text analysis, text and metadata visualization algorithms, LDA, NMF, word2vec.

Modern approaches to construction of metadata and search services for applications in the field of Earth Sciences

М.А. Попов, Е.Б. Кудашев, С.Ю. Марков, С.А. Станкевич
Abstract: The main aim of this paper is to improve the effectiveness of information retrieval and working with metadata. Approaches to metadata standards are considered, as well as the implementation methods of search procedures in modern web environment. Various aspects for the organization of services of management by spatial metadata are considered. Model for direct usage of heterogeneous spatial information are presented.
Keywords: efficient search, metadata, spatial and non-spatial information.
1 - 25 of 28 items 1 2 > >> 
Information
  • For Readers
  • For Authors
  • For Librarians
Make a Submission
Current Issue
  • Atom logo
  • RSS2 logo
  • RSS1 logo

Russian Digital Libraries Journal

ISSN 1562-5419

Information

  • About the Journal
  • Aims and Scopes
  • Themes
  • Author Guidelines
  • Submissions
  • Privacy Statement
  • Contact
  • eLIBRARY.RU
  • dblp computer science bibliography

Send a manuscript

Authors need to register with the journal prior to submitting or, if already registered, can simply log in and begin the five-step process.

Make a Submission
About this Publishing System

© 2015-2025 Kazan Federal University; Institute of the Information Society