Участник:Strijov/Drafts

Материал из MachineLearning.

Перейти к: навигация, поиск

Шаблон:Main article

Содержание

2023

Problem list

Roles

Student's workload depends on the group and can vary from 54 hours and up. A consultant is expected to make one-hour meetings weekly and promptly to student's questions. So it takes 12 to 16 hours. An expert is expected to state the problem and evaluate the delivery. It takes one-hour maximum. And we guess researchers are ready to discuss their favorite problems. It creates a negative workload: for a problem the expert solves as a daily routine, some delivery appears after several months of synchronized work. The quality of the stated problem matters.

Task 112

  • Name: Modeling an FMRI reading from a video of a shown person
  • Problem description: It is required to build a dependence model of the readings of FMRI sensors and the video sequence that a person is viewing at this moment.
  • Data: The sample for approximation is presented in the work of J. Berezutskay, in which there are various types of parallel signals.
  • Literature: Berezutskaya J., et al Open multimodal iEEG-fMRI dataset from naturalistic stimulation with a short audiovisual film // Sci Data 9, 91, 2022.
  • Predecessor code:
  • Base algorithm: Running code based on transformer models.
  • Novelty: Analysis of the relationship between sensor readings and human perceptions of the external world. It is required to test the hypothesis of the relationship between the data, as well as to propose a method for approximating FMRI readings based on the video sequence being viewed.
  • Authors: Expert Grabovoi Andrey

Task 113

  • Name: Modeling of the FMRI indication on the sound range that a person hears
  • Problem description: It is required to build a model of the dependence of the readings of the FMRI sensors and the sound accompaniment that a person is listening to at this moment.
  • Data: The sample for approximation is presented in the work of J. Berezutskay, in which there are various types of parallel signals.
  • Literature: Berezutskaya J., et al Open multimodal iEEG-fMRI dataset from naturalistic stimulation with a short audiovisual film // Sci Data 9, 91, 2022.
  • Predecessor code:
  • Base algorithm: Running code based on transformer models.
  • Novelty: Analysis of the relationship between sensor readings and human perceptions of the external world. It is required to test the hypothesis of the relationship between the data, as well as to propose a method for approximating the FMRI readings from the listening sound series.
  • Authors: Expert Grabovoi Andrey

Task 114

  • Name: Simulating the Dynamics of Physical Systems with Physics-Informed Neural Networks
  • Task: The Task of choosing the optimal model for predicting the dynamics of a physical system is being solved. Under the dynamics of the system is understood the change in time of the parameters of the system. Neural networks do not have a priori knowledge about the system being modeled, which does not allow obtaining optimal parameters that take into account physical laws. The Lagrangian neural network takes into account the law of conservation of energy when modeling dynamics. In this paper, a Noetherian Agrangian neural network is proposed that takes into account the laws of conservation of momentum and angular momentum in addition to the law of conservation of energy. It is shown that for this problem the Noetherian Lagrangian neural network is optimal among the fully connected neural network model, the neural network with long-term short-term memory and the Lagrangian neural network. The simulation comparison was carried out on artificially generated data for the double pendulum system, which is the simplest chaotic system. The results of the experiments confirm the hypothesis that the introduction of a priori knowledge about the physics of the system improves the quality of the model.
  • Problem description:Generate a set of convolutions from the available data and choose the best one using order and dimensionality reduction techniques.
  • Data: Biomedical accelerometer and gyroscope data, ocean currents, dune movement, air currents.
  • Literature: The base work contains references.
  • Base algorithm: Neural network, Lagrangian neural networks.
  • Solution: Nesterov neural network.
  • Novelty: The proposed network takes into account the symmetry.
  • Authors: Experts - Severilov, Strizhov V.V., consultant - Panchenko

Task 115

  • Name: Knowledge distillation in deep networks and alignment of model structures
  • Problem description: It is required to build a network of the simplest structure, a student model, using a high quality teacher model. Show how the student's accuracy and stability change. The result of the experiment is a graph complexity-accuracy-stability, where each model is accurate.
  • Data: CIFAR-10. It is assumed that the teacher has a structure open for analysis with a large number of layers.
  • Literature: Hinton's original work on distillation, work by Andrei Grabovoi, work by Maria Gorpinich
  • Base algorithm: for comparison

Training (models with a given structure of controlled complexity) without distillation. Teaching (ditto) with Hinton distillation. Layered learning. Neuronal transfer learning.

  • Solution: As in paragraph 2, only in layers. Building the path of least cost over neurons. We consider the covariance matrices of each neuron of each layer for the teacher and for the student. We propose an error function that includes the cost of the least cost path. We propose a way to construct the path of least cost. The main idea: the transfer goes through pairs of neurons and the most similar distributions (expectation and covariance matrix) from teacher to student.
  • Novelty: The proposed transfer significantly reduces complexity without loss of accuracy and solves the problem of interchangeability of neurons by identifying them.
  • Authors: Experts Bakhteev Oleg, Strizhov V.V., consultant - Gorpinich Maria

Task 116

  • Name: Neural differential equations for modeling physical activity - selection and generation of mathematical models
  • Task: Task of choosing the optimal mat. models as a task of genetic optimization. The optimality criterion is defined in terms of the accuracy, complexity, and stability of the model. The sampling procedure itself consists of two steps: generating a new structure and rejecting this structure if it does not satisfy the optimality criterion. Required on 'pendulum' type data - accelerometer, myogram, pulse wave - to choose the optimal model.
  • Data: WISDM, own collection of biomedical data
  • Literature: Neural CDE
  • Base algorithm: Neuro ODE/CDE on a two-layer neural network.
  • Solution: A number of experiments have already been performed, where sampling is performed by a genetic algorithm. Acceptable results have been obtained. It is proposed to analyze and improve them.
  • Solution: Algorithm for generating mathematical models in the form of ordinary differential equations. Comparison of models and solvers on biomedical data.
  • Authors: Expert – Strizhov V.V., consultant - Eduard Vladimirov

Task 117

  • Name: Search for dependencies of biomechanical systems (do people dance in pairs or independently?) and (Method of Convergence Cross-Mpping, Takens theorem)
  • Task: When forecasting complex time series that depend on exogenous factors and have multiple periodicity, it is required to solve the problem of identifying related pairs of series. It is assumed that the addition of these series to the model improves the quality of the forecast. In this paper, to detect relationships between time series, it is proposed to use the convergent cross-mapping method. With this approach, two time series are connected if their trajectory subspaces exist, the projections onto which are connected. In turn, the projections of series onto trajectory subspaces are related if the neighborhood of the phase trajectory of one series is mapped to the neighborhood of the phase trajectory of another series. The Task of finding trajectory subspaces that reveal the connection of series is set.
  • Literature:

Everything Sugihara wrote in Science and Nature (ask the collection) Usmanova K.R., Strizhov V.V. V.V. Detection of dependencies in time series in the problems of building predictive models // Systems and means of informatics, 2019, 29(2) Neural CDE

  • Data:

Accelerometer, gyroscope, and other data describing dynamic systems

  • Solution:

Basic in Karina's work Ours is to build the Neural ODE for both signals and decide if both models belong to the same dynamic system.

  • Authors: Expert Strizhov V.V., consultant - Vladimirov, Samokhina

Task 118

  • Name: Continuous time when building a BCI neural interface
  • Task: In signal decoding tasks, data is represented as multidimensional time series. When solving problems, a discrete representation of time is used. However, recent work on neural ordinary differential equations illustrates the ability to work with the hidden state of recurrent neural networks, as with solutions to differential equations. This allows us to consider time series as continuous in time.
  • Data: For classification: dataset P300, which was used to write an article with Alina, DEAP dataset dataset similar to it in the format of records, find a modern dataset, ask U.Grenoble-Alpes
  • Literature:: Neural CDE
  • Base algorithm: Alina Samokhina's algorithm
  • Solution: Using NeurODE variations to approximate the original signal. Comparative analysis of existing approaches to the application of differential equations for EEG classification. (Encoder-tensor decomposition, NeuroCDE decoder)
  • Novelty: suggests a way to construct a continuous signal representation. Working with the functional space of the signal, not its discrete representation. Using the parameters of the resulting function as a feature space of the resulting model.
  • Authors: Expert Strizhov V.V. (was Task 109), consultant - Tikhonov

Task 119

  • Name: Analysis of the dynamics of multiple learning
  • Problem description: Consider a supervised multiple learning problem in which the training set is not fixed, but is updated depending on the predictions of the trained model on the test set. For the process of multiple training, prediction and updating of the sample, we build a mathematical model and study the properties of this process based on the constructed model. Let f(x) be a feature distribution density function, G be an algorithm for training the model, generating predictions on the test set and mixing predictions into the training set, as a result of which the feature distribution changes. Let the space of non-negative smooth functions F(x) be given, whose integral on R^n is equal to one. f_{t+1}(x) = G(f_{t})(x), where G(f) is the evolution operator on the space of these functions F and the initial function f_0(x) is known. In general, G can be an arbitrary operator, not necessarily smooth and/or continuous.

Question 0. Find conditions on the operator G under which the image of G lies in the same class of distribution density functions F. In particular, should G be bounded, the operator norm ||G|| <= 1, so that the image of G(f) \in F is also a distribution density function for any f from F? Does there exist a unit in the space F with respect to the operator G, and what will be the identity function f in such F? Question 1. Under what conditions will there be a t_0 on G such that for all t > t_0 the tail of the sequence {f} will be bounded? Question 2. Under what conditions will the operator G have a fixed point? Data In a computational experiment, it is proposed to check the significance of the restriction / the significance of the conditions under which the answer to questions 0-2 is obtained. For example, for a problem of linear regression and/or regression with a multilevel fully connected neural network with different proportions of predictions mixed into the training set on synthetic data sets.

  • Authors: Expert - Khritankov A.S., Expert - Afanasiev A.P.
  • Literature:

Khritankov A., Hidden Feedback Loops in Machine Learning Systems: A Simulation Model and Preliminary Results, https://doi.org/10.1007/978-3-030-65854-0_5 Khritankov A.. Pilkevich A. Existence Conditions for Hidden Feedback Loops in Online Recommender Systems, https://doi.org/10.1007/978-3-030-91560-5_19 Katok A.B., Hasselblat B. Introduction to the modern theory of dynamical systems.1999. 768 p. ISBN 5-88688-042-9. Nemytsky V. V., Stepanov V. V. Qualitative theory of differential equations, published in 1974

Task 120

  • Name: Differentiated algorithm for searching ensembles of deep learning models with diversity control
  • Task: The Task of selecting an ensemble of models is considered. It is required to propose a method for controlling the diversity of basic models at the stage of application.
  • Data: Fashion-MNIST, CIFAR-10, CIFAR-100 datasets
  • Literature::

Neural Architecture Search with Structure Complexity Control Neural Ensemble Search via Bayesian Sampling DARTS: Differentiable Architecture Search

  • Base algorithm: It is proposed to use DARTS [3] as the basic algorithm.
  • Solution: To control the diversity of basic models, it is proposed to use a hypernet [1], which shifts the structural parameters in terms of the Jensen-Shannon divergence. At the application stage, base architectures are sampled with a given offset to build an ensemble.
  • Novelty: The proposed method allows building ensembles with any number of base models without additional computational costs relative to the base algorithm.
  • Authors: K.D. Yakovlev, Bakhteev Oleg

Task 121

  • Task: building predictive analytics for air pollution sensors.
  • Problem description: Data available for air quality monitoring stations in Moscow and the Moscow region (time series). The task is to check the achievable predictive ability to predict the time series of station readings by their history + when connecting additional features (take into account the stations in aggregate, taking into account their location, time of day and weekend / working day, history and weather forecast (wind))
  • Data: Real data and simulations for Moscow and Moscow Region
  • Authors: Artem Mikhailov, Vladimir Vanovsky

Task 122

  • Task: Reducing the dimension of space in a generative modeling problem using reversible models.
  • Problem description: An example of a generative modeling problem is image generation. Some kinds of new models, such as normalization flows or diffusion models, define reversible transformations. But at the same time they work in a space of very high dimensions. It is proposed to combine 2 approaches: dimensionality reduction and generative modeling.
  • Data: Any image dataset (MNIST/CIFAR10).
  • Novelty: By reducing the dimension, you can achieve a significant acceleration of generative models, which will reduce the complexity of such models.
  • Author: Roman Isachenko

Task 123

  • Task: Analysis of distribution bias in contrast distribution problem.
  • Problem description: There is the same problem as Representation learning. One of the most popular approaches to solving this problem is contrastive learning. At the same time, in the data we learn from, there are often markup errors: false positive/false negative. It is proposed to analyze various ways to eliminate these biases caused by errors. And also to explore the properties of the proposed models.
  • Data: Any image dataset (MNIST/CIFAR10).
  • Novelty: Current models are very error sensitive. If you manage to take into account the bias in the distributions, many methods of ranking products will greatly increase in quality.
  • Author: Roman Isachenko

Task 124

  • Task: Speed up sampling from diffusion models using adversarial networks
  • Problem description: The most popular generative model today is the diffusion model. Its main disadvantage is the speed of sampling. To sample 1 picture, you need to run 1 neural network 100-1000 times. There are ways to speed up this process. One such way is to use adversarial networks. It is proposed to develop this method and explore various ways to set the functional for sampling
  • Data: Any image dataset (MNIST/CIFAR10).
  • Novelty: By speeding up diffusion models, they will become even more popular and easier to use.
  • Author: Roman Isachenko

Task 125

  • Name: Influence of the lockdown on the dynamics of the spread of the epidemic
  • Task: The introduction of a lockdown is considered an effective measure to combat the epidemic. However, contrary to intuition, it turned out that under certain conditions, a lockdown can lead to an increase in the epidemic. This effect is absent for the classical models of the spread of the epidemic “on average”, but was revealed when modeling the epidemic on the contact graph. The task is to find formulaic and quantitative relationships between the parameters under which the lockdown can lead to an increase in the epidemic. It is necessary both to identify such relationships in the SEIRS/SEIR/SIS/etc models based on the SEIRS+ epidemiological distribution framework (and its modifications), and to theoretically substantiate the relationships obtained from specific implementations of the epidemia.
  • Data: Task involves working with model and synthetic data: there are ready-made data, and it is also possible to generate new ones in the process of solving the problem. This Task belongs to unsupervised learning, since the implementation of the epidemic on the contact graph has a high proportion of random events, and therefore requires analysis on average over many synthetically generated implementations of the epidemic
  • Literature: T. Harko, Francisco S. N. Lobo, and M. Mak. "Exact analytical solutions of the Susceptible-Infected-Recovered (SIR) epidemic model and of the SIR model with equal death and birth rates"
  • Authors: A.Yu. Bishuk, A.V. Zuhba

Task 126

  • Name: Machine generation style change detection
  • Problem description:It is required to propose a detection method
  • Data: The sample for approximation is presented in the work of J. Berezutskay, in which there are various types of parallel signals.
  • Literature::

G. Gritsay, A. Grabovoy, Y. Chekhovich. Automatic Detection of Machine Generated Texts: Need More Tokens // Ivannikov Memorial Workshop (IVMEM), 2022. M. Kuznetsov, A. Motrenko, R. Kuznetsova, V. Strijov. Methods for intrinsic plagiarism detection and author diarization // Working Notes of CLEF, 2016, 1609 : 912-919. RuATD competition.

  • Base algorithm:

Using the results of the RuATD competition as base models for classifying proposals. Use the method from Kuznetsov et all.

  • Novelty: Suggest a method for detecting machine-generated fragments in the text using methods for changing the writing style.
  • Authors: Expert Grabovoi Andrey

Task 128

  • Name: Build a deep learning model based on task data
  • Task: is considered Task optimization of the deep learning model for the new dataset. It is required to propose a model optimization method that allows generating new models for a new dataset with low computational costs.
  • Data: CIFAR10, CIFAR100
  • Literature:: variational inference for neural networks, hypernets, similar work tailored to change the model depending on a predetermined complexity
  • Base algorithm: Retrain the model directly. * Solution: The proposed method is to represent a deep learning model as a hypernet (a network that generates the parameters of another network) using a Bayesian approach. Probabilistic assumptions about the parameters of deep learning models are introduced, and a variational lower estimate of the Bayesian validity of the model is maximized. The variation estimate is considered as a conditional value, depending on the information about the problem data.
  • Novelty: The proposed method allows you to generate models in one-shot mode (practically without retraining) for the required task, which significantly reduces the cost of optimization and retraining.
  • Authors: Olga Grebenkova and Bakhteev Oleg

Task 129

  • Name: Spatiotemporal Prediction with Convolutional Networks and Tensor Decompositions
  • Problem description:Generate a set of convolutions from the available data and choose the best one using order and dimensionality reduction techniques.
  • Data: Consumption and price of electricity, ocean currents, dune movement, air currents
  • Literature:

[1](Tensor-based Singular Spectrum Analysis for Automatic Scoring of Sleep EEG) [2](Tensor based singular spectrum analysis for nonstationary source separation)

  • Base algorithm: Caterpillar, tensor caterpillar.
  • Solution: Find a multi-periodic time series, build its tensor representation, decompose into a spectrum, collect, show the forecast.
  • Novelty: Show that a multilinear model is a convenient way to construct convolutions for dimensions in space and time.
  • Authors: Expert - V. V. Strizhov V.V., consultant - Nadezhda Alsakhanova

Task 130

  • Name: Automatic highlighting of terms for topic modeling
  • Task: Build an ATE (Automatic Term Extraction) model for automatic extraction of phrases that are terms of the subject area in the texts of scientific articles. It is supposed to use effective collocation detection methods (TopMine or more modern) and thematic models to determine the "thematic" of the phrase. The model must be trained without a teacher (unsupervised).
  • Data: Collection of scientific articles in the field of machine learning. Marked up articles with highlighted terms for evaluating models.
  • Literature::

El-Kishky A., Song Y., Wang C., Voss C. R., Han J. Scalable topical phrase mining from text corpora // Proc. VLDB Endowment. _ 2014._ Vol. 8, no. 3._Pp. 305_316. Vorontsov K. V. "Probabilistic thematic modeling: theory, models, algorithms and the BigARTM project" (http://www.machinelearning.ru/wiki/images/d/d5/Voron17survey-artm.pdf) Nikolay Shatalov. Unsupervised learning methods for automatically highlighting compound terms in text collections. 2019. VMK MSU. Vladimir Polushin. Topic models for ranking text content recommendations. 2017. VMK MSU. Hanh Thi Hong Tran, Matej Martinc, Jaya Caporusso, Antoine Doucet, Senja Pollak. The Recent Advances in Automatic Term Extraction: A survey. 2023. https://arxiv.org/abs/2301.06767

  • Base algorithm: TopMine collocation search method • BigARTM thematic modeling library. • Modern methods based on neural network language models
  • Solution: Application of the TopMine collocation search algorithm followed by filtering by topic. Selection of thematic model hyperparameters and thematicity criterion. Comparison of this approach with modern methods based on neural network models of the language.
  • Novelty: Previous studies of the proposed approach have shown good results both in terms of completeness and computational efficiency. However, they have not yet been compared with neural network models.
  • Authors: Polina Potapova, Vorontsov K.V.

Task 131

  • Name: Iterative improvement of the topic model with user feedback
  • Task: Topic modeling is widely used in socio-humanitarian research to understand the thematic structure of large text collections. A typical use case would involve the user rating topics as relevant, irrelevant, and junk. If the number of garbage topics is too large, then the user tries to build another model. The task is to use custom markup for each such rebuild in such a way that relevant topics are preserved, new relevant ones stand out from irrelevant and garbage topics if possible, and there are as few garbage topics as possible.
  • Data: Any collection of natural language texts about which the thematic structure is known (about how many topics, how many documents on different topics) is suitable as data. For example, you can take a collection of Lenta news, a Wikipedia dump, posts from Habrahabr, 20 Newsgroups, Reuters, articles from PostNauka. The subject of the collection should be of interest to the researcher himself, so that there is motivation to evaluate topics manually.
  • Literature:: Vorontsov K. V. "Probabilistic thematic modeling: theory, models, algorithms and the BigARTM project" (http://www.machinelearning.ru/wiki/images/d/d5/Voron17survey-artm.pdf ), Alekseev V. et al. "TopicBank: Collection of coherent topics using multiple model training with their further use for topic model validation" (https://www.sciencedirect.com/science/article/pii/S0169023X21000483)
  • Solution: Using the BigARTM theme modeling library. Use of smoothing and decorrelation regularizers. Development of methods of initialization when rebuilding thematic models. Finding a ready-made tool or developing a simple, fast, convenient way to view and markup topics.
  • Novelty: The problem of non-uniqueness and instability of models still does not have a final solution in probabilistic thematic modeling. The proposed study is an important step towards building models with the maximum number of interpretable topics that are meaningfully useful from the point of view of humanitarian research.
  • Authors: Vasily Alekseev, Vorontsov K. V.

Task 132

  • Name: Ranking of scientific articles for semi-automatic summarization
  • Task: Build a ranking model that takes a selection of texts of scientific articles as input and outputs the sequence of their mention in the abstract.
  • Data: - Overview sections (for example, Introduction and Related Work) of articles from the S2ORC collection (81.1M English-language articles) are used as a training sample. The object of the training set is a sequence of references to articles from the bibliography mentioned in the review sections. For each document there is a set of metadata - year of publication, journal, number of citations, number of citations of the author, etc. Also, there is an abstract and, possibly, the full text of the article. - Kendall's rank correlation coefficient is used as a metric.
  • Literature:: Kryzhanovskaya S. Yu. "Technology of semi-automatic summation of thematic collections of scientific articles", Vlasov A. V. "Methods of semi-automatic summation of collections of scientific articles", [Kryzhanovskaya S. Yu., Vorontsov K. V "Technology for semi-automatic summarization of thematic collections of scientific articles"] (http://www.machinelearning.ru/wiki/images/f/ff/Idp22.pdf, p. 371), S2ORC: The Semantic Scholar Open Research Corpus
  • Base algorithm: Pair-wise ranking methods. Gradient boosting.
  • Solution: The simplest solution is to rank the articles in chronological order, according to the year they were published. To solve the problem, it is proposed to build a ranking model based on gradient boosting. As signs, you can use the year of publication, the citation of the article, the citation of its authors, the semantic proximity of the publication to the review, to its local context, etc.
  • Novelty: Task is the first step for semi-automatic summarization of thematic collections of scientific publications (machine aided human summarization, MAHS). After the abstract script is built, the system generates prompt phrases for each article, from which the user selects phrases to continue his abstract.
  • Author: Kryzhanovskaya Svetlana, Vorontsov K. V.

Task 133

  • Name: Diffusion models in the problem of generating the structure of a molecule with optimal energy
  • Task: For an organic small molecule (the number of atoms is less than 100), knowing only the topology of the molecular graph is not enough to obtain the spatial structure. A molecule can have many possible configurations (conformers), each of which corresponds to a local minimum of the potential. In practice, of greatest interest are the most stable conformers, which have the lowest energy. Recent studies show the success of the application of diffusion models for the generation of molecular structures. This approach shows advanced results in the problem of generating molecules and their conformers for a small number of heavy atoms (QM9 dataset - up to 9 heavy atoms in a molecule), as well as in assessing the binding of a molecule and a protein. It is proposed to build a model for the generation of conformers with minimum energy for larger molecules.
  • Data: Base dataset - QM9
  • Literature::

Different theoretical approaches to the diffusion model: https://arxiv.org/abs/2011.13456 Diffusion in molecular generation: https://arxiv.org/abs/2203.17003 Diffusion in the problem of binding a protein and a molecule: https://arxiv.org/abs/2210.01776 Diffusion in the problem of conformer generation: https://arxiv.org/abs/2203.02923 Tutorial on equivariant neural networks: https://arxiv.org/abs/2207.09453

  • Base algorithm: GeoDiff[4].
  • Solution: Implement conformer generation similar to DiffDock[3] for QM9 dataset. Check the performance of the model for larger molecules.
  • Novelty: The novelty of the work lies in the design of a model for generating large conformers, which is of great practical importance.
  • Author: Philip Nikitin

Task 134

  • Name: Combining distillation of models and data
  • Task: Knowledge distillation - the transfer of knowledge from a more meaningful representation to a compact, concise representation. There are two kinds of knowledge distillation. The first is the distillation of models. In this case, the large model transfers knowledge (distilled) to the small model. The second is data distillation. In this case, a minimum data set is created, on which, after training the model, it achieves a quality comparable to training on a full sample. At the moment, there is no solution that can implement simultaneous distillation of model and knowledge. Therefore, the goal of the task is to propose a basic solution for model distillation and compare with approaches to model distillation and data distillation.
  • Data: MNIST handwritten digit sampling, CIFAR-10 image sampling
  • Literature::

A collection of various papers on the distillation of data. Review on methods of distillation models. Basic knowledge distillation solution Basic solution for model distillation

  • Base algorithm: Basic Model Distillation Solution - Hinton Distillation Basic Dataset Distillation Solution - Dataset Distillation by Matching Training Trajectories
  • Solution: It is proposed to implement data distillation as a basic algorithm. Then train a larger model on the data and distill it into a smaller model. Next, compare with the original model and the model trained on distilled data.
  • Novelty: The novelty of the work lies in the combination of two distillation approaches, which has not been implemented before
  • Authors: Andrey Filatov

Task 135

  • Name: Proximity measures in self-supervised learning tasks
  • Task: The idea of self-supervised learning is to solve an artificially selected task to get useful representations of data without markup. One of the most popular approaches is the use of contrastive learning, during which the model is trained to minimize the distance between representations of augmented copies of the same object. The purpose of the task is to investigate the quality of the resulting representations depending on the choice of the proximity measure (similarity measure) used in training, and to offer our own version of distance measurement
  • Data: CIFAR-100
  • Literature::

Solution using squared Euclidean distance Solution using cosine similarity Decision based on the information principle

  • Base algorithm: VicReg, Barlow Twins, SimSiam
  • Solution: One of the distance options that can be proposed is an analogue of the Vaserstein metric, which would allow taking into account the dependencies between features.
  • Novelty: Propose a new way to determine the measure of proximity, which would be theoretically justified / contributed to obtaining representations with given properties
  • Authors: Polina Barabanshchikova

Task 136

  • Name: Stochastic Newton with Arbitrary Sampling
  • Task: We analyze second order methods solving Empirical Risk Minimization problem of the form min f(x) in R^d. Here x is a parameter vector of some Machine Learning model, f_i(x) is a loss function on i-th training point (a_i,b_i). Our desire to solve it using Newton-type method that requires access to only one data point per iteration. We investigate different sampling strategies of index i_k on iteration k. See description in PDF.
  • Data: It is proposed to use open SVM library as a data for experimental part of the work.
  • References:

Stochastic Newton and Cubic Newton Methods with Simple Local Linear-Quadratic Rates Parallel coordinate descent methods for big data optimization

  • Base algorithm: As a base method it is proposed to use Algorithm 1 from the paper Stochastic Newton and Cubic Newton Methods with Simple Local Linear-Quadratic Rates.
  • Solution: Is is proposed to adjust existing sampling strategies from Parallel coordinate descent methods for big data optimization in this work.
  • Novelty: In the literature of Second Order methods there are a few works on incremental methods. The idea is to analyze the existing method by applying different sampling strategies. It is known that the proper sampling strategies may improve the performance of a method.
  • Authors: Islamov Rustem, Vadim Strijov

Task 137

  • Task: Binary Neural Networks. Lossless picture quality for binary neural networks in pixel-level tasks. (Binary networks in img-to-img tasks)
  • Authors: Ilya Zharikov

Task 138

  • Task: Post Training Quantization. Flexible continuous modification for SOTA post training quantization methods to make them lossless.
  • Authors: Ilya Zharikov

Task 139

  • Name: Distillation of models on multidomain selections.
  • Task: The problem of reducing the complexity of the approximating model when transferred to new data of lower power is investigated.
  • Data: Samples MNIST, CIFAR-10, CIFAR-100, Amazon products.
  • Literature:: Diploma Kamil Bayazitov
  • Base algorithm: The basic solution and experiments are presented in the thesis.
  • Authors: Grabovoi Andrey

Task 140

  • Name: Tailoring the architecture of a performance-controlled deep learning model
  • Task: considers Task adapting the structure of a trained deep learning model for limited computing resources. It is assumed that the resulting architecture (or several architectures) should work efficiently on several types of computing servers (for example, on different GPU models or different mobile devices). It is required to propose a model search method that allows controlling its complexity taking into account the target performance characteristics.
  • Data: MNIST, CIFAR
  • Literature::

Grebenkova O.S., Bakhteev Oleg O., Strizhov V.V. V.V. Variational optimization of a deep learning model with complexity control // Informatics and its applications, 2021, 15(2). PDF Yakovlev K. D. et al. Neural Architecture Search with Structure Complexity Control //Recent Trends in Analysis of Images, Social Networks and Texts: 10th International Conference, AIST 2021, Tbilisi, Georgia, December 16–18, 2021, Revised Selected Papers. - Cham: Springer International Publishing, 2022. - pp. 207-219. FBNet: choosing a model architecture based on target characteristics

  • Base algorithm: FBNet and random search of model substructure
  • Solution: The proposed method is to use a differentiable neural network architecture search algorithm (FBNet) with parameter complexity control using a hypernet. A hypernetwork is a model that generates the structure of the model depending on the input parameters. It is proposed to use the normalized running time of basic operations on target computing resources as hypernet parameters. Thus, the resulting model will allow adapting the architecture of the model for an arbitrary device. * Novelty: The proposed method allows you to control the complexity of the model, in the process of searching for an architecture without additional heuristics.
  • Authors: Konstantin Yakovlev, Bakhteev Oleg

2022

My first scientific paper, spring semester 2022

List of tasks

Task 100 (group)

  • Name: Multi-model representation of dynamical systems
  • Task: The system described by attractors in several phase spaces is considered. Particular models are constructed that approximate measurements of the state of the system in each space. A matching multimodel is built. The parameters of private models are specified.
  • Data: Human motion video, accelerometer, gyroscope, electroencephalogram signals
  • Literature: Our work on accelerometers and BCI, dissertations by Motrenko, Isachenko, Grabovoi
  • Base algorithm: Particular models are neural networks, multimodel is canonical correlation analysis and multimodel is distilled.
  • Solution: Generalize canonical correlation analysis and distillation to the case of an arbitrary number of models.
  • Novelty: Alignment space built for a set of heterogeneous models
  • Authors: A.V. Grabovoi, Strizhov V.V.

Task 90

  • Name: Hand movement recovery from video
  • Task: A skeletal representation of a person's pose is restored from the video sequence. The trajectory of the movement of human limbs sets the initial phase space. The accelerometer signal from the limbs sets the target phase space. Build a model that connects the attractors of the trajectories of the source and target spaces.
  • Data: The initial sample is collected by the authors of the project. Parts of the selection are in the library examples.
  • Literature:
  • Base algorithm:
  • Solution: Theoretical part executed by the extended command. Perform a theoretical study: show that the canonical correlation analysis method (and in particular the PLS, NNPLS, seq2seq, Neur ODE methods) are special cases of the Sugihara convergent cross mapping method.
  • Novelty: A reversible model has been introduced that maps the coordinates recovered from the video sequence into the accelerations of the mobile phone's accelerometer.
  • Authors: A.D. Kurdyukova, R.I. Isachenko, Strizhov V.V.

Task 91

  • Name: Clustering human movement trajectories
  • Task: This paper analyzes the periodic signals in the time series to recognize human activity by using a mobile accelerometer. Each point in the timeline corresponds to a segment of historical time series. This segments form a phase trajectory in phase space of human activity. The principal components of segments of the phase trajectory are treated as feature descriptions at the point in the timeline. The paper introduces a new distance function between the points in new feature space. To reval changes of types of the human activity the paper proposes an algorithm. This algorithm clusters points of the timeline by using a pairwise distances matrix. The algorithm was tested on synthetic and real data. This real data were obtained from a mobile accelerometer
  • Data: USC-HAD, new accelerometer samples
  • Literature: Grabovoy A.V., Strijov V.V. Quasi-periodic time series clustering for human activity recognition // Lobachevskii Journal of Mathematics, 2020, 41 : 333-339.
  • Base algorithm: Caterpillar
  • Solution: Bring Grabovoi's article from the Lobachevsky Journal of Mathematics to perfection
  • Novelty: Use Neuro ODE to plot the phase trajectory and classify it
  • Authors: A.V. Grabovoi (ask!!), Strizhov V.V.

Task 97

  • Name: Anti-distillation or teacher training: knowledge transfer from a simple model to a complex one
  • Task: The task of adapting the model to a new sample with a large amount of information is considered. For adaptation, it is proposed to build a new model of greater complexity with further transfer of information from a simple model to it. When transferring information, it is necessary to take into account not only the quality of the forecast on the original sample, but also the adaptability of the new model to the new sample and the robustness of the solution obtained.
  • Data: MNIST handwritten digit sampling, CIFAR-10 image sampling
  • Literature: Original distillation problem statement: Hinton G. et al. Distilling the knowledge in a neural network //arXiv preprint arXiv:1503.02531
  • Base algorithm: It is proposed to increase the complexity of the model by including constant values close to zero in the model. This approach is basic, because can lead to a decrease in the robustness of the model and worse adaptability to a new sample.
  • Solution: It is proposed to consider several approaches to increase the complexity of the model, including both probabilistic (adding noise to new parameters, taking into account operational requirements) and algebraic (expanding the parametric space of the model, taking into account the requirements for robustness and constant Lipschitz of the original model)
  • Novelty: obtaining a method that allows you to adapt the existing model to complicate the training sample without losing information
  • Authors: Bakhteev, Grabovoi, Strijov V.V.

Task 98

  • Name: Deep learning model selection with expert model matching control
  • Task: is considered Task classification. An expert model of low complexity is specified. It is required to build a deep learning model that gives a high quality of the forecast and is similar in behavior to the expert model.
  • Data: Sociological samples, CIFAR image sample
  • Literature: Yakovlev Konstantin, Grebenkova Olga, Bakhteev Oleg, Strijov Vadim. Neural architecture search with structure complexity control // Communications in Computer and Information Science (Proceedings of the 10th International Conference on Analysis of Images, Social Networks and Texts), 2021
  • Base algorithm: building an expert model.
  • Solution: The proposed method consists in hypernetworks with control of the consistency of the found model with the expert model. A hypernetwork is a deep learning model that generates the parameters of the target model.
  • Novelty: the proposed method allows to take into account expert judgment in the process of model selection and architecture search.
  • Authors: Grebenkova, Bakhteev, Strijov V.V.

Task 99

  • Name: Selection of interpretable convolutional deep learning models
  • Task: Considers the Task of choosing an interpretable deep learning classification model. Interpretability is understood as the ability of the model to: a) return the most significant features of an object for classification, b) determine clusters of objects that are similar from the point of view of the classifier
  • Data: MNIST handwritten digit sampling, CIFAR-10 image sampling
  • Literature:
  1) Exact and Consistent Interpretation for Piecewise Linear Neural Networks: A Closed Form Solution
  2) "Why Should I Trust You?": Explaining the Predictions of Any Classifier
  • Base algorithm: The LIME(1) algorithm interprets the model by local approximation
  • Solution: A solution based on the method described in (2) is proposed. In this paper, a generalization of the multilayer perzpetron model with a piecewise linear activation function was proposed. Such an activation function allows us to consider the classifier for each sample object as a locally linear one, without using approximation. It is proposed to generalize the proposed approach to the main nonlinear functions used in convolutional neural networks: convolution, pooling and normalization functions.
  • Novelty: is to obtain a new class of neural models that lend themselves to good interpretation.
  • Authors: Yakovlev, Bakhteev, Strijov V.V.

Task 101

  • Title: Stochastic Newton with Arbitrary Sampling
  • Problem: We analyze second order methods solving Empirical Risk Minimization problem of the form min f(x) in R^d. Here x is a parameter vector of some Machine Learning model, f_i(x) is a loss function on i-th training point (a_i,b_i). Our desire to solve it using Newton-type method that requires access to only one data point per iteration. We investigate different sampling strategies of index i_k on iteration k. See description in PDF.
  • Dataset: It is proposed to use open SVM library as a data for experimental part of the work.
  • References:
 1. Stochastic Newton and Cubic Newton Methods with Simple Local Linear-Quadratic Rates
 2. Parallel coordinate descent methods for big data optimization
  • Base algorithm: As a base method it is proposed to use Algorithm 1 from the paper Stochastic Newton and Cubic Newton Methods with Simple Local Linear-Quadratic Rates.
  • Solution: Is is proposed to adjust existing sampling strategies from Parallel coordinate descent methods for big data optimization in this work.
  • Novelty: In the literature of Second Order methods there are a few works on incremental methods. The idea is to analyze the existing method by applying different sampling strategies. It is known that the proper sampling strategies may improve the performance of a method.
  • Authors: Islamov Rustem, Vadim Strijov

Task 107

  • Title: Compression for Federated Random Reshuffling
  • Problem: We analyze first order methods solving Empirical Risk Minimization problem of the form min f(x) in R^d. Here x is a parameter vector of some Machine Learning model, f_i(x) is a loss function on i-th training point (a_i,b_i). We focus on distributed setting of this problem. We are going to apply compression techniques to reduce number of communicated bits to overcome communication bottleneck. Also we want to combine it with server-side updates. We desire to generalize and get improvement in theory and practice.
  • Dataset: It is proposed to use open SVM library as a data for experimental part of the work.
  • References:
 1. Federated Random Reshuffling with Compression and Variance Reduction
 2. Proximal and Federated Random Reshuffling
 3. Server-Side Stepsizes and Sampling Without Replacement Provably Help in Federated Optimization
  • Base algorithm: As a base method we use Algorithm 3 from Proximal and Federated Random Reshuffling.
  • Solution: Is is proposed to combine the method with two stepsizes with compression operators.
  • Novelty: This would be the first method combining 4 popular federated learning techniques: local steps, compression, reshuffling of data and two stepsizes.
  • Authors: Grigory Malinovsky

Task 108

  • Name: Distillation of knowledge using sample representation in the common latent space of models
  • Task: Considers the Task of distillation - the transfer of information from one or more teacher models to the student. A special case is considered when teachers have incomplete information about the sample, and each model has useful information only about some subset.
  • Data: Sample CIFAR-10 images; sampling of handwritten MNIST digits
  • Literature:
  1. Hinton G. et al. Distilling the knowledge in a neural network //arXiv preprint arXiv:1503.02531. - 2015. - Vol. 2. - No. 7.
  2. Oki H. et al. Triplet Loss for Knowledge Distillation //2020 International Joint Conference on Neural Networks (IJCNN). - IEEE, 2020. - P. 1-7.
  • Base algorithm: Hinton distillation [1].
  • Solution: It is proposed to consider hidden representations of teachers and students obtained using dimensionality reduction algorithms. To align the model spaces, it is proposed to use the autoencoder model with triplet constraints (see, for example, [2]).
  • Novelty: The proposed method will allow the distillation of heterogeneous models, using information from several teachers.
  • Authors: Gorpinich, Bakhteev, Strijov V.V.

Task 93

  • Name: Estimating the risk of forest fires using machine learning methods.
  • Task: Wildfire risk prediction based on climate variables (water/air temperature, atmospheric pressure) since 1991. Forecasting is carried out (a) in the short-term range (2-5 years; stationary time series) and (b) in the long-term range (up to 50 years; non-stationary time series). A feature of forecasting in the long range is the (probable) significant change in the behavior of climate variables (CMIP5 scenarios).

The key features of problem (1) are the need for a sufficiently accurate prediction of extreme risk values (maximum values of the time series), while the algorithm can make a significant number of errors in the region of small values of the series. (2) the spatial data structure of the series.

  • Data:
  1. Google Earth Data - data on climate variables and landscape available via API (there is a jupyter notebook through which you can download data locally)
  2. CMIP5 climate scenarios (there is a jupyter notebook through which you can download data locally)
  3. Wildfire Risk Database
  4. Severe Weather Dataset
  • Literature:
  1. Daizong Ding, Mi Zhang, Xudong Pan, Min Yang, Xiangnan He. Modeling Extreme Events in Time Series Prediction. KDD-2019.
  2. Roman Kail, Alexey Zaytsev, Evgeny Burnaev. Recurrent Convolutional Neural Networks help to predict the location of Earthquakes.
  3. Nikolay Laptev, Jason Yosinski, Li Erran Li, Slawek Smyl. Time-series Extreme Event Forecasting with Neural Networks at Uber.
  • Base algorithm: (1) method from article 1, (2). ST-LSTM
  • Solution: is proposed to solve the problem in two steps. At the first step, Algorithm 1 (with the addition of a spatial component) restores (averaged over a certain range) the behavior of the time series. Next, the discrepancy between the values of the series and the model is analyzed. Based on this, the noise distribution is restored and a probabilistic model is built to achieve a certain level of risk in a given territory in the required time range.
  • Novelty: (geo)-spatial time series prediction is an open area with great potential for theoretical and practical work. In particular, fire risk assessment is necessary for (1) predicting the probability of accidents (electric power industry, gas transport complex); (2) prioritization of fire prevention measures by region; (3) assessing the financial risks of companies operating in the region.
  • Authors: Yuri Maksimov, Alexey Zaitsev
  • Consultants: Yuri Maksimov, Alexey Zaitsev, Alexander Lukashevich.

Task 94

  • Name: Hail forecast using graph neural networks
  • Task: Hail risk prediction based on climate variables (water/air temperature, atmospheric pressure) since 1991. Forecasting is carried out (a) in the short-term range (2-5 years; stationary time series) and (b) in the long-term range (up to 50 years; non-stationary time series). A feature of forecasting in the long range is the (probable) significant change in the behavior of climate variables (CMIP5 scenarios).

Key features of the task (1) rare events, the case of hail in Russia over the past 30 years was less than 700 throughout the country (2) the spatial structure of the data series.

  • Data:
  1. Google Earth Data - data on climate variables and landscape available via API (there is a jupyter notebook through which you can download data locally)

https://developers.google.com/earth-engine/datasets/catalog/IDAHO_EPSCOR_TERRACLIMATE

  2. CMIP5 climate scenarios (there is a jupyter notebook through which you can download data locally) https://www.worldclim.org/data/cmip6/cmip6_clim2.5m.html
  3. NOAA Storm Events Database https://www.ncdc.noaa.gov/stormevents/ftp.jsp
  4. European Severe Weather Database https://eswd.eu/cgi-bin/eswd.cgi
  5. Severe Weather Dataset https://www.visualcrossing.com/weather/weather-data-services
  • Literature:
  1. Ayush, Kumar, et al. "Geography-aware self-supervised learning." Proceedings of the IEEE/CVF International Conference on Computer Vision. 2021. https://openaccess.thecvf.com/content/ICCV2021/papers/Ayush_Geography-Aware_Self-Supervised_Learning_ICCV_2021_paper.pdf
  2. Cachay, Salva Rühling, et al. "Graph Neural Networks for Improved El Ni\~ no Forecasting." arXiv preprint arXiv:2012.01598 (2020). NeurIPS Clima Workshop. https://arxiv.org/pdf/2012.01598.pdf
  3. Cai, Lei, et al. "Structural temporal graph neural networks for anomaly detection in dynamic graphs." Proceedings of the 30th ACM International Conference on Information & Knowledge Management. 2021. https://dl.acm.org/doi/pdf/10.1145/3459637.3481955
  • Base algorithm: classification with extremely rare events, the most basic variant of log-regression + SMOTE. The paper proposes to take a combination of algorithms from articles 2 and 3 as a basis.
  • Solution: suggests that a combination of the algorithms from articles 2 and 3 can improve classification in such Tasks with exceptionally rare events. In addition, it is supposed to use physical information to regularize the classifier (combination of temperature/humidity factors at which hail is most likely)
  • __New: (geo)-spatial time series prediction is an open area with great potential for theoretical and practical work. In particular, fire risk assessment is necessary for (1) predicting the probability of damage (agriculture, animal husbandry); (2) assessment of insurance and financial risks.
  • Authors: Yuri Maksimov (point of contact), Alexey Zaitsev
  • Consultants: Yuri Maksimov (point of contact), Alexey Zaitsev, Alexander Bulkin.

Results

Author Topic Links Consultant Letters
Pilkevich Anton Existence conditions for hidden feedback loops in recommender systems GitHub, LinkReview,

Paper, Slides, Video, Video

Khritankov AILB.P-X+R-B-H1CVO.T-EM.H1WJSF
Vladimirov Eduard Восстановление траектории движения руки по видео GitHub, LinkReview,

Paper, Slides

Isachenko (B.O.H1M)ALI+PXRBС+V+TED?
Petrushina Ksenia Anti-Distillation: Knowledge Transfer from Simple Model to a Complex One GitHub, LinkReview,

Paper, Slides

Grabovoi (B.O.H1M)ALIPXRBСVTED
Kornilov Nikita Winterstorm risk prediction via machine learning methods GitHub, LinkReview,

Paper, Slides

Yuri Maksimov (B.O.H1M?)ALIPXRBСV+TE0D
Aliyev Alen Geometric Deep Learning for Protein-Protein Binding Affinity Prediction GitHub, LinkReview,

Paper, Slides

Ilya Igashov (B.O.H1M?)ALIPXRBСVTED?
Lukyanenko Ivan Hail Prediction Using Graph Neural Networks GitHub, [3],

Paper, Slides

Yuri Maksimov (B.O.H1M?)ALIPXRBСV+TED?
Gaponov Maxim Choosing Interpretable Recurrent Deep Learning Models GitHub, LinkReview,

Paper, Slides

Bakhteev Oleg (B.O.H1M)AL+IPXRBСVT???ED
Melnikov Igor Stochastic Newton with Arbitrary Sampling GitHub, LinkReview,

Paper, Slides

Rustem Islamov (B.O.H1M)ALIPXСRBVTED
Zmushko Philip Continuous time when building a BCI neural interface GitHub, LinkReview,

Paper, Slides

Samokhina (B.O.H1M)ALI0P0XR?BСVTE?D?
Tishchenko Evgeny Cross-language duplicate search GitHub, LinkReview,

Paper, Slides

Konstantin Vorontsov (B.O.H1M)ALIPXRB0СV0T?E?D?
Antyshev Tikhon Compression for Federated Random Reshuffling GitHub, LinkReview,

Paper, Slides

Malinovsky (B.O.H1_M?)ALI-PXRBСVT?
Pyzh Vladislav Flood risk prediction via machine learning methods GitHub, LinkReview,

Paper, Online Draft, Slides

Yuri Maksimov (B.O.H10M?)ALI0P0XRBСVT0ED?
Zharov Georgy Forest fire risk assessments using machine learning methods GitHub, LinkReview,

Paper, Slides

Yuri Maksimov (B.O.H1)ALIPX0R0B0С0V0T?E0D?
Muradov Timur Choosing Interpretable Convolutional Deep Learning Models GitHub, LinkReview,

Paper, Slides

Bakhteev (B.O.H1)ALI0P0XRBСV0T0E?D?
Pavlov Dmitry Machine learning approach to startup success prediction GitHub, Online Draft,

Paper, Slides

Anton Moiseev, Yuri Ammosov (B.O.H10M?)ALI?P?XRBСV?T0E0D0

2021

Author Topic Links Consultant Letters Reviewer
Grebenkova Olga (example) Variational optimization of deep learning models with model complexity control LinkReview

GitHub Paper Slides Video

Oleg Bakhteev AILP+UXBR+HCV+TEDWSS Shokorov Vyacheslav

Review

Pilkevich Anton Existence conditions for hidden feedback loops in recommender systems GitHub

LinkReview Paper Slides Video

Khritankov Anton AILB*P-X+R-B-H1CVO*T-EM*H1WJSF Gorpinich Maria

Review

Antonina Kurdyukova| Determining the phase and disorder of human movement based on the signals of wearable devices LinkReview

GitHub Paper Slides Video

Georgy Kormakov AILB*PXBRH1CVO*TEM*WJSF Pilkevich Anton

Review

Yakovlev Konstantin A differentiable search algorithm for model architecture with control over its complexity LinkReview

GitHub Paper Slides Video

Grebenkova Olga AILB*PXBRH1CVO*TEM*WJSF Pyrau Vitaly

Review

Gorpinich Maria Trajectory Regularization of Deep Learning Model Parameters Optimization Based on Knowledge Distillation LinkReview

GitHub Paper Slides Video

Oleg Bakhteev AILB*P+XBRC+VH1O*TEM*WJSF Kulakov Yaroslav

Review

Alexandr Tolmachev Analysis of the QPFS Feature Selection Method for Generalized Linear Models LinkReview

GitHub Paper Slides Video

Aduenko Alexander AILB*PXB-R-H1CVO*TEM*WJSF Antonina Kurdyukova

Review

Kulakov Yaroslav BCI: Selection of consistent models for building a neural interface LinkReview

GitHub Paper Slides Video

Isachenko Roman AILB*PXBRH1CVO*TEM*WJ0SF Zverev Egor

Review

Pyrau Vitaly Experimental comparison of several problems of operational planning of biochemical production. LinkReview

GitHub Paper Slides Video

Trenin Sergey Alekseevich AILB*PXBRH1CVO*TEM*WJSF Yakovlev Konstantin

Review

Bazhenov Andrey Search for the boundaries of the iris by the method of circular projections LinkReview

GitHub Paper Slides Video

Matveev Ivan Alekseevich AILB*PXB0RH1CVO*TEM*WJ0SF
Zverev Egor Learning co-evolution information with natural language processing for protein folding problem LinkReview

GitHub Paper Slides Video

Sergei Grudinin, Ilya Igashov AILB*PXBRH1CVO*TEM*WJSF Alexandr Tolmachev

Review

Gorchakov Vyacheslav Importance Sampling for Chance Constrained Optimization LinkReview

Github Paper Video

Yuri Maksimov AILB*PX0B0R0H1C0V0O*0T0E0M*0W0JS0F Bazhenov Andrey

Review

Lindemann Nikita Training with an expert for a sample with many domains LinkReview

Github Paper Slides

Andrey Grabovoi AILPXBRH1C0V0O*TE0M*0W0J0SF0

Task 74

  • Name: Existence conditions for hidden feedback loops in recommender systems
  • Problem description: In recommender systems, the effect of artificially inadvertently limiting the user's choice due to the adaptation of the model to his preferences (echo chamber / filter bubble) is known. The effect is a special case of hidden feedback loops. (see - Analysis H.F.L.). It is expressed in the fact that by recommending the same objects of interest to the user, the algorithm maximizes the quality of its work. The problem is a) lack of variety b) saturation / volatility of the user's interests.
  • Task: It is clear that the algorithm does not know the interests of the user and the user is not always honest in his choice. Under what conditions, what properties of the learning algorithm and dishonesty (deviation of the user's choice from his interests) will the indicated effect be observed? Clarification. The recommendation algorithm gives the user a_t objects to choose from. The user selects one of them c_t from Bernoulli from the model of interest mu(a_t) . Based on the user's choice, the algorithm changes its internal state w_t and gives the next set of objects to the user. On an infinite horizon, you need to maximize the total reward sum c_t. Find the conditions for the existence of an unlimited growth of user interest in the proposed objects in a recommender system with the Thomson Sampling (TS) MAB algorithm under conditions of noisy user choice c_t. Without noise, it is known that there is always unlimited growth (in the model) [1].
  • Data: are created as part of the experiment (simulation model) by analogy with the article [1], external data is not required.
  • References:
    1. Jiang, R., Chiappa, S., Lattimore, T., György, A. and Kohli, P., 2019, January. Degenerate feedback loops in recommender systems. In Proceedings of the 2019 AAAI/ACM Conference on AI, Ethics, and Society (pp. 383-390).
    2. Khritankov, A. (2021). Hidden Feedback Loops in Machine Learning Systems: A Simulation Model and Preliminary Results. In International Conference on Software Quality (pp. 54-65). Springer, Cham.
    3. Khritankov A. (2021). Hidden feedback loop experiment demo. https://github.com/prog-autom/hidden-demo
  • Basic algorithm: The initial mathematical model of the phenomenon under study is described in the article [1]. The method of experimental research is in the article [2]. The base source code is available at [3]
  • Solution: It is necessary to derive conditions for the existence of positive feedback for the Thomson Sampling Multi-armed Bandit algorithm based on the known theoretical properties of this algorithm. Then check their performance in the simulation model. For verification, a series of experiments is performed with the study of parameter ranges and the estimation of the error (variance) of the simulation. The results are compared with the previously constructed mathematical model of the effect. There is an implementation of the experiment system that can be improved for this task.
  • Novelty: The studied positive feedback effect is observed in real and model systems and is described in many publications as an undesirable phenomenon. There is his model for the limited case of the absence of noise in the user's actions, which is not implemented in practice. Under the proposed conditions, Task has not previously been posed and not solved for recommender systems. For the regression problem, the solution is known.
  • Authors: Expert, consultant - Anton Khritankov

Task 77

  • Name: Determining the phase and disorder of human movement by signals from wearable devices
  • Task: A wide class of periodic movements of a person or an animal is investigated. It is required to find the beginning and end of the movement. It is required to understand when one type of movement ends and another begins. For this, the Task of segmentation of time series is solved. The phase trajectory of one movement is constructed and its actual dimension is found. The purpose of the work is to describe a method for finding the minimum dimension of the phase space. By repetition of the phase, segment the periodic actions of a person. It is also necessary to propose a method for extracting the zero phase in a given space for a specific action. Bonus: find the discord in the phase trajectory and indicate the change in the type of movement. Bonus 2: do this for different phone positions by proposing invariant transformation models.
  • Data: The data consists of time series read from a three-axis accelerometer with an explicit periodic class (walking, running, walking up and down stairs, etc.). It is possible to get your own data from a mobile device, or get model data from the dataset UCI HAR
  • References:
    1. A. P. Motrenko, V. V. Strijov. Extracting fundamental periods to segment biomedical signals // Journal of Biomedical and Health Informatics, 2015, 20(6).P. 1466–1476
1.(Сегментация временных рядов с периодическими действиями: решалась Task сегментации с использованием фазового пространства фиксированной размерности.) PDFURL
    2. A.D. Ignatov, V. V. Strijov. Human activity recognition using quasi-periodic time series collected from a single triaxial accelerometer. // Multimedia Tools and Applications, 2015, P. 1–14.
( Классификация человеческой активности с помощью сегментации временных рядов
: исследовались классификаторы над получаемыми сегментами.) PDFURL
    3. Grabovoy, A.V., Strijov, V.V. Quasi-Periodic Time Series Clustering for Human Activity Recognition. Lobachevskii J Math 41, 333–339 (2020). (Сегментация временных рядов на квазипериодические сегменты
: исследовались методы сегментации с использованием анализа главных компонент and перехода в фазовое пространство.) Text Slides DOI
  • Basic algorithm: The basic algorithm is described in 1 and 3 works, code here, work code 3 author.
  • Solution: It is proposed to consider various dimensionality reduction algorithms and compare different spaces in which the phase trajectory is constructed. Develop an algorithm for finding the minimum dimension of the phase space in which the phase trajectory has no self-intersections up to the standard deviation of the reconstructed trajectory.
  • Novelty: In Motrenko's article, the space dimension is equal to two. This shortcoming must be corrected. The phase trajectory must not intersect itself. And if we can distinguish one type of movement from another within one period (switched from running to a step and realized this within one and a half steps), it will be great.
  • Authors: 
consultants: Kormakov G.V., Tikhonov D.M., Expert Strizhov V.V.

Task 78

  • Name: Importance Sampling for Scenario Approximation of Chance Constrained Optimization
  • Task: Optimization problems with probabilistic constraints are often encountered in engineering practice. For example, the Task of minimizing energy generation in energy networks, with (randomly fluctuating) renewable energy sources. In this case, it is necessary to comply with safety restrictions: voltages at generators and consumers, as well as currents on the lines, must be less than certain thresholds. However, even in the simplest situations, the Task cannot be resolved exactly. The best-known approach is the chance constrained optimization methods, which often give a good approximation. An alternative approach is sampling the network operation modes and solving the problem on the data set of the classification problem: separating bad modes from good ones with a given error of the second kind. At the same time, for a sufficiently accurate solution, a very large amount of data is required, which often makes the problem numerically inefficient. We suggest using “importance sampling” to reduce the number of scenarios. Importance sampling consists of substituting a sample from a nominal solution, which often carries no information since all bad events are very rare, with a synthetic distribution that samples the sample in a neighborhood of bad events.            
  • Problem statement: find the minimum of a convex function (price) under probabilistic constraints (the probability of exceeding a certain threshold for a system of linear/quadratic functions is small) and numerically show the effectiveness of sampling in this problem.
  • Data: Data is available in the pypower and matpower packages as csv files.
  • References: The proposed algorithms are based on 3 articles:
    1. Owen, Maximov, Chertkov. Importance Sampling for the Union of Rare Events with Applications to Power Systems LINK
    2. A. Nemirovski. On safe tractable approximations of chance constraints [4]
    3. S. Tong, A. Subramanyam, and Vi. Rao. Optimization under rare chance constraints. LINK
    4. In addition, the authors of the problem have a draft of the article, in which you need to add a numerical part.
  • Basic algorithm: A list of basic algorithms is provided in this lecture LINK
  • Solution: in numerical experiments, you need to compare the sample size requirements for standard methods (scenario approximation) and using importance sampling to obtain a solution of comparable quality (and inverse Task, having equal sample lengths, compare the quality of the solution)
  • Novelty: Task has long been known in the community and scenario approximation is one of the main methods. At the same time, importance sampling helps to significantly reduce the number of scenarios. We have recently received a number of interesting results on how to calculate optimal samplers, with their use the complexity of the problem will be significantly reduced
  • Authors: Expert – Yuri Maksimov, consultant – Yuri Maksimov and Alexander Lukashevich, student.

Task 79

  • Name: Improving Bayesian Inference in Physics Informed Machine Learning
  • Task: Machine learning methods are currently widely used in physics, in particular, in solving turbulence problems or analyzing the stability of physical networks. At the same time, the key issue is which modes to choose for training models. A frequent choice is a sequence of points that uniformly covers the admissible set. However, often such sequences are not very informative, especially if analytical methods give a region where the system is guaranteed to be stable. The problem proposes several methods of sampling: allowing to take into account this information. Our goal is to compare them and find the one that requires the smallest sample size (empirical comparison).
  • Data: The experiment is proposed to be carried out on model and real data. The simulation experiment consists in analyzing the stability of (slightly non-linear) differential equations (synthetic data is self-generated). The second experiment is to analyze the stability of energy systems (data from matpower, pypower, GridDyn).
  • References:
    1. Art Owen. Quasi Monte Carlo Sampling. LINK 
    2. Jian Cheng & Marek J. Druzdzel. Computational Investigation of Low-Discrepancy Sequences in Simulation Algorithms for Bayesian Networks [5]
    3. A. Owen, Y Maximov, M. Chertkov. Importance Sampling for the Union of Rare Events with Applications to Power Systems [6]
    4. Polson and Solokov. Deep Learning: A Bayesian Perspective [7]
    5. In addition: the authors of the problem have a draft work on this topic
  • Basic algorithm: The basic algorithm we are improving is Quasi Monte Carlo (QMC, LINK ). Task to construct low discrepancy sequences not covering the polyhedral region and the region given by the intersection of the quadratic constraints. Another algorithm with which we need a comparison:

E. Gryazina, B. Polyak. Random Sampling: a Billiard Walk Algorithm LINK и с алгоритмами типа Hit and Run [8]

  • Solution: sampling methods by importance, in particular the extension of the approach (Boy, Ryi, 2014) and (Owen, Maximov, Chertkov, 2017) and their applications to ML/DL for physical problems
  • Novelty: in a significant reduction in sample complexity and the explicit use of existing and analytical results and learning to solve physical problems, before that ML approaches and analytical solutions were mostly parallel courses
  • Authors: Expert Yuri Maksimov, consultant Yuri Maksimov and Alexander Lukashevich, student.

 

Task 81

  • Name: NAS — Generation and selection of neural network architectures
  • Task: The task of choosing the optimal neural network architecture is set as the Task of sampling the vector of structural parameters. The optimality criterion is defined in terms of the accuracy, complexity and stability of the model. The sampling procedure itself consists of two steps: generating a new structure and rejecting this structure if it does not satisfy the optimality criterion. It is proposed to explore various methods of sampling. The formulation of the problem of choosing the optimal structure is described in Potanin-1
  • Data: : Two separate sets are offered as data. The first one consists of one element, this is the popular MNIST dataset. Pros - is a strong and generally accepted baseline, was used as a benchmark for the WANN article, quite large (multi-class classification). The second set is a set of datasets for the regression task. Size varies from very small to quite large. Here is a link to the dataset and laptop to download the data data.
  • References:
    1. Potanin - 1
    2. Potanin - 2. One more work, the text is given to the interested student, but without publication.
    3. Strizhov Factory laboratory Error function
    4. Informtica
    5. WANN
    6. DARTS
    7. Symbols
    8. NEAT
  • Basic algorithm: Closest project, and its code. Actual code from consultant.
  • Solution: A number of experiments have already been performed, where sampling is performed by a genetic algorithm. Acceptable results have been obtained. It is proposed to analyze and improve them. Namely, to distinguish two modules: generation and deviation and compare several types of sampling. Basic - Importance sampling, desirable - Metropolis-Hastings (or even Metropolis-Langevin) sampling. Since the genetic algorithm is considered by us as a process with jumps, it is proposed to take this into account when designing the sampling procedure. The bonus of MH is that it has a Bayesian interpretation. The first level of Bayesian inference as applied to MH is described in [Informatica]. It is required either to rewrite it in terms of the distribution of structural parameters, or to describe both levels in general, moving the structural parameters to the second level (by the way, approximately the same will be in the Aduenko problem).
  • Novelty: Neural networks excel at the tasks of computer vision, reinforcement learning, and natural language processing. One of the main goals of neural networks is to perform well tasks that are currently solved exclusively by humans, that is, natural human neural networks. Artificial neural networks still work very differently from natural neural networks. One of the main differences is that natural neural networks evolve over time, changing the strength of connections and their architecture. Artificial neural networks can adjust the strength of connections using weights, but cannot change their architecture. Therefore, the task of choosing the optimal structures of neural networks for specific tasks seems to be an important step in the development of the capabilities of neural network models.
  • Authors: consultant Mark Potanin, Expert Strizhov V.V.

Task 82

  • Name: Training with an Expert for a sample with many domains.
  • Task: The Task of approximating a multi-domain sample by a single multi-model - a mixture of Experts is considered. As data, it is supposed to use a sample that contains several domains. There is no domain label for each object. Each domain is approximated by a local model. The paper considers a two-stage Task optimization based on the EM algorithm.
  • Data: Samples of reviews from the Amazon site for different types of goods are used as data. It is supposed to use a linear model as a local model, and use tf-idf vectors within each domain as an indicative description of reviews.
  • References:
    1. https://arxiv.org/pdf/1806.00258.pdf
    2. http://www.mysmu.edu/faculty/jingjiang/papers/da_survey.pdf
    3. https://dl.acm.org/doi/pdf/10.1145/3400066
  • Basic algorithm and Solution: The basic solution is presented here. The work uses the expert mixture method for the Multi-Soruce domain adaptation problem. The code for the article is available link.
  • Novelty: At the moment, in machine learning there are more and more tasks related to data that are taken from different sources. In this case, there are samples that consist of a large number of domains. At the moment, there is no complete theoretical justification for constructing mixtures of local models for approximating such types of samples.
  • Authors: Grabovoi A.V., Strizhov V.V.

Task 17

  • Name: BCI: Selection of consistent models for building a neural interface
  • Task: When building brain-computer interface systems, simple, stable models are used. An important step in building an interface is such a model is an adequate choice of model. A wide range of models is considered: linear, simple neural networks, recurrent networks, transformers. The peculiarity of the problem is that when making a prediction, it is required to model not only the initial signal taken from the cerebral cortex, but also the target signal taken from the limbs. Thus, two models are required. In order for them to work together, a space of agreements is being built. It is proposed to explore the properties of this space and the properties of the resulting forecast (neural interface) on various pairs of models.
  • Data: ECoG/EEG brain signal data sets.
    1. Need ECoG (dataset 25 contains EEG, EOG and hand movements) http://bnci-horizon-2020.eu/database/data-sets
    2. neyrotycho — our old data.
  • References::
    1. Yaushev F.Yu., Isachenko R.V., Strizhov V.V. Latent space matching models in the forecasting problem // Systems and Means of Informatics, 2021, 31(1). PDF
    2. Isachenko R.V. Choice of a signal decoding model in high-dimensional spaces. Manuscript, 2021. PDF
    3. Isachenko R.V. Choice of a signal decoding model in high-dimensional spaces. Slides, 2020. [9]
    4. Isachenko R.V., Vladimirova M.R., Strijov V.V. Dimensionality reduction for time series decoding and forecasting problems // DEStech Transactions on Computer Science and Engineering, 2018, 27349 : 286-296. PDF
    5. Isachenko R.V., Strijov V.V. Quadratic Programming Optimization with Feature Selection for Non-linear Models // Lobachevskii Journal of Mathematics, 2018, 39(9) : 1179-1187. PDF
    6. Motrenko A.P., Strijov V.V. Multi-way feature selection for ECoG-based brain-computer interface // Expert Systems with Applications, 2018, 114(30) : 402-413. PDF
    7. Eliseyev A., Aksenova T. Stable and artifact-resistant decoding of 3D hand trajectories from ECoG signals using the generalized additive model //Journal of neural engineering. – 2014.
  • Basic algorithm: Described in the first work. The code is available. In that work, the data is two parts of an image. In our work, the signal of the brain and the movement of the hands. SuperTask: to finish the first job. Also the code and works here.
  • Solution: The case is considered when the initial data are heterogeneous: the spaces of the independent and target variables are of different nature. It is required to build a predictive model that would take into account the dependence in the source space of the independent variable, as well as in the space of the target variable. It is proposed to investigate the accuracy, complexity and stability of pairs of various models. Since the inverse Task is solved when building a forecast, it is required to build inverse transformations for each model. To do this, you can use both basic techniques (PLS) and streams.
  • Novelty: Analysis of the prediction and latent space obtained by a pair of heterogeneous models.
  • Authors: Consultant Roman Isachenko, Expert Strizhov V.V.

Task 69

  • Name: Graph Neural Network in Reaction Yield prediction
  • Task: There are disconnected graphs of source molecules and products in a chemical reaction. The yield of the main product in the reaction is known. It is required to design an algorithm that predicts yield by solving the regression task on given disconnected graphs.
  • Data: Database of reaction from US patents [10]
  • References::
    • [11] A general overview.
    • [12] Relational Graph Convolution Neural Network
    • [13] Transformer architecture
    • [14] Graph neural network learning for chemical compounds synthesis
  • Basic algorithm: Transformer model. The input sequence is a SMILES representation of the source and product molecules.
  • Solution: A pipeline for working with disconnected graphs is proposed. The pipeline includes the construction of extended graph with molecule and reaction representation, Relational Graph Convolution Neural Network, Encoder of Transformer. The method is applied to solve yield predictions.
  • Novelty: A solution for regression problem on the given disconnected graph is constructed; the approach demonstrates better performance compared with other solutions
  • Authors:: Nikitin Filipp, Isayev Olexandr, Strizhov V.V.

Task 84

  • Name: Trajectory Regularization of Deep Learning Model Parameters Optimization Based on Knowledge Distillation
  • Task: The problem of optimizing the parameters of a deep learning model is considered. The case is considered when the responses of a more complex model (teacher model) are available during optimization. The classical approach to solving such a problem is learning based on the responses of a complex model (knowledge distillation). Assignment of hyperparameters is made empirically based on the results of the model on delayed sampling. In this paper, we propose to consider a modification of the approach to knowledge distillation, in which the coefficient of significance of the distilling term, as well as its gradients, act as hyperparameters. Both of these groups of parameters allow you to adjust the optimization of the model parameters. To optimize hyperparameters, it is proposed to consider the optimization problem as a two-level optimization problem, where at the first level of optimization the Task of optimizing the model parameters is solved, and at the second level the Task of optimizing hyperparameters is approximately solved by the value of the loss function on the delayed sample.
  • Data: Sampling of CIFAR-10 images
  • References:
    1. Distillation of knowledge
    2. Hyperparameter Optimization in a Bilevel Problem: Greedy Method
    3. Hyperparameter Optimization in a Bilevel Problem: Comparison of Approaches
    4. Meta Optimization: neural network instead of optimization operator
  • Basic algorithm: Model optimization without distillation and with standard distillation approach
  • Solution: Using a two-level problem for model optimization. The combination of gradients for both terms is processed by a separate model (LSTM)
  • Novelty: A new approach to model distillation will be proposed to significantly improve the performance of models trained in privileged information mode. It is also planned to study the dynamics of changes in hyperparameters in the optimization process.
  • Authors: Oleg Bakhteev, Strizhov V.V.

Task 85

  • Name: A differentiable search algorithm for model architecture with control over its complexity
  • Task: The problem of choosing the structure of a deep learning model with a predetermined complexity is considered. It is required to propose a method for searching for a model that allows controlling its complexity with low computational costs.
  • Data: MNIST, CIFAR
  • References:
    1. Grebenkova O.S., Oleg Bakhteev, Strizhov V.V.Variational optimization of a deep learning model with complexity control // Informatics and its applications, 2021, 15(2). PDF
    2. DARTS
    3. hypernets
  • Basic algorithm: DARTS
  • Solution: The proposed method is to use a differentiable neural network architecture search algorithm (DARTS) with parameter complexity control using a hypernet.
  • Novelty: The proposed method allows you to control the complexity of the model, in the process of searching for an architecture without additional heuristics.
  • Authors: Oleg Bakhteev, Grebenkova O. S.

Task 86

  • Name: Learning co-evolution information with natural language processing for protein folding problem
  • Task: One of the most essential problems in structural bioinformatics is protein fold recognition since the relationship between the protein amino acid sequence and its tertiary structure is revealed by protein folding. A specific protein fold describes the distinctive arrangement of secondary structure elements in the nearly-infinite conformation space, which denotes the structural characteristics of a protein molecule.
  • Problem description:: request
  • Authors: Sergei Grudinin, Maria Kadukova.

Task 87

  • Name: Bayesian choice of structures of generalized linear models
  • Task: The work is devoted to testing methods for feature selection. It is assumed that the sample under study contains a significant number of multicollinear features. Multicollinearity is a strong correlation between the features selected for analysis that jointly affect the target vector, which makes it difficult to estimate regression parameters and identify the relationship between features and the target vector. There is a set of time series containing the readings of various sensors that reflect the state of the device. The readings of the sensors correlate with each other. It is necessary to choose the optimal set of features for solving the forecasting problem.
  • Novelty: One of the most preferred feature selection algorithms has been published. It uses structural parameters. But there is no theoretical justification. It is proposed to build a theory by describing and analyzing various functions of a priori distribution of structural parameters. In works on the search for structures of neural networks, there is also no clear theory and a list of a priori assumptions.
  • Data: Multivariate time series with readings from various sensors from paper 4, for starters, all samples from paper 1.
  • References: Keywords: bootstrap aggregation, Belsley method, vector autoregression.
    1. Katrutsa A.M., Strijov V.V. Comprehensive study of feature selection methods to solve multicollinearity problem according to evaluation criteria // Expert Systems with Applications, 2017, 76 : 1-11. PDF
    2. Katrutsa A.M., Strijov V.V. Stresstest procedure for feature selection algorithms // Chemometrics and Intelligent Laboratory Systems, 2015, 142 : 172-183.  PDF
    3. Strizhov V.V. Error function in regression recovery problems // Factory laboratory. material diagnostics, 2013, 79(5) : 65-73. PDF
    4. Зайцев А.А., Strizhov V.V., Tokmakova A.A. Estimation of hyperparameters of regression models by the maximum likelihood method // Information technologies, 2013, 2 : 11-15. PDF
    5. Kuznetsov M.P., Tokmakova A.A., Strijov V.V. Analytic and stochastic methods of structure parameter estimation // Informatica, 2016, 27(3) : 607-624. PDF
    6. Катруца А.М., Strizhov V.V. The problem of multicollinearity in the selection of features in regression problems // Information technologies, 2015, 1 : 8-18.  PDF
    7. Нейчев Р.Г., Катруца А.М., Strizhov V.V. Selection of the optimal set of features from a multicorrelated set in the forecasting problem. Zavodskaya Lab. material diagnostics, 2016, 82(3) : 68-74. PDF
  • Basic algorithm: Described in Reference 1: Quadratic Programming for QPFS Feature Selection. Code from Roman Isachenko.
  • Solution: It is proposed to consider the structural parameters used in QPFS at the second level of Bayesian inference. Introduce informative a priori distributions of parameters and structural parameters. Compare different a priori assumptions.
  • Novelty: Statistical Analysis of Structural Parameter Space and Visualization
  • Authors: Alexander Aduenko — consultant, Strizhov V.V.

Task 88

  • Name: Search for the boundaries of the iris by the method of circular projections
  • Task: Given a monochrome bitmap of the eye, см. examples. The approximate position of the center of the pupil is also known. The word "approximate" means that the calculated center of the pupil is no more than half of its true radius from the true one. It is necessary to determine the approximate positions of the circles approximating the pupil and iris. The algorithm must be very fast.
  • Data: About 200 thousand eye images. For each, the position of the true circles is marked - for the purpose of training and testing the method being created.
  • Basic algorithm: To speed up work with the image, it is proposed to aggregate data using circular projections of brightness. Circular projection is a function that depends on the radius, the value of which P(r) is equal to the integral of the directed image brightness gradient over a circle of radius r (or along an arc of a circle). Example for one arc (right quadrant) and for four arcs. Having built some circular projections, based on them, you can try to determine the position of the inner and outer borders of the iris (ring) using heuristics and / or a neural network. It is interesting to evaluate the capabilities of the neural network in this task.
  • References: Matveev I.A. Detection of Iris in Image By Interrelated Maxima of Brightness Gradient Projections // Applied and Computational Mathematics. 2010. V.9. N.2. P.252-257 PDF
  • Author: Matveev I.A.

Task 53

  • Name: Solution of an optimization problem combining classification and regression to estimate the binding energy of a protein and small molecules.
  • Task: The goal of the problem is to solve an optimization problem with classification and regression loss functions applied to biological data.
  • Data: Approximately 12,000 complexes of proteins with small molecules. For classification, for each of them there is 1 correct position in space and 18 incorrect ones generated, for regression, each complex corresponds to the value of the binding constant (proportional to energy). The main descriptors are histograms of distributions of distances between different atoms.
  • References::
  • Basic algorithm: In the classification task, we used an algorithm similar to linear SVM, whose relationship with the energy estimate, which is outside the scope of the classification task, is described in the article https://hal.inria.fr/hal-01591154/. For MSE, there is already a formulated dual Task as a regression loss function, with the implementation of which we can start.
  • Solution: The first step is to solve the problem with the MSE in the loss function using a solver that is convenient for you. The main difficulty may be the large dimensionality of the data, but they are sparse. Further it will be possible to change the wording of the problem.
  • Novelty: Many models used to predict the interactions of proteins with ligands are "retrained" for some task. For example, models that are good at predicting binding energies may be poor at selecting a protein-binding molecule from a variety of non-binding ones, and models that are good at determining the correct geometry of the complex may be poor at predicting energies. In this problem, we propose to consider a new approach to combat such overfitting, since the combination of classification and regression loss functions seems to us to be a very natural regularization.
  • Authors: Sergei Grudinin, Maria Kadukova.

Task 75

  • Name: Alignment of image elements using metric models.
  • Task: Character set specified. Each symbol is represented by one file - an image. Image pixel size may vary. All images are known to belong to the same class, such as faces, letters, flowers, or cars. (A more complicated option is to one class, which we are studying and noise classes.) It is known that each image can be combined with another with the help of an equalizing transformation up to noise, or up to some average image. (This image may or may not be present in the sample). This leveling transformation is specified in the base case by a neural network, and in the proposed case - by a parametric transformation from some given class (the first is a special case of the second). The aligned image is compared with the original one using the distance function. If the distance between two images is statistically significant, it is concluded that the images belong to the same class. It is required to 1) propose an adequate model of the alignment transformation that takes into account the assumptions about the nature of the image (for example, only rotation and proportional scaling), 2) propose a distance function, 3) propose a method for finding the average image.
  • Data: Synthetic and real 1) pictures - faces and symbols with rotation and stretch transformation, 2) faces and cars with 3D rotation transformation with 2D projection. Synthetic images are proposed to be created manually using 1) photographs of a sheet of paper, 2) photographs of the surface of the drawing on a balloon.
  • References:
    1. support work - alignment of images using 2D DTW,
    2. support work - alignment of images using neural networks,
    3. DTW alignment work in 2D,
    4. parametric alignment work.
  • Basic algorithm: from work 1.
  • Solution: In the attached file pdf.
  • Novelty: Instead of multidimensional image alignment, parametric alignment is proposed.
  • Authors: Alexey Goncharov, Strizhov V.V.

Task 80

  • Name: Detection of correlations between activity in social networks and capitalization of companies
  • Task: At present, the significant impact on stock quotes, company capitalization and the success or failure of an IPO depends on social factors such as public opinion expressed on social media. A recent notable example is the change in GameStore quotes caused by the surge in activity on Reddit. Our task at the first stage is to identify quotes between the shares of companies in different segments and activity in social networks. That is, it is necessary to identify correlations between significant changes in the company's capitalization and previous bursts (positive or negative) of its discussion in social networks. That is, it is necessary to find the minimum of the loss function when restoring the dependence in various classes of models (parametrics, neural networks, etc.). This Task is part of a large project to analyze the analysis of markets and the impact of social factors on risks (within a team of 5-7 professors), which will lead to a series of publications sufficient to defend a dissertation.
  • Data: Task has a significant engineering context, the data is downloads from quotes on the Moscow Exchange, as well as NYT and reddit data (crawling and parsing is done by standard tools). The student working on this task must have strong engineering skills and a desire to engage in both the practice of machine learning and the engineering parts of the task.
  • References:
    1. Paul S. Adler and Seok-Woo Kwon. Social Capital: Prospects for a new Concept. [15]   
    2. Kim and Hastak. Social network analysis: Characteristics of online social networks after a disaster LINK
    3. Baumgartner, Jason, et al. "The pushshift reddit dataset." Proceedings of the International AAAI Conference on Web and Social Media. Vol. 14. 2020. [16]
  • Basic algorithm: The basic algorithms are LSTM and Graph neural networks.
  • Solution: Let's start by using LSTM, then try some of its standard extensions
  • Novelty: In this area, there are a lot of economic, model solutions, but the accuracy of these solutions is not always high. The use of modern ML/DL models is expected to significantly improve the quality of the solution.
  • Authors: Expert Yuri Maksimov, consultant Yuri Maksimov, student.

Task 88b

  • Name: Finding a Pupil in an Eye Image Using the Luminance Projection Method
  • Task: Given a monochrome bitmap of the eye, examples. It is necessary to determine the approximate coordinates of the center of the pupil. The word "approximate" means that the calculated pupil center must lie inside a circle centered at the pupil's true center and half the true radius. The algorithm must be very fast.
  • Data: About 200 thousand eye images. For each, the position of the true circle is marked - for the purpose of training and testing the method being created.

Basic algorithm: To speed up work with the image, it is proposed to aggregate data using brightness projections. Image brightness is a function of two discrete arguments. Its projection on the horizontal axis is equal to. Similarly, projections are constructed on axes with an inclination. Having built several projections (two, four), based on them, you can try to determine the position of the pupil (compact dark area) using heuristics and / or a neural network. It is interesting to evaluate the capabilities of the neural network in this task.

  • References: Zhi-Hua Zhou, Xin Geng Projection functions for eye detection // Pattern Recognition. 2004. V.37ю N.5. P.1049-1056. PDF
  • Author: Matveev I.A.

Task 88c

  • Name: Searching for a century in an image as a parabolic contour using the projection method.
  • Task: Given a monochrome bitmap of the eye, examples. It is necessary to find the contour of the upper eyelid as a parabola, that is, to determine the parameters.
  • Data: About 200 thousand eye images. For some (about 2500), a human expert marked the position of a parabola that approximates the eyelid.
  • Basic algorithm: The first step is pre-processing the image with a vertical gradient filter with further binarization, below is a typical result. There are various options for the next step. For example, if the coordinates of the pupil are known, you can set the region of interest (from above) and in it, using the selected points, construct a parabola by approximation using the least squares method. An example result is given below. More subtle methods are possible, such as finding a parabola using the Hough transform (see Wikipedia). Another way is to use projective methods (Radon transform). The main idea: after specifying the coefficient , apply a coordinate transformation to the image, as a result of which all parabolas of the form formula turn into lines of the form , then, given the coefficient , apply the coordinate transformation where , after which the oblique lines of the formula form become horizontal, which are easy to determine, for example, by horizontal projection (by summing the values in the rows of the matrix of the resulting image. If the coefficients are guessed correctly, the perabola representing the eyelid will give a clear maximum in the projection. By going through the formula (having a physical meaning), you can find those that give the maximum projection value, and consider that the desired parabola - eyelid.
  • References: Wikipedia, articles "Hough Transform", "Radon Transform".
  • Author: Matveev I.A.

Task 62

  • Name: Construction of a method for dynamic alignment of multidimensional time series, resistant to local signal fluctuations.
  • Task: In the process of working with multidimensional time series, the situation of close proximity of sensors corresponding to different measurement channels is common. As a result, small signal shifts in space can lead to signal peak fixation by neighboring sensors, which leads to significant differences in measurements in terms of L2 distance.
    Thus, small signal shifts lead to significant fluctuations in the readings of the sensors. The Task of constructing a distance function between points of time series that is resistant to noise generated by small spatial signal shifts is considered. It is necessary to consider the problem in the approximation of the presence of a map of the location of the sensors.
  • Data:
  • References::
  • Basic algorithm: L2 distance between a pair of measurements.
  • Solution: Use the DTW distance function between two multidimensional time series. Two time axes are aligned, while inside the DTW functional, the distance between the i-th and j-th measurements is chosen such that it is resistant to local “shifts” of the signal. It is required to offer such functionality. The basic solution is L2, the improved solution is DTW between the i-th and j-th dimensions (dtw inside dtw).
    You can suggest some modification, for example, the distance between the hidden layers of the autoencoder for points i and j.
  • Novelty: A method for aligning multidimensional time series is proposed that takes into account small signal fluctuations in space.
  • Authors: Expert - Strizhov V.V., consultants - Gleb Morgachev, Alexey Goncharov.

Task 58

  • Name: Transformation of the Gerchberg-Saxton algorithm using Bayesian neural networks. (or Neural network approach in the problem of phase search for images from the European synchrotron)
  • Task: The aim of the project is to improve the quality of resolution of images of nanosized objects obtained in the laboratories of the European Synchrotron Radiation Foundation.
  • Data: Contact an advisor for data (3GB).

References::

  • Basic algorithm: The transition from direct space to reciprocal space occurs using the Fourier transform. The Fourier transform is a linear transformation. Therefore, it is proposed to approximate it with a neural network. For example, an autoencoder for modeling forward and inverse Fourier transforms.
  • Solution: Transformation of the Gerchberg-Saxton algorithm using Bayesian neural networks. Use of information on physical limitations and expertise.
  • Novelty: Use of information about physical constraints and expert knowledge in the construction of the error function.
  • Authors:: Experts Sergei Grudinin, Yuri Chushkin, Strizhov V.V., consultant Mark Potanin

Task 63

  • Name: Hierarchical alignment of time sequences.
  • Task: Task of alignment of sequences of difficult events is considered. An example is the complex behavior of a person: when considering data from IMU sensors, one can put forward a hypothesis: there is an initial signal, there are aggregates of “elementary actions” and there are aggregates of “actions” of a person. Each of the indicated levels of abstraction can be distinguished and operated on exactly by it.
    In order to accurately recognize the sequence of actions, it is possible to use metric methods (for example, DTW, as a method that is resistant to time shifts). For a more accurate quality of timeline alignment, it is possible to carry out alignment at different levels of abstraction.
    It is proposed to explore such a hierarchical approach to sequence alignment, based on the possibility of applying alignment algorithms to objects of different structures, having a distance function on them.
  • References:
  • Basic algorithm: classic DTW.
  • Solution: It is proposed to perform the transition from one level of abstraction to another by using convolutional and recurrent neural networks. Then the object at the lower level of abstraction is the original signal. At the second level - a signal from the hidden layer of the model (built on the objects of the lower level), the dimension of which is much less, and the upper layer - a signal from the hidden layer of the model (built on the objects of the middle level).
    In this case, DTW is calculated separately between the lower , between the middle and between the upper levels, but the formation of objects for calculating the distance is carried out taking into account the alignment path between the objects of the previous level.
    This method is considered as a way to increase the interpretability of the alignment procedure and the accuracy of the action classification in connection with the transition to higher-level patterns. In addition, a significant increase in speed is expected.
  • Novelty: The idea of aligning time sequences simultaneously at several levels of abstraction is proposed. The method should significantly improve the interpretability of alignment algorithms and increase their speed.
  • Authors: Strizhov V.V. - Expert, Gleb Morgachev, Alexey Goncharov - consultants.

Task 57

  • Name:Additive Regularization and in the Tasks of Privileged Learning in Solving the Problem of Predicting the State of the Ocean
  • Task: There is a sample of data from ocean buoys, it is required to predict the state of the ocean at different points in time.
  • Data: The buoys provide data on wave height, wind speed, wind direction, wave period, sea level pressure, air temperature and sea surface temperature with a resolution of 10 minutes to 1 hour.
  • References:
  • Basic algorithm: Using a simple neural network.
  • Solution:Adding to the basic algorithm (a simple neural network) a system of differential equations. Explore the properties of the parameter space of teacher and student according to the preferred approach.
  • Novelty: Investigation of the parameter space of the teacher and the student and their change. It is possible to set up separate teacher and student models and track the change in their parameters in the optimization process - variance, change in the quality of the student when adding teacher information, complexity.
  • Authors:: Strizhov V.V., Mark Potanin


Task 52

  • Name: Predicting the quality of protein models using spherical convolutions on 3D graphs.
  • Task: The purpose of this work is to create and study a new convolution operation on three-dimensional graphs in the framework of solving the problem of assessing the quality of three-dimensional protein models (task regression on graph nodes).
  • Data: Models generated by CASP competitors are used (http://predictioncenter.org).
  • References::
    • [22] More about the task.
    • [23] Relational inductive biases, deep learning, and graph networks.
    • [24] Geometric deep learning: going beyond euclidean data.
  • Basic algorithm: As a basic algorithm, we will use a neural network based on the graph convolution method, which is generally described in [25].
  • Solution: The presence of a peptide chain in proteins makes it possible to uniquely introduce local coordinate systems for all graph nodes, which makes it possible to create and apply spherical filters regardless of the graph topology.
  • Novelty: In the general case, graphs are irregular structures, and in many graph learning tasks, the sample objects do not have a single topology. Therefore, the existing operations of convolutions on graphs are greatly simplified or do not generalize to different topologies. In this paper, we propose to consider a new method for constructing a convolution operation on three-dimensional graphs, for which it is possible to uniquely choose local coordinate systems associated with each node.
  • Authors: Sergei Grudinin, Ilya Igashov.

Task 44+

  • Name: Early prediction of sufficient sample size for a generalized linear model.
  • Task: The problem of experiment planning is investigated. The Task of estimating a sufficient sample size according to the data is solved. The sample is assumed to be simple. It is described by an adequate model. Otherwise, the sample is generated by a fixed probabilistic model from a known class of models. The sample size is considered sufficient if the model is restored with sufficient confidence. It is required, knowing the model, to estimate a sufficient sample size at the early stages of data collection.
  • Цель: On a small simple iid sample, predict the error on a replenished large one. The predictive model is smooth monotonic in two derivatives. The choice of model is a complete enumeration or genetics. The model depends on the reduced (explore) covariance matrix of the GLM parameters.
  • Data: For the computational experiment, it is proposed to use classical samples from the UCI repository. Link to selections https://github.com/ttgadaev/SampleSizeEstimation/tree/master/datasets
  • References::
    1. Overview of Methods, Motivation and Problem Statement for Sample Size Estimation
    2. http://svn.code.sf.net/p/mlalgorithms/code/PhDThesis/.
    3. Bootstrap method. https://projecteuclid.org/download/pdf_1/euclid.aos/1.

Bishop, C. 2006. Pattern Recognition and Machine Learning. Berlin: Springer. 758 p.

  • Basic algorithm: We will say that the sample size is sufficient if the log-likelihood has a small variance, on a sample of size m calculated using the bootstrap.

We are trying to approximate the dependence of the average value of log-likelihood and its variance on the sample size.

  • Solution: The methods described in the review are asymptotic or require a deliberately large sample size. The new method should be to predict volume in the early stages of experiment design, i.e. when data is scarce.
  • Authors: consultant - Malinovsky G., Strizhov V.V. (Expert)


Task 12

  • Name: Machine translation training without parallel texts.
  • Task: The Task of building a text translation model without the use of parallel texts is considered, i.e. pairs of identical sentences in different languages. This Task occurs when building translation models for low-resource languages (that is, languages for which there is not much data in the public domain).
  • Data: A selection of articles from Wikipedia in two languages.
  • References::
    • [26] Unsupervised Machine Translation Using Monolingual Corpora Only
    • [27] Sequence to sequence.
    • [28] Autoencoding.
    • [29] Training with Monolingual Training Data.
  • Basic algorithm: Unsupervised Machine Translation Using Monolingual Corpora Only.
  • Solution: As a translation model, it is proposed to consider a combination of two auto-encoders, each of which is responsible for presenting sentences in one of the languages. The models are optimized in such a way that the latent spaces of autoencoders for different languages match. As an initial representation of sentences, it is proposed to consider their graph description obtained using multilingual ontologies.
  • Novelty: A method for constructing a translation model is proposed, taking into account graph descriptions of sentences.
  • Authors: Oleg Bakhteev, Strizhov V.V.,


Task 8

  • Name: Generation of features using locally approximating models (Classification of human activities according to measurements of fitness bracelets).
  • Task: It is required to check the feasibility of the hypothesis about the simplicity of sampling for the generated features. Features are the optimal parameters of approximating models. Moreover, the entire sample is not simple and requires a mixture of models to approximate it. Explore the information content of the generated features - the parameters of the approximating models trained on the segments of the original time series. According to the measurements of the accelerometer and gyroscope, it is required to determine the type of activity of the worker. It is assumed that the time series of measurements contain elementary movements that form clusters in the space of time series descriptions. The characteristic duration of the movement is seconds. Time series are labeled with activity type labels: work, leisure. The typical duration of activity is minutes. It is required to restore the type of activity according to the description of the time series and cluster.
  • Data: WISDM accelerometer time series (Time series (library of examples), section Accelerometry).
    • WISDM (Kwapisz, J.R., G.M. Weiss, and S.A. Moore. 2011. Activity recognition using cell phone accelerometers. ACM SigKDD Explorations Newsletter. 12(2):74–82.), USC-HAD или сложнее. Данные акселерометра (Human activity recognition using smart phone embedded sensors: A Linear Dynamical Systems method, W Wang, H Liu, L Yu, F Sun - Neural Networks (IJCNN), 2014)
  • References::
    • Motrenko A.P., Strijov V.V. Extracting fundamental periods to segment human motion time series // Journal of Biomedical and Health Informatics, 2016, Vol. 20, No. 6, 1466 - 1476. URL
    • Карасиков М.Е., Strizhov V.V. Classification of time series in the space of parameters of generating models // Informatics and its applications, 2016.URL
    • Kuznetsov M.P., Ivkin N.P. Algorithm for Classifying Accelerometer Time Series by Combined Feature Description // Machine Learning and Data Analysis. 2015. T. 1, No. 11. C. 1471 - 1483. URL
    • Isachenko R.V., Strizhov V.V. Metric learning in Taskx multiclass classification of time series // Informatics and its applications, 2016, 10(2) : 48-57. URL
    • Zadayanchuk A.I., Popova M.S., Strizhov V.V. Choosing the optimal model for classifying physical activity based on accelerometer measurements // Information technologies, 2016. URL
    • Ignatov A., Strijov V. Human activity recognition using quasiperiodic time series collected from a single triaxial accelerometer // Multimedia Tools and Applications, 2015, 17.05.2015 : 1-14. URL
  • Basic algorithm: Basic algorithm described in [Karasikov, Strizhov: 2016] and [Kuznetsov, Ivkin: 2014].
  • Solution: It is required to build a set of locally approximating models and choose the most adequate ones. Find the optimal segmentation method and the optimal description of the time series. Construct a metric space of descriptions of elementary motions.
  • Novelty: A standard for building locally approximating models has been created. The connection of two characteristic times of the description of human life, the combined statement of the problem.
  • Authors: Expert - Strizhov V.V., consultants - Alexandra Galtseva, Danil Sayranov.

2020

Author Topic Links Consultant Letters Reviewer
Grebenkova Olga Variational optimization of deep learning models with model complexity control LinkReview

GitHub Paper Slides Video

Oleg Bakhteev AILP+UXBR+HCV+TEDWS Shokorov Vyacheslav

Review

Shokorov Vyacheslav Text recognition based on skeletal representation of thick lines and convolutional networks LinkReview

GitHub Paper Slides Video

Denis Ozherelkov AIL Grebenkova Olga

Review

Filatov Andrey Intention forecasting. Investigation of the properties of local models in the spatial decoding of brain signals LinkReview

GitHub Paper Slides Video

Valery Markin AILPHUXBRCVTEDWS Hristolubov Maxim

Review

Islamov Rustem Analysis of the properties of an ensemble of locally approximating models LinkReview

GitHub Paper Slides Video

Andrey Grabovoi AILPHUXBRCVTEDWS Gunaev Ruslan

Review

Zholobov Vladimir Early prediction of sufficient sample size for a generalized linear model. LinkReview

GitHub Paper Slides Video

Grigory Malinovsky AILPHUXBRCVTEWSF Vayser Kirill

Review

Vayser Kirill Additive regularization and its meta parameters when choosing the structure of deep learning networks LinkReview

GitHub Paper Slides Video

Mark Potanin AILP+HUX+BRCV+TEDWS Zholobov Vladimir

Review

Bishuk Anton Solution of an optimization problem combining classification and regression to estimate the binding energy of a protein and small molecules. LinkReview

GitHub Paper Slides Video

Maria Kadukova AILPHUXBRCVTEDH Filippova Anastasia
Filippova Anastasia Step detection for IMU navigation via deep learning LinkReview

GitHub Paper Slides EnglishPaper Video

Tamaz Gadaev AIL0PUXBRCVSF Bishuk Anton

Review

Savelev Nickolay Distributed optimization under Polyak-Loyasievich conditions LinkReview

GitHub Paper Slides Video

A. N. Beznosikov AILPHUXBRCVTEDWS Khary Alexandra

Review

Khary Alexandra Theoretical validity of the application of metric classification methods using dynamic alignment (DTW) to spatiotemporal objects. LinkReview

GitHub Paper Slides Video

Gleb Morgachev, Alexey Goncharov AILPHUXBRCVTEDCWS Savelev Nickolay

Review

Hristolubov Maxim Generating features using locally approximating models (Classification of human activities by measurements of fitness bracelets) LinkReview

GitHub Paper Slides Video

Alexandra Galtseva, Danil Sayranov AILPH Filatov Andrey

Review

Mamonov Kirill Nonlinear ranking of exploratory information search results. LinkReview

GitHub Paper Slides Video

Maxim Eremeev AILPHU+XBRC+V+TEDHWJSF
Pavlichenko Nikita Predicting the quality of protein models using spherical convolutions on 3D graphs. LinkReview

GitHub Paper Slides Video

Sergei Grudinin, Ilya Igashov AILPUXBRHCVTEDH
Sodikov Mahmud, Skachkov Daniel Agnostic neural networks Code

Paper Slides Video

Radoslav Neichev AILPHUXBRC+VTEDHWJSF Kulagin Petr

Review

Gunaev Ruslan Graph Neural Network in Reaction Yield prediction LinkReview

Github Paper Slides Video

Philip Nikitin AILPUXBRHCVTEDHWSF Islamov Rustem

Review

Yaushev Farukh Investigation of ways to match models by reducing the dimension of space LinkReview

Github Paper Slides Video

Roman Isachenko AILPUXBRHCVTEDHWJS Zholobov Vladimir

Review

Task 51

  • Name: Analysis of the properties of an ensemble of locally approximating models.
  • Task: In this paper, we consider the task of constructing a universal approximator --- a multimodel, which consists of a given finite set of local models. Each local model approximates a connected region in feature space. It is assumed that the set of local models cover the entire space of objects. A convex combination of local models is considered as an aggregating function. As the coefficients of the convex combination, we consider a function depending on the object --- the gate function.
  • Required: To construct an algorithm for optimizing the parameters of local models and parameters of the gate function. It is required to propose a metric in the space of objects, a metric in the space of models.
  • Data:
    1. Synthetically generated data.
    2. Energy consumption forecasting data. It is proposed to use the following models as local models: working day, day off. (Energy Consumption, Turk Electricity Consumption German Spot Price).
  • References::
    1. Overview of methods for estimating sample size
    2. Vorontsov's lectures on compositions
    3. Vorontsov's lectures on compositions
    4. Esen Y.S., Wilson J., Gader P.D. Twenty Years of Mixture of Experts. IEEE Transactions on Neural Networks and Learning Systems. 2012. Issues. 23. No 8. P. 1177-1193.
    5. Pavlov K.V. Selection of multilevel models in Tasks classification, 2012
  • Basic algorithm: As a basic algorithm, it is proposed to use a two-level optimization problem, where local models are optimized at one iteration and at the next iteration, the parameters of the gate function are optimized.
  • Authors: Grabovoi A.V. (consultant), Strizhov V.V. (Expert)

Task 54

It is necessary to determine the approximate coordinates of the center of the pupil. The word "approximate" means that the calculated pupil center must lie inside a circle centered at the pupil's true center and half the true radius. The algorithm must be very fast.

  • Data: About 200 thousand eye images. For each, the position of the true circle is marked - for the purpose of training and testing the method being created.
  • Basic algorithm: To speed up work with the image, it is proposed to aggregate data using brightness projections. Image brightness is a function of two discrete arguments I(x, y). Its projection onto the horizontal axis is P(x)=\sum \limits_y I(x,y). Similarly, projections are constructed on axes with an inclination. Having built several projections (two, four), based on them, you can try to determine the position of the pupil (compact dark area) using heuristics and / or a neural network. It is interesting to evaluate the capabilities of the neural network in this task.
  • References:: Zhi-Hua Zhou, Xin Geng Projection functions for eye detection // Pattern Recognition. 2004. V.37ю N.5. P.1049-1056. https://doi.org/10.1016/j.patcog.2003.09.006
  • Authors: Matveev I.A.

Task 55

  • Name: Search for the boundaries of the iris by the method of circular projections
  • Task: Given a monochrome bitmap of the eye, see examples (https://cloud.mail.ru/public/2DBu/5c6F6e3LC). The approximate position of the center of the pupil is also known. The word "approximate" means that the calculated center of the pupil is no more than half of its true radius from the true one. It is necessary to determine the approximate positions of the circles approximating the pupil and iris. The algorithm must be very fast.
  • Data: About 200 thousand eye images. For each, the position of the true circle is marked - for the purpose of training and testing the method being created.
  • Basic algorithm: To speed up work with the image, it is proposed to aggregate data using circular projections of brightness. Circular projection is a function that depends on the radius, the value of which P(r) is equal to the integral of the directed image brightness gradient over a circle of radius r (or along an arc of a circle). Example for one arc (right quadrant) and for four arcs. Having built some circular projections, based on them, you can try to determine the position of the inner and outer borders of the iris (ring) using heuristics and / or a neural network. It is interesting to evaluate the capabilities of the neural network in this task.
  • References:: Matveev I.A. Detection of Iris in Image By Interrelated Maxima of Brightness Gradient Projections // Applied and Computational Mathematics. 2010. V.9. N.2. P.252-257. https://www.researchgate.net/publication/228396639_Detection_of_iris_in_image_by_interrelated_maxima_of_brightness_gradient_projections
  • Authors: Matveev I.A.

Task 56

  • Name: Construction of local and universal interpretable scoring models
  • Task: Build a simple and interpretable scoring system as a superposition of local models, taking into account the requirements for the system to retain knowledge about key customers and features (in other words, take into account new economic phenomena). The model must be a superposition, and each element must be controlled by its own quality criterion. Introduce a schedule for optimizing the structure and parameters of the model: the system must work in a single optimization chain. Propose an algorithm for selecting features and objects.
  • Data:
  1. Data from OTP Bank. The sample contains records of 15,223 clients classified into two classes: 1 - there was a response (1812 clients), 0 - there was no response (13411 clients). Feature descriptions of clients consist of 50 features, which include, in particular, age, gender, social status in relation to work, social status in relation to pension, number of children, number of dependents, education, marital status, branch of work. The data are available at the following addresses: www.machinelearning.ru/wiki/images/2/26/Contest_MMRO15_OTP.rar (sample A), www.machinelearning.ru/wiki/images/5/52/Contest_MMRO15_OTP_(validation).rar (sample B).
  2. Data from Home Credit: https://www.kaggle.com/c/home-credit-default-risk/data
  • References::
  1. Strijov V.V. Error function in regression analysis // Factory Laboratory, 2013, 79(5) : 65-73
  2. Bishop C. M. Linear models for classification / В кн.: Pattern Recognition and Machine Learning. Под ред.: M. Jordan, J. Kleinberg, B. Scholkopf. – New York: Springer Science+Business Media, 2006, pp--203 – 208
  3. Tokmakova A.A. Obtaining Stable Hyperparameter Estimates for Linear Regression Models // Machine Learning and Data Analysis. — 2011. — № 2. — С. 140-155
  4. S. Scitovski and N. Sarlija. Cluster analysis in retail segmentation for credit scoring // CRORR 5. 2014. 235–245
  5. Goncharov A.V. Building Interpretable Deep Learning Models in the Social Ranking Problem
  • Basic algorithm: Iterative weighted least squares (described in (2))
  • Solution: It is proposed to build a scoring system containing such a preprocessing block as a block for generating metric features. It is proposed to investigate the influence of the non-equivalence of objects on the selection of features for the model, to investigate the joint selection of features and objects when building a model. It is required to implement a schedule for optimizing the model structure using an algorithm based on the analysis of covariance matrices of model hyperparameters. The schedule includes a phased replenishment of the set of features and objects. The feature sample size will be determined by controlling the error variance. The main criterion for the quality of the system: ROC AUC (Gini).
  • Novelty:
  1. The model structure optimization schedule must satisfy the requirement to rebuild the model at any time without losing its characteristics.
  2. Accounting for the unequal value of objects in the selection of features
  • Authors: Pugaeva I.V. (consultant), Strizhov V.V. (Expert)

Task 59

  • Name: Distributed optimization under Polyak-Loyasievich conditions
  • Task: The task is to efficiently solve large systems of nonlinear equations using a network of calculators.
  • Solution: A new method for decentralized distributed solution of systems of nonlinear equations under Polyak-Loyasievich's conditions is proposed. The approach is based on the fact that the distributed optimization problem can be represented as a composite optimization problem (see 2 from the literature), which in turn can be solved by analogs of the similar triangles or sliding method (see 2 from the literature).
  • Basic algorithm: The proposed method is compared with gradient descent and accelerated gradient descent
  • References:
  1. Linear Convergence of Gradient and Proximal-GradientMethods Under the Polyak- Lojasiewicz Condition https://arxiv.org/pdf/1608.04636.pdf
  2. Linear Convergence for Distributed Optimization Under the Polyak-Łojasiewicz Condition https://arxiv.org/pdf/1912.12110.pdf
  3. Optimal Decentralized Distributed Algorithms for Stochastic ConvexOptimization https://arxiv.org/pdf/1911.07363.pdf
  4. Modern numerical optimization methods, universal gradient descent method https://arxiv.org/ftp/arxiv/papers/1711/1711.00394.pdf
  • Novelty: Reduction of a distributed optimization problem to a composite optimization problem and its solution under Polyak-Loyasievich conditions
  • Authors: Expert — А.В. Гасников, consultant — А.Н. Безносиков
  • Comment: it is important to set up a computational experiment in this task, otherwise the task will be poorly compatible with the course.

Task 17

  • Name: Intention forecasting. Investigation of the properties of local models in the spatial decoding of brain signals
  • Task: When building brain-computer interface systems, simple, stable models are used. An important stage in the construction of such a model is the construction of an adequate feature space. Previously, such a Task was solved by extracting features from the frequency characteristics of signals.
  • Data: ECoG/EEG brain signal data sets.
  • References::
    1. Motrenko A.P., Strijov V.V. Multi-way feature selection for ECoG-based brain-computer Interface // Expert systems with applications. - 2018.
    2. Eliseyev A., Aksenova T. Stable and artifact-resistant decoding of 3D hand trajectories from ECoG signals using the generalized additive model //Journal of neural engineering. – 2014.
  • Basic algorithm: The comparison is proposed to be made with the partial least squares algorithm.
  • Solution: In this paper, it is proposed to take into account the spatial dependence between sensors that read data. To do this, it is necessary to locally model the spatial impulse/signal and build a predictive model based on the local description.
  • Novelty: An essentially new way of constructing a feature description in the problem of signal decoding is proposed. Bonus: analysis of changes in the structure of the model, adaptation of the structure when the sample changes.
  • Authors: Strizhov V.V., Roman Isachenko - Experts, consultants – Valery Markin, Alina Samokhina

Task 9

  • Name: Text recognition based on skeletal representation of thick lines and convolutional networks
  • Task: It is required to build two CNNs, one recognizes a raster representation of an image, the other a vector one.
  • Data: Fonts in raster representation.
  • References::List of works [30], in particular arXiv:1611.03199 and
    • Goyal P., Ferrara E. Graph embedding techniques, applications, and performance: A survey. arXiv:1705.02801, 2017.
    • Cai H., Zheng V.W., Chang K.C.-C. A comprehensive survey of graph embedding: Problems, techniques and applications. arXiv:1709.07604, 2017.
    • Grover A., Leskovec J. node2vec: Scalable Feature Learning for Networks. arXiv:1607.00653, 2016.
    • Mestetskiy L., Semenov A. Binary Image Skeleton - Continuous Approach // Proceedings 3rd International Conference on Computer Vision Theory and Applications, VISAPP 2008. P. 251-258. URL
    • Kushnir O.A., Seredin O.S., Stepanov A.V. Experimental study of regularization parameters and approximation of skeletal graphs of binary images // Machine Learning and Data Analysis. 2014. Т. 1. № 7. С. 817-827. URL
    • Zhukova K.V., Reyer I.A. Basic Skeleton Connectivity and Parametric Shape Descriptor // Machine Learning and Data Analysis.2014. Т. 1. № 10. С. 1354-1368. URL
    • Kushnir O., Seredin O. Shape Matching Based on Skeletonization and Alignment of Primitive Chains // Communications in Computer and Information Science. 2015. V. 542. P. 123-136. URL
  • Basic algorithm: Convolution network for bitmap.
  • Solution: It is required to propose a method for collapsing graph structures, which allows generating an informative description of the thick line skeleton.
  • Novelty: A method is proposed for improving the quality of recognition of thick lines due to a new method for generating their descriptions.
  • Authors: Experts Reyer I.A., Strizhov V.V., Mark Potanin, consultant Denis Ozherelkov

Task 60

  • Name: Variational optimization of deep learning models with model complexity control
  • Task: The task of optimizing a deep learning model with a predetermined model complexity is considered. It is required to propose a model optimization method that allows generating new models with a given complexity and low computational costs.
  • Data:MNIST, CIFAR
  • References:
  • Basic algorithm: Random search
  • Solution: The proposed method is to represent a deep learning model as a hypernet (a network that generates the parameters of another network) using a Bayesian approach. Probabilistic assumptions about the parameters of deep learning models are introduced, and a variational lower estimate of the Bayesian validity of the model is maximized. The variation estimate is considered as a conditional value depending on the external parameter of complexity.
  • Novelty: The proposed method allows generating models in one-shot mode (practically without retraining) with the required model complexity, which significantly reduces the cost of optimization and retraining.
  • Authors: Oleg Bakhteev, Strizhov V.V.

Task 61

  • Name: Selecting a deep learning model based on the triplet relationship of model and sample
  • Task: Task one-shot of choosing a deep learning model is considered: choosing a model for a specific sample, issued from some general population, should not be computationally expensive.
  • Data:MNIST, synthetic data
  • References:
  • Basic algorithm: Random search
  • Solution: It is proposed to consider the space of parameters and models as two domains with their own generative models. To obtain a connection between domains, a generalization of the variational derivation to the case of triplet constraints is used.
  • Novelty: New one-shot model training method
  • Authors: Oleg Bakhteev, Strizhov V.V.

Task 64

  • Name: Theoretical validity of the application of metric classification methods using dynamic alignment (DTW) to spatiotemporal objects.
  • Task: It is necessary to study the existing theoretical justifications for applying dynamic alignment methods to various objects, and explore the use of such methods for space-time series.
    When proving the applicability of alignment methods, it is proved that the function generated by the dynamic alignment algorithm is the core. Which, in turn, justifies the use of metric classification methods.
  • References:
  • Solution: For different formulations of the DTW method (when the internal function of the distance between time series samples is different) - find and collect evidence that the function is the kernel in one place.
    For a basic set of datasets with time series (on which the accuracy of distance functions is checked ) check the fulfillment of the conditions from the Mercer theorem (positive definiteness of the matrix). Do this for various modifications of the DTW distance function. (Sakoe-Chiba band, Itakura band, weighted DTW.)
  • Novelty: Investigation of theoretical justifications for applying the dynamic alignment algorithm (DTW) and its modifications to space-time series.
  • Authors: Strizhov V.V. - Expert, Gleb Morgachev, Alexey Goncharov - consultants.

Task 66

  • Name: Agnostic neural networks
  • Task: Introduce a metric space into the problem of automatic construction (selection) of agnostic networks.
  • Data: Data from the Reinforcement learning area. Preferably the type of cars on the track.
  • References::
  • Basic algorithm: Networks from an archived article. Symbolic regression from an article in ESwA (you need to restore the code).
  • Solution: We create a model generator in the framework of symbolic regression. We create a model generator as a variational autoencoder (we won’t have time during the course). We study the metric properties of sample spaces (Euclidean) and models (Banach). We create a GAN pair - a generator-discriminator for predicting the structures of predictive models.
  • Novelty: So far, no one has succeeded. Here they discussed Tommi Yaakkola, how he came to us in Yandex. He hasn't succeeded yet either.
  • Authors: Expert Strizhov V.V., Radoslav Neichev - consultant

Task 13

  • Name: Deep learning for RNA secondary structure prediction
  • Task: RNA secondary structure is an important feature which defines RNA functional properties. Its importance can be illustrated by the fact, that it is evolutionary preserved and some types of functional RNAs always * have the same secondary structure, for example all tRNAs fold into cloverleaf. As secondary structure often defines functions, knowing RNAs secondary structure may help investigate functions of novel RNA molecules. RNA folding is not as easy as DNA folding, because RNA is single stranded molecule which forms complicated base-pairing interactions, while DNA mostly exists as fully base paired double helices. Current methods of RNA structure prediction rely on experimentally evaluated thermodynamic rules, but with thermodynamics alone only 80% of structures can be accurately predicted. We propose an AI-driven method for predicting RNA secondary structure inspired by neural machine translation model.
  • Data: RNA sequences in form of strings of characters
  • References:: https://arxiv.org/abs/1609.08144
  • Basic algorithm: https://www.ncbi.nlm.nih.gov/pubmed/16873527
  • Solution: Deep learning recurrent encoder-decoder model with attention
  • Novelty: Currently RNA secondary structure prediction still remains unsolved problem and to the best of our knowledge DL approach has never been introduced in the literature before
  • Authors: consultant Maria Popova, Alexander Isaev (we are waiting for a response from them, without a response task is removed)

Task 65

  • Name: Approximation of low-dimensional samples by heterogeneous models
  • Task: The problem of knowledge transfer (Hinton's distillation, Vapnik's privileged learning) from one network to another is investigated.
  • Data: UCI samples, see what samples are used in papers on this topic
  • References::
  • Basic algorithm: described in the work of Neichev
  • Novelty: Exploring different sampling methods
  • Solution:Try different models that are in the lectures, from non-parametric to deep ones, compare and visualize the likelihood functions
  • Authors: consultants Mark Potanin, (ask Andrey Grabovoi for help) Strizhov V.V.

Task 67

  • Name: Selection of topics in topic models for exploratory information retrieval.
  • Task: Test the hypothesis that when searching for similar documents by their topic vectors, not all topics are informative, so discarding some topics can increase the accuracy and completeness of the search. Consider the alternative hypothesis that instead of discarding topics, one can compare vectors by a weighted cosine proximity measure with adjustable weights.
  • Data: Text collections of sites habr.com and techcrunch.com. Labeled selections: queries and related documents.
  • References::
    1. Vorontsov K. V. Probabilistic Topic Modeling: An Overview of Models and Additive Regularization.
    2. Ianina A., Vorontsov K. Regularized Multimodal Hierarchical Topic Model for Document-by-Document Exploratory Search // FRUCT ISMW, 2019.
  • Basic algorithm: The topic model with regularizers and modalities described in the article (source code available).
  • Novelty:The question of informativeness of topics for vector search of thematically related documents has not been studied before.
  • Solution: Evaluate the individual informativeness of topics by throwing them out one at a time; then sort the topics by individual informativeness and determine the threshold for cutting off non-informative topics. A suggestion as to why this should work: background themes are not informative, and discarding them increases search accuracy and recall by a few percent.
  • Authors: Vorontsov K. V.в, consultant Anastasia Yanina.

Task 68

  • Name: Meta-learning of topic classification models.
  • Task: Develop universal heuristics for a priori assignment of modality weights in thematic models of text classification.
  • Data: Description of datasets, Folder with datasets.
  • References::
    1. Vorontsov K. V. Probabilistic Topic Modeling: An Overview of Models and Additive Regularization.
  • Basic algorithm: Thematic classification models for several datasets.
  • Novelty:In topic modeling, the problem of automatic selection of modality weights has not yet been solved.
  • Solution: Optimize the weights of modalities according to the quality criterion of text classification. Investigate the dependence of the optimal relative weights of modalities on the dimensional characteristics of the problem. Find formulas for estimating the initial values of modality weights without explicitly solving the problem. To reproduce datasets, apply sampling of fragments of source documents.
  • Authors: Vorontsov K. V., consultant Yulian Serdyuk.

Task 70

  • Name: Investigation of the structure of the target space when building a predictive model
  • Task:The problem of forecasting a complex target variable is studied. Complexity means the presence of dependencies (linear or non-linear). It is assumed that the initial data are heterogeneous: the spaces of the independent and target variables are of different nature. It is required to build a predictive model that would take into account the dependence in the source space of the independent variable, as well as in the space of the target variable.
  • Data: Heterogeneous data: picture - text, picture - speech and so on.
  • Basic algorithm: As basic algorithms, it is proposed to use a linear model, as well as a nonlinear neural network model.
  • Authors: Strizhov V.V. - Expert, consultant: Isachenko Roman.

Task 71

  • Name: Investigation of ways to match models by reducing the dimension of space
  • Task: The task of predicting a complex target variable is investigated. Complexity means the presence of dependencies (linear or non-linear). It is proposed to study ways to take into account dependencies in the space of the target variable, as well as the conditions under which these dependencies affect the quality of the final predictive model.
  • Data: Synthetic data with known data generation hypothesis.
  • Basic algorithm: As basic algorithms, it is proposed to use space dimensionality reduction methods (PCA, PLS, autoencoder) and linear matching models.
  • Authors: Strizhov V.V. - Expert, consultant: Isachenko Roman.

Task 72

  • Name: Construction of a single latent space in the problem of modeling heterogeneous data.
  • Task: The task of predicting a complex target variable is investigated. Complexity means the presence of dependencies (linear or non-linear). It is proposed to build a single latent space for the independent and target variables. Model matching is proposed to be carried out in the resulting low-dimensional space.
  • Data: Heterogeneous data: picture - text, picture - speech and so on.
  • Basic algorithm: As basic algorithms, it is proposed to use space dimensionality reduction methods (PCA, PLS, autoencoder) and linear matching models.
  • Authors: Strizhov V.V. - Expert, consultant: Isachenko Roman.

Task 73

  • Name: Nonlinear ranking of exploratory information search results.
  • Task: Develop an algorithm for recommending the reading order of documents (reading order, reading list) found using exploratory information retrieval. Documents should be ranked from simple to complex, from general to specific, that is, in the order in which it will be easier for the user to understand a new subject area for him. The algorithm must build a reading graph - a partial order relation on the set of found documents; in particular, it can be a collection of trees (document forest).
  • Data: Part of Wikipedia and reference reading graph derived from Wikipedia categories.
  • References::
    1. Vorontsov K. V. Probabilistic Topic Modeling: An Overview of Models and Additive Regularization.
    2. Georgia Koutrika, Lei Liu, and Steven Simske. Generating reading orders over document collections. HP Laboratories, 2014.
    3. James G. Jardine. Automatically generating reading lists. Cambridge, 2014.
  • Basic algorithm: described in the article G.Koutrika.
  • Novelty: Task has been little studied in the literature. Regularized multimodal topic models (ARTM, BigARTM) have never been applied to this problem.
  • Solution: The use of ARTM topic models in conjunction with estimates of the cognitive complexity of the text.
  • Authors: Vorontsov K. V., consultant Maxim Eremeev.

2019

Author Topic Links Consultant Reviewer
Severilov Pavel Task of searching characters in texts LinkReview

code paper slides video

Murat Apishev
Grigoriev Alexey Text recognition based on skeletal representation of thick lines and convolutional networks LinkReview

code, paper, slides video

Ilya Zharikov review Varenyk Natalia
Grishanov Alexey Automatic configuration of BigARTM parameters for a wide class of tasks LinkReview code, paperslides

video

Viktor Bulatov review Gerasimenko Nikolay
Yusupov Igor Dynamic alignment of multivariate time series LinkReview code paper slides video Alexey Goncharov
Varenyk Natalia Spherical CNN for QSAR prediction LinkReview, code, paper, slides video Maria Popova review Grigoriev Alexey
Beznosikov Alexander Z-learning of linearly-solvable Markov Decision Processes LinkReview

paper code slides video

Yury Maximov
Panchenko Svyatoslav Obtaining a simple sample at the output of the neural network layer LinkReview,

code, paper, slides

Gadaev Tamaz
Veselova Evgeniya Deep Learning for reliable detection of tandem repeats in 3D protein structures Code link review paper slides video Guillaume Pages, Sergei Grudinin
Aminov Timur Quality Prediction for a Feature Selection Procedure LinkReview code paper

slides

Roman Isachenko
Markin Valery Investigation of the properties of local models in the spatial decoding of brain signals LinkReview

code paper slides video

Roman Isachenko
Abdurahmon Sadiev Generation of features using locally approximating models LinkReview

code, paper, slides video

Anastasia Motrenko
Tagir Sattarov Machine translation training without parallel texts. LinkReview code paper, slides video Oleg Bakhteev
Gerasimenko Nikolay Thematic search for similar cases in the collection of acts of arbitration courts. LinkReview code paper slides video Ekaterina Artyomova reviewGrishanov Alexey

Task 40

  • Name: Quality prediction for the feature selection procedure.
  • Task: The solution of the feature selection problem is reduced to enumeration of binary cube vertices. This procedure cannot be performed for a sample with a large number of features. It is proposed to reduce this problem to optimization in a linear space.
  • Data: Synthetic data + simple samples
  • References::
    1. Bertsimas D. et al. Best subset selection via a modern optimization lens //The annals of statistics. – 2016. – Т. 44. – №. 2. – С. 813-852.
    2. Luo R. et al. Neural architecture optimization //Advances in Neural Information Processing Systems. – 2018. – С. 7827-7838.
  • Basic algorithm: Popular feature selection methods.
  • Solution: In this paper, it is proposed to build a model that, based on a set of features, predicts the quality on a test sample. To do this, a mapping of a binary cube into a linear space is constructed. After that, the quality of the model in linear space is maximized. To reconstruct the solution of the problem, the model of inverse mapping into a binary cube is used.
  • Novelty: A constructively new approach to solving the problem of choosing models is proposed.
  • Authors: Strizhov V.V., Tetiana Aksenova, consultant – Roman Isachenko

Task 42

  • Name: Z-learning of linearly-solvable Markov Decision Processes
  • Task: Adapt Z-learning from [1] to the case of Markov Decision Process discussed in [2] in the context of energy systems. Compare it with standard (in reinforcement learning) Q-learning.
  • Data: We consider a Markov Process described via transition probability matrix. Given initial state vector (probability of being in a state at time zero), we generate data for the time evolution of the state vector. See [2] for an exemplary process describing evolution of an ensemble of energy consumers.
  • References::
    1. E. Todorov. Linearly-solvable Markov decision problems https://homes.cs.washington.edu/~todorov/papers/TodorovNIPS06.pdf
    2. Ensemble Control of Cycling Energy Loads: Markov Decision Approach. Michael Chertkov, Vladimir Y. Chernyak, Deepjyoti Deka. https://arxiv.org/abs/1701.04941
    3. Csaba Szepesvári. Algorithms for Reinforcement Learning. https://sites.ualberta.ca/~szepesva/papers/RLAlgsInMDPs.pdf
  • Basic algorithm: Principal comparison should be made with Q learning described in [3]
  • Solution: We suppose that plugging in algorithm from [1] directly into [2] gives faster and more reliable solution.
  • Novelty: In the area of power systems there is a huge demand on fast reinforcement learning algorithms, but there is still a lack of that (in particular the ones respect the physics/underlying graph)
  • Authors: Yury Maximov (consultant, expert), Michael Chertkov (expert)

Task 1

  • Name: Forecasting the direction of movement of the price of exchange instruments according to the news flow.
  • Task: Build and explore a model for predicting the direction of price movement. Given a set of news S and a set of timestamps T corresponding to the time of publication of news from S. 2. Time series P, corresponding to the price of an exchange instrument, and time series V, corresponding to the volume of sales for this instrument, for a period of time T'. 3. The set T is a subset of the time period T'. 4. Time intervals w=[w0, w1], l=[l0, l1], d=[d0, d1], where w0 < w1=l0 < l1=d0 < d1. It is required to predict the direction of movement of the price of an exchange instrument at the time t=d0 according to the news released in the period w.
  • Data:
    1. Financial data: data on quotes (at one tick interval) of several financial instruments (GAZP, SBER, VTBR, LKOH) for the 2nd quarter of 2017 from the Finam.ru website; for each point of the series, the date, time, price and volume are known.
    2. Text data: economic news for the 2nd quarter of 2017 from Forexis; each news is a separate html file.
  • References:
    1. Usmanova K.R., Kudiyarov S.P., Martyshkin R.V., Zamkovoy A.A., Strijov V.V. Analysis of relationships between indicators in forecasting cargo transportation // Systems and Means of Informatics, 2018, 28(3).
    2. Kuznetsov M.P., Motrenko A.P., Kuznetsova M.V., Strijov V.V. Methods for intrinsic plagiarism detection and author diarization // Working Notes of CLEF, 2016, 1609 : 912-919.
    3. Aysina Roza Munerovna, Thematic modeling of financial flows of corporate clients of a bank based on transactional data, final qualification work.
    4. Lee, Heeyoung, et al. "On the Importance of Text Analysis for Stock Price Prediction." LREC. 2014.
  • Basic algorithm: Method used in the article (4).
  • Solution: Using topic modeling (ARTM) and local approximation models to translate a sequence of texts corresponding to different timestamps into a single feature description. Quality criterion: F1-score, ROC AUC, profitability of the strategy used.
  • Novelty: To substantiate the connection of time series, the Converging cross-mapping method is proposed.
  • Authors: Ivan Zaputlyaev (consultant), Strizhov V.V., K.V. Vorontsov (Experts)

Task 3

  • Name: Dynamic alignment of multidimensional time series.
  • Task: A characteristic multidimensional time series is the trajectory of a point in 3-dimensional space. The two trajectories need to be optimally aligned with each other. For this, the distance DTW between two time series is used. In the classical representation, DTW is built between one-dimensional time series. It is necessary to introduce various modifications of the algorithm for working with high-dimensional time series: trajectories, corticograms.
  • Data: The data describes 6 classes of time series from the mobile phone's accelerometer. https://sourceforge.net/p/mlalgorithms/code/HEAD/tree/Group274/Goncharov2015MetricClassification/data/
  • References:
    1. Multidimensional DTW: https://pdfs.semanticscholar.org/76d3/5bd5a52453ebde80faaa1467d7effd74426f.pdf
  • Basic algorithm: Using L_p distances between two dimensions of a time series, their modifications.
  • Solution: Investigation of distances resistant to change of coordinate order, studies of distances unstable to change of coordinate order. Experiments with other types of distances (cosine, RBF, others).
  • Novelty: There is no complete review and study of methods for working with multivariate time series. The dependence of the quality of the solution on the selected distances between measurements has not been studied.
  • Authors: Alexey Goncharov - consultant, Expert, Strizhov V.V. - Expert

Task 43

  • Name: Getting a simple sample at the output of the neural network layer
  • Task: The output of the neural network is usually a generalized linear model over the outputs of the penultimate layer. It is necessary to propose a way to test the simplicity of the sample and its compliance with the generalized linear model (linear regression, logistic regression) using a system of statistical criteria.
  • Data: For the computational experiment, it is proposed to use classical samples from the UCI repository. Link to samples https://github.com/ttgadaev/SampleSize/tree/master/datasets
  • References:: http://www.ccas.ru/avtorefe/0016d.pdf c 49-63 Bishop, C. 2006. Pattern Recognition and Machine Learning. Berlin: Springer. $758
  • Basic algorithm: White test, Wald test, Goldfeld-Quantum test, Durbin-Watson, Chi-square, Fry-Behr, Shapiro-Wilk
  • Solution: The system of tests for checking the simplicity of the sample (and the adequacy of the model), the independent variables are not random, the dependent variables are distributed normally or binomially, there are no gaps and outliers, the classes are balanced, the sample is approximated by a single model. The variance of the error function does not depend on the independent variable. The study is based on synthetic and real data.
  • Authors: Gadaev T. T. (consultant) Strizhov V.V., Grabovoi A.V. (Experts)

Task 14

  • Name: Deep Learning for reliable detection of tandem repeats in 3D protein structures more in PDF
  • Task: Deep learning algorithms pushed computer vision to a level of accuracy comparable or higher than a human vision. Similarly, we believe that it is possible to recognize the symmetry of a 3D object with a very high reliability, when the object is represented as a density map. The optimization problem includes i) multiclass classification of 3D data. The output is the order of symmetry. The number of classes is ~10-20 ii) multioutput regression of 3D data. The output is the symmetry axis (a 3-vector). The input data are typically 24x24x24 meshes. The total amount of these meshes is of order a million. Biological motivation : Symmetry is an important feature of protein tertiary and quaternary structures that has been associated with protein folding, function, evolution, and stability. Its emergence and ensuing prevalence has been attributed to gene duplications, fusion events, and subsequent evolutionary drift in sequence. Methods to detect these symmetries exist, either based on the structure or the sequence of the proteins, however, we believe that they can be vastly improved.
  • Data: Synthetic data are obtained by ‘symmetrizing’ folds from top8000 library (http://kinemage.biochem.duke.edu/databases/top8000.php).
  • References:: Our previous 3D CNN: [33] Invariance of CNNs (and references therein): 01630265/document, [34]
  • Basic algorithm: A prototype has already been created using the Tensorflow framework [4], which is capable of detecting the order of cyclic structures with about 93% accuracy. The main goal of this internship is to optimize the topology of the current neural network prototype and make it rotational and translational invariant with respect to input data. [4] [35]
  • Solution: The network architecture needs to be modified according to the invariance properties (most importantly, rotational invariance). Please see the links below [36], [37] The code is written using the Tensorflow library, and the current model is trained on a single GPU (Nvidia Quadro 4000)of a desktop machine.
  • Novelty: Applications of convolutional networks to 3D data are still very challenging due to large amount of data and specific requirements to the network architecture. More specifically, the models need to be rotationally and transnationally invariant, which makes classical 2D augmentation tricks loosely applicable here. Thus, new models need to be developed for 3D data.
  • Authors: Expert Sergei Grudinin, consultants Guillaume Pages

Task 46

  • Name: Task of searching characters in texts
  • Task: In the simplest case, this Task is reduced to the Sequence Labeling task on a labeled selection. The difficulty lies in obtaining a sufficient amount of training data, that is, it is required to obtain a larger sample from the existing small Expert markup (automatically by searching for patterns or by compiling a simple and high-quality markup instruction, for example, in Toloka). The presence of markup allows you to start experimenting with the selection of the optimal model, various neural network architectures (BiLSTM, Transformer, etc.) may be of interest here.
  • Data: Dictionary of symbols, Marked artistic texts
  • References: http://www.machinelearning.ru/wiki/images/0/05/Mmta18-rnn.pdf
  • Basic algorithm: HMM, RNN
  • Solution: It is proposed to compare the work of several state-of-the-art algorithms. Propose a classifier quality metric for characters (character/non-character). Determine applicability of methods.
  • Novelty: The proposed approach to text analysis is used by Experts in manual mode and has not been automated
  • Authors: M. Apishev (consultant), D. Lemtyuzhnikova

Task 47

  • Name: Deep learning for RNA secondary structure prediction
  • Task: RNA secondary structure is an important feature which defines RNA functional properties. Its importance can be illustrated by the fact, that it is evolutionary preserved and some types of functional RNAs always * have the same secondary structure, for example all tRNAs fold into cloverleaf. As secondary structure often defines functions, knowing RNAs secondary structure may help investigate functions of novel RNA molecules. RNA folding is not as easy as DNA folding, because RNA is single stranded molecule which forms complicated base-pairing interactions, while DNA mostly exists as fully base paired double helices. Current methods of RNA structure prediction rely on experimentally evaluated thermodynamic rules, but with thermodynamics alone only 80% of structures can be accurately predicted. We propose an AI-driven method for predicting RNA secondary structure inspired by neural machine translation model.
  • Data: RNA sequences in form of strings of characters
  • References:: https://arxiv.org/abs/1609.08144
  • Basic algorithm: https://www.ncbi.nlm.nih.gov/pubmed/16873527
  • Solution: Deep learning recurrent encoder-decoder model with attention
  • Novelty: Currently RNA secondary structure prediction still remains unsolved problem and to the best of our knowledge DL approach has never been introduced in the literature before
  • Authors: consultant Maria Popova Chapel-Hill

Task 4

  • Name: Automatic setting of ARTM parameters for a wide class of tasks.
  • Task: The bigARTM open library allows you to build topical models using a wide class of possible regularizers. However, this flexibility makes the task of setting the coefficients very difficult. This tuning can be greatly simplified by using the relative regularization coefficients mechanism and automatic selection of N-grams. We need to test the hypothesis that there is a universal set of relative regularization coefficients that gives "reasonably good" results on a wide class of problems. Several datasets are given with some external quality criterion (for example, classification of documents into categories or ranking). We find the best parameters for a particular dataset, giving the "locally the best model". We find the bigARTM initialization algorithm that produces thematic models with quality comparable to the "locally best model" on its dataset. Comparability criterion in quality: on this dataset, the quality of the "universal model" is no more than 5% worse than that of the "locally best model".
  • Data: Victorian Era Authorship Attribution Data Set, uci.edu/ml/datasets/Twenty+Newsgroups 20 Newsgroups, ICD-10, search/ranking triplets.
  • References:
    1. WRC by Nikita Doykov: http://www.machinelearning.ru/wiki/images/9/9f/2015_417_DoykovNV.pdf
    2. Presentation by Viktor Bulatov at a scientific seminar: https://drive.google.com/file/d/19pJ21LRPeeOxY4mkcSnQCRm93zOO4J5b/view
    3. Draft with formulas: https://drive.google.com/open?id=1AqS7snUsSJ18ZYBtC-6uP_2dMTDJSGeD
  • Basic algorithm: PLSA / LDA / logregression.
  • Solution: bigARTM with background themes and smoothing, sparseness and decorrelation regularizers (coefficients picked up automatically), as well as automatically selected N-grams.
  • Novelty: The need for automated tuning of model parameters and the lack of such implementations in the scientific community.
  • Authors: consultant Viktor Bulatov, ExpertVorontsov K. V..

Task 50

  • Name: Thematic search for similar cases in the collection of acts of arbitration courts.
  • Task: Build an information retrieval algorithm for a collection of acts of arbitration courts. The request can be an arbitrary document of the collection (the text of the act). The search result should be a list of documents in the collection, ranked in descending order of relevance.
  • Data: collection of text documents — acts of arbitration courts http://kad.arbitr.ru.
  • References:
    1. Anastasia Yanina. Thematic exploratory information search. 2018. FIVT MIPT.
    2. Ianina A., Golitsyn L., Vorontsov K. Multi-objective topic modeling for exploratory search in tech news. AINL-2017. CCIS, Springer, 2018.
    3. Ahmed El-Kishky, Yanglei Song, Chi Wang, Clare Voss, Jiawei Han. Scalable Topical Phrase Mining from Text Corpora. 2015.
  • Basic algorithm: BigARTM with decorrelation, smoothing, sparse regularizers. Search by TF-IDF of words, by TF-IDF of UPA links, by thematic vector representations of documents, using a cosine proximity measure. TopMine algorithm for collocation detection.
  • Solution: Add modality of links to legal acts. Add modality of legal terms. Choose the optimal number of topics and regularization strategy. Organize the process of marking pairs of documents. Implement the evaluation of the quality of the search for a labeled sample of pairs of documents.
  • Novelty: The first attempt to use ARTM for thematic search of legal texts.
  • Authors: consultant Ekaterina Artyomova, Expert Vorontsov K. V..

Group 2

Author Topic Links Consultant Reviewer
Vishnyakova Nina Optimal Approximation of Non-linear Power Flow Problem LinkReview paper code presentation video Yury Maximov reviewer Loginov Roman

review

Kudryavtseva Polina Intention forecasting. Building an optimal signal decoding model for modeling a brain-computer interface. code

LinkReview paper video presentation

Roman Isachenko Nechepurenko Ivan

review

Loginov Roman Multi-simulation as a universal way to describe a general sample code

LinkReview paper ChatInvite presentation video

Aduenko A. A. Макаров Михаил review
Mikhail Makarov Location determination by accelerometer signals code

LinkReview paper presentation video

Anastasia Motrenko Cherepkov Anton: review
Kozinov Alexeyй Task of finding characters in images LinkReview

paper code

M. Apishev,

D. Lemtyuzhnikova

Gracheva Anastasia review
Buchnev Valentin Early prediction of sufficient sample size for a generalized linear model. LinkReview

paper code presentation video

Grabovoi A.V. reviewer
Nechepurenko Ivan Multisimulation, privileged training code,

paper, LinkReview presentation

R. G. Neichev Kudryavtseva Polina
Gracheva Anastasia Estimation of binding energy of protein and small molecules code

paper LinkReview presentation video

Sergei Grudinin,

Maria Kadukova

reviewer
Cherepkov Anton Privileged learning in the problem of iris boundary approximation paper, slides, code, LinkReview

video

R. G. Neichev Lepekhin Mikhail

preliminary review

Lepekhin Mikhail Creation of ranking models for information retrieval systems. Algorithm for Predicting the Structure of Locally Optimal Models code

LinkReview paper presentation video

Andrey Kulunchakov Vishnyakova Nina, review
Gridasov Ilya Automatic construction of a neural network of optimal complexity LinkReview

paper Presentation code

O. Yu. Bakhteev, Strizhov V.V. Buchnev Valentin
Telenkov Dmitry Brain signal decoding and intention prediction LinkReview

git The paper Presentation code

Andrey Zadayanchuk reviewer

Task 18

  • Name: Forecasting intentions. Building an optimal signal decoding model for modeling a brain-computer interface.
  • Task: The Brain Computer Interface (BCI) allows you to help people with disabilities regain their mobility. According to the available description of the device signal, it is necessary to simulate the behavior of the subject.
  • Data: Data sets of ECoG/EEG brain signals.
  • References::
    • Motrenko A.P., Strijov V.V. Multi-way feature selection for ECoG-based brain-computer Interface // Expert systems with applications. - 2018.
  • Basic algorithm: It is proposed to compare with the partial least squares algorithm.
  • Solution: In this work, it is proposed to build a single system that solves the problem of signal decoding. As stages of building such a system, it is proposed to solve the problems of data preprocessing, feature space extraction, dimensionality reduction and selection of a model of optimal complexity. It is proposed to use the tensor version of PLS with feature selection.
  • Novelty: In the formulation of the problem, the complex nature of the signal is taken into account: a continuous trajectory of movement, the presence of discrete structural variables (fingers or joint movement), the presence of continuous variables (position of a finger or limb).
  • Authors: Strizhov V.V., Tetiana Aksenova, consultant – Roman Isachenko

Task 41

  • Name: Optimal Approximation of Non-linear Power Flow Problem
  • Task: Our goal is to approximate the solution of non-linear non-convex optimal power flow problem by solving a sequence of convex optimization problems (aka trust region approach). On this way we propose to compare various approaches for an approximate solution of this problem with adaptive approximation of the power flow non-linearities with a sequence of quadratic and/or piece-wise linear functions
  • Data: Matpower module from MATLAB contains all necessary test cases. Start considering IEEE 57 bus case.
  • References::
    1. Molzahn, D. K., & Hiskens, I. A. (2019). A survey of relaxations and approximations of the power flow equations. Foundations and Trends in Electric Energy Systems, 4(1-2), 1-221. https://www.nowpublishers.com/article/DownloadSummary/EES-012
    2. The QC Relaxation: A Theoretical and Computational Study on Optimal Power Flow. Carleton Coffrin ; Hassan L. Hijazi; Pascal Van Hentenryck https://ieeexplore.ieee.org/abstract/document/7271127/
    3. Convex Relaxations in Power System Optimization: A Brief Introduction. Carleton Coffrin and Line Roald. https://arxiv.org/pdf/1807.07227.pdf
    4. Optimal Adaptive Linearizations of the AC Power Flow Equations. Sidhant Misra, Daniel K. Molzahn, and Krishnamurthy Dvijotham https://molzahn.github.io/pubs/misra_molzahn_dvijotham-adaptive_linearizations2018.pdf
  • Basic algorithm: A set of algorithms described in [1] should be considered to compare with, details behind the proposed method would be shared by the consultant (a draft of the paper)
  • Solution: to figure out the quality of the solution we propose to compare it with the ones given by IPOPT and numerous relaxations, and do some reverse engineering regarding to our method
  • Novelty: The OPF is a truly hot topic in power systems, and is of higher interest by the discrete optimization community (as a general QCQP problem). Any advance in this area is of higher interest by the community
  • Authors: Yury Maximov (consultant and expert), Michael Chertkov (expert)
  • Notes: the problem has both the computational and the theoretical focuses, so 2 students are ok to work on this topic

Task 2

  • Name: Investigation of reference objects in the problem of metric classification of time series.
  • Task: The DTW function is the distance between two time series that can be non-linearly warped relative to each other. It looks for the best alignment between two objects, so it can be used in a metric object classification problem. One of the methods for solving the problem of metric classification is measuring distances to reference objects and using the vector of these distances as an indicative description of the object. The DBA method is an algorithm for constructing centroids (reference objects) for time series based on the DTW distance. When plotting the distance between the time series and the centroid, different pairs of values (eg peak values) are more specific to one of the classes, and the impact of such coincidences on the distance value should be higher.

It is necessary to explore various ways of constructing reference objects, as well as determining their optimal number. The criterion is the quality of the metric classifier in the task. In the DBA method, for each centroid, it is proposed to create a weight vector that demonstrates the "significance" of the measurements of the centroid, and use it in the modified weighted-DTW distance function.

Literature research and a combination of up-to-date methods.

  • Novelty: There has not been a comprehensive study of various methods of constructing centroids and reference elements along with the choice of their optimal number.
  • Authors: Alexey Goncharov - consultant, Expert, Strizhov V.V. - Expert

Task 7

  • Name: Privileged learning in the iris boundary approximation problem
  • Task: Based on the image of the human eye, determine the circles approximating the inner and outer border of the iris.
  • Data: Bitmap monochrome images, typical size 640*480 pixels (however other sizes are possible)[38], [39].
  • References::
    • Aduenko A.A. Selection of multi-models in Tasks classification (supervisor Strizhov V.V.). Moscow Institute of Physics and Technology, 2017. [40]
    • K.A. Gankin, A.N. Gneushev, I.A. Matveev Segmentation of the iris image based on approximate methods with subsequent refinements // Izvestiya RAN. Theory and control systems, 2014, no. 2, p. 78–92.
    • Duda, R. O. Use of the Hough transformation to detect lines and curves in pictures / R. O. Duda, P. E. Hart // Communications of the ACM. 1972 Vol. 15, no. 1.Pp.
  • Basic algorithm: Efimov Yury. Search for the outer and inner boundaries of the iris in the eye image using the paired gradient method, 2015.
  • Solution: See iris_circle_problem.pdf
  • Novelty: A fast non-enumerative algorithm for approximating boundaries using linear multimodels is proposed. Additionally, capsule neural networks.
  • consultant: Radoslav Neichev (by Strizhov V.V., Expert Matveev I.A.)

Task 44

  • Name: Early prediction of sufficient sample size for a generalized linear model.
  • Task: The problem of designing an experiment is being investigated. The Task of estimating a sufficient sample size according to the data is solved. The sample is assumed to be simple. It is described by an adequate model. Otherwise, the sample is generated by a fixed probabilistic model from a known class of models. The sample size is considered sufficient if the model is restored with sufficient confidence. It is required, knowing the model, to estimate a sufficient sample size at the early stages of data collection.
  • Data: For the computational experiment, it is proposed to use classical samples from the UCI repository. Link to samples https://github.com/ttgadaev/SampleSize/tree/master/datasets
  • References::
    1. [Overview of methods for estimating sample size]
    2. http://svn.code.sf.net/p/mlalgorithms/code/PhDThesis/.
    3. Bootstrap method. https://projecteuclid.org/download/pdf_1/euclid.aos/1.

Bishop, C. 2006. Pattern Recognition and Machine Learning. Berlin: Springer. $758

  • Basic algorithm: We will say that the sample size is sufficient if the log-likelihood has a small variance, on a sample of size m calculated using bootstrap.

We are trying to approximate the dependence of the average value of log-likelihood and its variance on the sample size.

  • Solution: The methods described in the review are asymptotic or require a deliberately large sample size. The new method should be to predict volume in the early stages of experiment design, i.e. when data is scarce.
  • Authors: Grabovoi A.V. (consultant), Gadaev T. T. Strizhov V.V. (Experts)
  • Note: to determine the simplicity of the sample, a new definition of complexity is proposed (Sergey Ivanychev). This is a separate work, +1 Task 44a (? Katruza).

Task 15

  • Name: Formulation and solution of an optimization problem combining classification and regression to estimate the binding energy of a protein and small molecules. Task description [41]
  • Task: From a bioinformatics point of view, the Task is to estimate the free energy of protein binding to a small molecule (ligand): the best ligand in its best position has the lowest free energy of interaction with the protein. (Following a large text, see the file at the link above.)
  • Data:
    • Data for binary classification. Approximately 12,000 protein-ligand complexes: for each of them there is 1 native position and 18 non-native ones. The main descriptors are histograms of distributions of distances between different atoms of the protein and ligand, the dimension of the vector of descriptors is ~ 20,000. In the case of continued research and publication in a specialized journal, the set of descriptors can be expanded. The data will be provided as binary files with a python script to read.
    • Data for regression. For each of the presented complexes, the value of the quantity is known, which can be interpreted as the binding energy.
  • References::
  • Basic algorithm: [45] In the classification problem, we used an algorithm similar to linear SVM, whose relationship with the energy estimate is beyond the scope of the classification problem, described in the above article. Various loss functions can be used in a regression problem.
  • Solution: It is necessary to connect the previously used optimization problem with the regression problem and solve it using standard methods. Cross-validation will be used to check the operation of the algorithm. There is a separate test set consisting of (1) 195 complexes of proteins and ligands, for which it is necessary to find the best ligand pose (the algorithm for obtaining ligand positions differs from that used in training), (2) complexes of proteins and ligands, for which native poses it is necessary to predict the energy binding, and (3) 65 proteins for which the most strongly binding ligand is to be found.
  • Novelty: First of all, the interest is combining classification and regression problems. The correct assessment of the quality of protein and ligand binding is used in drug development to search for molecules that interact most strongly with the protein under study. Using the classification problem described above to predict the binding energy results in an insufficiently high correlation of predictions with experimental values, while using the regression problem alone leads to overfitting.
  • Authors Sergei Grudinin, Maria Kadukova

Task 27

  • Name: Creation of ranking models for information retrieval systems. Algorithm for Predicting the Structure of Locally Optimal Models
  • Task: It is required to predict a time series using some parametric superposition of algebraic functions. It is proposed not to cost the prognostic model, but to predict it, that is, to predict the structure of the approximating superposition. A class of considered superpositions is introduced, and on the set of such structural descriptions, a search is made for a locally optimal model for the problem under consideration. The task consists in 1) searching for a suitable structural description of the model 2) describing the search algorithm for the structure that will correspond to the optimal model 3) describing the algorithm for inverse construction of the model according to its structural description. For an already existing example of the answer to questions 1-3, see the works of A. A. Varfolomeeva.
  • Data:
    • Collection of text documents TREC (!)
    • A set of time series, which implies the restoration of functional dependencies. It is proposed to first use synthetic data or immediately apply the algorithm to forecasting time series 1) electricity consumption 2) physical activity with subsequent analysis of the resulting structures.
  • References::
    1. (!) Kulunchakov A.S., Strijov V.V. Generation of simple structured Information Retrieval functions by genetic algorithm without stagnation // Expert Systems with Applications, 2017, 85: 221–230.
    2. A. A. Varfolomeeva Selection of features when marking up bibliographic lists using structural learning methods, 2013, [46]
    3. Bin Cao, Ying Li and Jianwei Yin Measuring Similarity between Graphs Based on the Levenshtein Distance, 2012, [47]
  • Basic algorithm: Described in [1]. Developed in the work of the 974 group team. It is proposed to use their code and experiment.
  • Solution: It is proposed to try to repeat the experiment of A. A. Varfolomeeva for a different structural description in order to understand what is happening. The superposition of algebraic functions defines an ortree, on the vertices of which the labels of the corresponding algebraic functions or variables are given. Therefore, the structural description of such a superposition can be its DFS-code. This is a string consisting of vertex labels, written in the order in which the tree is traversed by depth-first search. Knowing the arities of the corresponding algebraic functions, we can restore any such DFS-code in O(n) and get back the superposition of functions. On the set of similar string descriptions, it is proposed to search for the string description that will correspond to the optimal model.
  • Authors: consultant Andrey Kulunchakov (Inria Montbonnot), Expert Strizhov V.V.

Task 26

  • Name: Accelerometer positioning
  • Task: Given initial coordinates, accelerometer signals, additional information (gyroscope, magnetometer signals). Possibly inaccurate map given (Task SLAM)
  • Data: from [1], self-collected data.
  • References::
    1. https://arxiv.org/pdf/1712.09004.pdf
    2. https://ieeexplore.ieee.org/document/1528431
  • Basic algorithm: from [1].
  • Solution: Search for a priori and additional information that improves positioning accuracy.
  • Novelty: Statement of the problem in terms of Projection to Latent Spaces
  • Authors: consultant Anastasia Motrenko, Expert Ilya Gartseev , Strizhov V.V.

Task 45

  • Name: Task of searching characters in images
  • Task: This Task in one of the formulation options can be reduced to two sequential operations: 1) searching for objects in the image and determining their class 2) searching the database for information about the symbolic meaning of the found objects. The main difficulty in solving the problem lies in the search for objects in the image. However, the following classification may also be difficult due to the fact that the image of the object may be incomplete, unusually stylized, and the like.
  • Data: Dictionary of Symbols Museum Sites Image-net
  • References:
    1. http://www.machinelearning.ru/wiki/images/e/e2/IDP18.pdf (p. 116)
    2. http://www.image-net.org
  • Basic algorithm: CNN
  • Solution: It is proposed to compare the work of several state-of-the-art algorithms. Suggest a quality metric for searching and classifying objects. Determine applicability of methods.
  • Novelty: The proposed image analysis approach is used by Experts in manual mode and has not been automated
  • Authors: M. Apishev (consultant), D. Lemtyuzhnikova

Task 28

  • Name: Multi-simulation as a universal way to describe a general sample
  • Task: Build a method for incremental refinement of the multimodel structure when new objects appear. Development and comparison of different algorithms for updating the structure of multimodels. Construction of an optimal scheme for refining the structure of a multimodel depending on the total sample size.
  • Data: At the initial stage of work, synthetic data with a known statistical structure is used. Testing of the developed methods is carried out on real data from the UCI repository.
  • References:
  1. Bishop, Christopher M. "Pattern recognition and machine learning." Springer, New York (2006).
  2. Gelman, Andrew, et al. Bayesian data analysis, 3rd edition. Chapman and Hall/CRC, 2013.
  3. MacKay, David JC. "The evidence framework applied to classification networks." Neural computation 4.5 (1992): 720-736.
  4. Aduenko A. A. "Choice of multimodels in Task classification" Ph.D. thesis
  5. Motrenko, Anastasiya, Strizhov V.V., and Gerhard-Wilhelm Weber. "Sample size determination for logistic regression." Journal of Computational and Applied Mathematics 255 (2014): 743-752.
  • Basic algorithm: Algorithm for constructing adequate multi-models from #4.
  • Solution: Bayesian approach to the problem of choosing models based on validity. Analysis of the properties of validity and its relationship with statistical significance.
  • Novelty: A method is proposed for constructing an optimal scheme for updating the structure of a multimodel when new objects appear. The relationship between validity and statistical significance for some classes of models has been studied.
  • Authors: Strizhov Vadim Viktorovich, Aduenko Alexander Alexandrovich (GMT-5)

Task 11

Task 48

Task 49

  • Name: Brain signal decoding and intention prediction
  • Task: It is required to build a model that restores the movement of the limbs according to the corticogram.
  • Data: neurotycho.org [9] (or fingers)
  • References:
    • Neichev R.G., Katrutsa A.M., Strizhov V.V. Selection of the optimal set of features from a multicorrelated set in the forecasting problem. Zavodskaya Lab. Materials Diagnostics, 2016, 82(3) : 68-74. [10]
    • Isachenko R.V., Strijov V.V. Quadratic Programming Optimization with Feature Selection for Non-linear Models // Lobachevskii Journal of Mathematics, 2018, 39(9) : 1179-1187. article
  • Basic algorithm: Partial Least Squares[11]
  • Solution: Create a feature selection algorithm alternative to PLS and taking into account the non-orthogonal feature interdependence structure.
  • Novelty: A feature selection method is proposed that takes into account the regularities of both the and independent variable and the dependent variable. Bonus: Explore changes in model structure as the nature of the sample changes.
  • Authors: Andrey Zadayanchuk, Strizhov V.V.

2018

Autumn 2018

Number Project name materials Team
0 (Example) Metric classification of time series code,

LinkReview, Discussion

Alexey Goncharov*, Maxim Savinov
1 Forecasting the direction of movement of the price of exchange instruments according to the news flow Code,

LinkReview, Slides, Report

Alexander Borisov,

Drobin Maxim, Govorov Ivan, Mukhitdinova Sofia, Valentin Rodionov, Valentin Akhiyarov

2 Construction of reference objects for a set of multidimensional time series Code

LinkReview

Iskhakov Rishat,

Korepanov Georgy, Solodnev Samirkhanov Danil

3 Dynamic alignment of multivariate time series Code

LinkReview Slides Report

Gleb Morgachev,

Vladislav Smirnov, Tatiana Lipnitskaya

4 Automatic adjustment of ARTM parameters for a wide class of tasks Code,

LinkReview, Presentation

Golubeva Tatiana,

Ivanova Ekaterina, Matveeva Svetlana, Trusov Anton, Tsaritsyn Mikhail, Chernonog Vyacheslav

5 Finding paraphrases Code,

LinkReview

Stas Okrug, Nikita Mokrov

Fedor Kitashov, Polina Proskura, Natalia Basimova, Roman Krasnikov, Akhmedkhan Shabanov

6 On conformational changes of proteins using collective motions in torsion angle space and L1 regularization Code,

LinkReview Presentation

Ryabinina Raisa, Emtsev Daniil
7 Privileged training in the problem of approximating the borders of the iris Code,

LinkReview

Pavel Fedosov, Alexey Gladkov,

Genrikh Kenigsberger, Ivan Korostelev, Nikolay Balakin

8 Generation of features using locally approximating models Code,

LinkReview

Ibrahim Kurashov, Nail Gilmutdinov,

Albert Mulyukov, Valentin Spivak

9 Text recognition based on skeletal representation of thick lines and convolutional networks Code, LiteratureReview, Slides, report Kutsevol Polina

Lukoyanov Artem Korobov Nikita Boyko Alexander Litovchenko Leonid Valukov Alexandr Badrutdinov Kamil Yakushevskiy Nikita Valyukov Nikolay Tushin Kirill


10 Comparison of neural network and continuous-morphological methods in the problem of text detection Code, LinkReview, Discussion, Presentation Gaiduchenko Nikolay

Torlak Artyom Akimov Kirill Mironova Lilia Gonchar Daniel

11 Automatic construction of a neural network of optimal complexity Code, LinkReview, report, slides Nikolai Goryan

Alexander Ulitin Tovkes Artem Taranov Sergey Gubanov Sergey Krinitsky Konstantin Zabaznov Anton Valery Markin

12 Machine translation training without parallel texts. Code,

LinkReview, Report, Slides

Alexander Artemenkov

Angelina Yaroshenko Andrey Stroganov Egor Skidnov Anastasia Borisova Ryabov Fedor Mazurov Mikhail

13 Deep learning for RNA secondary structure prediction Code

Link Review

Dorokhin Semyon

Pastukhov Sergey Pikunov Andrey Nesterova Irina Anna chat

14 Deep Learning for reliable detection of tandem repeats in 3D protein structures Code

Link Review

Veselova Evgeniya
15 Formulation and solution of an optimization problem combining classification and regression to estimate the binding energy of a protein and small molecules Code

Link Review

Merkulova Anastasia

Plumite Elvira Zhiboyedova Anastasia chat

16 Estimation of the optimal sample size for research in medicine Code

Link Review

Artemy Kharatyan,

Mikhail Mikheev, Evgin Alexander, Seppar Alexander, Konoplyov Maxim, Murlatov Stanislav, Makarenko Stepan

17 Intention forecasting. Investigation of the properties of local models in the spatial decoding of brain signals Code,

LinkReview, Presentation

Natalia Bolobolova,

Alina Samokhina, Shiyanov Vadim

18 Intention forecasting. Building an optimal signal decoding model for modeling a brain-computer interface. Code,

LinkReview, Presentation, Article

Ivan Nasedkin, Galiya Latypova,

Nestor Sukhodolsky, Alexander Shemenev Ivan Borodulin,

19 Investigation of the dependence of the quality of recognition of ontological objects on the depth of hyponymy. Code,

Report, LinkReview, Presentation

Вячеслав Резяпкин, Alexey Russkin,

Victoria Dochkina, Miron Kuznetsov, Yarmoshyk Demyan

20 Comparison of the quality of end-to-end trainable models in the task of answering questions in a dialogue, taking into account the context Code

LinkReview Report, Presentation

Agafonov Alexey, Ryakin Ilya,Litvinenko Vladimir,

Khokhlov Ivan, Velikovsky Nikita, Anufrienko Oleg

21 High order convex optimization methods Code,

LinkReview, Slides

Selikhanovich Daniel,

Sokolov Igor

23 Fractal analysis and synthesis of optical images of sea waves code,

LinkReview, Presentation Report

Kanygin Yuri
24 Entropy maximization for various types of image transformations code,

LinkReview, report, slides

Nikita Voskresensky,

Alisa Shabalina, Yaroslav Murzaev, Alexey Khokhlov, Alexey Kazakov, Olga Gribova, Alexander Belozertsev

25 Automatic detection and recognition of objects in images code,

code_A, Slides_for_demo, Report2018Project25_30 Report2018Project25_31 slides_30 slides_25_31 LinkReview

Julia Demidova

Ivan Razumov Vladislav Tominin Yaroslav Tominin Nikita Dudorov Leonid Erlygin Proshutinsky Dmitry Baimakov Vladimir Zubkov Alexander Chernenkova Elena

26 Location determination by accelerometer signals Code,

LinkReview, Slides, Text

Elvira Zainulina

Fateev Dmitry Vitaly Protasov Nikita Bozhedomov

28 Multimodelling as a universal way to describe a general sample Code,

Linkreview, Slides, Report

Vladimir Kachanov

Evgenia Strelkova

29 Cross-Language Document Extractive Summarization with Neural Sequence Model Code,

Linkreview, Report, Slides

Pavel Zakharov

Pavel Kvasha Evgeny Dyachkov Evgeny Petrov Ilya Selnitsky

31 Pairwise energy matrix construction for inverse folding problem Code,

LinkReview Report Slides

Rubinshtein Alexander
32 Smooth orientation-dependent scoring function Code

Отчёт

Noskova Elizaveta

Kachkov Sergey Sidorenko Anton

Task 5

  • Name: Finding paraphrases.
  • Task: Paraphrases are different variations of the same and the same text, identical in meaning, but differing lexically and grammatically, for example: "Where did the car go" and "Which direction did the car go". The task of detecting paraphrases is to select clusters in a set of texts, such that each cluster contains only paraphrases of the same and the same sentence.

The easiest way to extract paraphrases is to cluster texts, where each text is represented by a "bag of words".

  • . Data: There are open datasets of questions for testing and training on kaggle.com, there are open datasets for testing from semeval conferences.
  • References:
    1. Will be later
  • Basic algorithm: Use one of the document clustering algorithms to extract paraphrases, where each document is represented by a bag of words or tf-idf.
  • Solution: Use neural network architectures to search for paraphrases, use phrases extracted with parsers as features, use multilevel clustering.
  • Novelty: Lack of implementations for the Russian language that will use parsers for a similar task, all current solutions are quite "simple".
  • Authors: Artyom Popov.

Task 6

  • Name: On conformational changes of proteins using collective motions in torsion angle space and L1 regularization.
  • Task: Torsion angles are the most natural degrees of freedom for describing motions of polymers, such as proteins. This is because bond lengths and bond angles are heavily constrained by covalent forces. Thus, multiple attempts have been done to describe protein dynamics in the torsion angle space. For example, one of us has developed an elastic network model (ENM) [1] in torsion angle space called Torsional Network Model (TNM) [2]. Functional conformational changes in proteins can be described in the Cartesian space using just a subset of collective coordinates [3], or even a sparse representation of these [4]. The latter requires a solution of a LASSO optimization problem [5]. The goal of the current project is to study if a sparse subset of collective coordinates in the torsion subspace can describe functional conformational changes in proteins. This will require a solution of a ridge regression problem with a L1 regularization constraint. The starting point will be the LASSO formulation.
  • . Data: Experimental conformations will be extracted from the Protein Docking Benchmark v5 (https://zlab.umassmed.edu/benchmark/) and a few others. The TNM model can be downloaded from https://ub.cbm.uam.es/tnm/tnm_soft_main.php
  • References:
    1. Tirion MM. (1996) Large Amplitude Elastic Motions in Proteins from a Single-Parameter, Atomic Anal- ysis. Phys Rev Lett. 77:1905–1908.
    2. Mendez R, Bastolla U. (2011) Torsional network model: normal modes in torsion angle space better correlate with conformation changes in proteins. Phys Rev Lett. 2010 104:228103.
    3. SwarmDock and the use of normal modes in protein-protein docking. IH Moal, PA Bates - International journal of molecular sciences, 2010
    4. Modeling protein conformational transition pathways using collective motions and the LASSO method. TW Hayes, IH Moal - Journal of chemical theory and computation, 2017
    5. https://en.wikipedia.org/wiki/Lasso_(statistics)
    6. E. Frezza, R. Lavery, Internal normal mode analysis (iNMA) applied to protein conformational flexibility, Journal of Chemical Theory and Computation 11 (2015) 5503–5512.
  • Basic algorithm: The starting point will be a combination of methods from references 2 and 4. It has to be a LASSO formulation with the direction vectors reconstructed from the internal coordinates. The quality will be computed based on the RMSD measure between the prediction and the solution on several benchmarks. Results will be presented with statistical plots (see examples in references 3-4.
  • Novelty: This is an important and open question in computational structural bioinformatics - how to efficiently represent transitions between protein structures. Not much has been done in the torsional angle subspace (internal coordinates)[6] and nearly nothing has been done using L1 regularization [4].
  • Authors: Ugo Bastolla on the torsional subspace (https://ub.cbm.uam.es/home/ugo.php), Sergei Grudinin on L1 minimization (https://team.inria.fr/nano-d/team-members/sergei-grudinin/)

Task 10

  • Name: Comparison of neural network and continuous-morphological methods in the problem of text detection (Text Detection).
  • Task: Automatically Detect Text in Natural Images.
  • Data: Synthetic generated data + prepared sample of photos + COCO-Text dataset + Конкурс Avito 2014.
  • References:: COCO benchmark, One of a state-of-the-art architecture
  • Basic algorithm: code + morphological methods, Avito 2014 winner’s solution.
  • Solution: It is proposed to compare the performance of several state-of-the-art algorithms that need a large training set with morphological methods that require a small amount of data. It is proposed to determine the limits of applicability of certain methods.
  • Novelty: propose an algorithm based on the use of both neural network and morphological methods (solution of the word detection problem).
  • Authors: I. N. Zharikov.
  • Expert: L. M. Mestetsky (morphological methods).

Task 16

  • Name: Estimate of the optimal sample size for research in medicine
  • Task: In conditions of an insufficient number of expensive measurements, it is required to predict the optimal size of the replenished sample.
  • Data: Samples of measurements in medical diagnostics, in particular, a sample of immunological markers.
  • References::
  • Basic algorithm: A series of empirical sample size estimation algorithms.
  • Solution: Investigation of the properties of the parameter space when replenishing the sample.
  • Novelty: A new methodology for sample size forecasting is proposed, justified in terms of classical and Bayesian statistics.
  • Authors: A.M. Katrutsa, Strizhov V.V., coordinator Tamaz Gadaev

Task 19

  • Name: Study of the dependence of the quality of recognition of ontological objects on the depth of hyponymy.
  • Task: It is necessary to investigate the dependence of the quality of recognition of ontological objects at different levels of concept hyponymy. The classic formulation of the problem of named entity recognition: https://en.wikipedia.org/wiki/Named-entity_recognition
  • Data: Hyponyms from https://wordnet.princeton.edu/ , texts from different domains presumably from WebOfScience.
  • References: Relevant articles for classical staging http://arxiv-sanity.com/search?q=named+entity+recognition
  • Basic algorithm: https://arxiv.org/pdf/1709.09686.pdf or its simplified version can be used as an algorithm, studies are performed using the DeepPavlov library.
  • Solution: It is necessary to collect a dataset of hyponymy (nesting of concepts) of objects using WordNet, to automatically mark up ontological objects of texts of various domains for several levels of generalization of concepts, to conduct a series of experiments to determine the quality of recognition of ontological objects for different levels of nesting.
  • Novelty: Similar studies have not been carried out, there are no ready-made datasets with a hierarchical markup of objects. Recognition of ontological objects at various levels of hyponymy can be used to produce additional features when solving various NLP (Natural language processing) tasks, as well as determining whether objects are a hyponym-hypernym pair.
  • Authors: Burtsev Mikhail Sergeevich (Expert), Baimurzina Dilyara Rimovna (consultant).

Task 20

  • Name: Сравнение качества end-to-end обучаемых моделей в задаче ответа на вопросы в диалоге с учетом контекста
  • Task: Задан фрагмент текста and несколько последовательных вопросов. Ответы на первые n вопросов известны. Нужно сформировать ответ на n+1 вопрос. В качестве ответа нужно указать непрерывный промежуток в тексте заданного фрагмента текста (номера начального and конечного слов). При оценке качества ответа Task сводится к классификации символов фрагмента на класс 0 (не входит в ответ) and 1 (входит в ответ).
  • Data: Предоставляется размеченный датасет с фрагментами текста and наборами вопросов с ответами в диалоге
  • References: Статья Bi-directional Attention Flow for Machine Comprehension (BiDAF2017) описывает end-to-end модель ответов на вопросы по фрагменту без учета контекста диалога. Статья QuAC: Question Answering in Context (QuAC2018) описывает набор данных, содержит описание используемого базового алгоритма с учетом контекста диалога. Статьи с описанием других моделей вопрос-ответных систем (R-Net, DrQA)
  • Basic algorithm: Basic algorithm описан статьях and реализован (QuAC2018, BiDAF2017).
  • Solution: Предлагается изучить механизмы учета контекста (k-ctx, append, etc) and исследовать возможность их добавления в другие модели (DrQA, R-NET), либо предложить собственные для повышения качества по мере F1. Для изучения поведения модели используется визуализация внимания (attention visualization), обучаемых эмбеддингов, а также анализ ошибочных ответов. Предоставляется доступ к вычислительным ресурсам, используемые фреймворки: TensorFlow, PyTorch или Keras.
  • Novelty: Исследование проводится на новом датасете, для которого на данный момент имеется только Basic algorithm. Подтверждение повышения качества от применения механизмов учета контекста диалога в других моделях указывает на применимость предлагаемых подходов для решения более широкого круга задач.
  • Authors: Антон Сергеевич Хританков

Task 21

  • Name: High order convex optimization methods
  • Task: High-order methods are effectively (up to n ~ 10^3 sometimes even up to n ~ 10^4) used for convex problems of not very large dimensions. Until recently, it was generally accepted that these are second-order methods (using the second derivatives of the function being optimized). However, at the beginning of 2018 Yu.E. Nesterov [1] proposed an efficient third-order method in the theory, which works according to almost optimal estimates. In the manual [3] in exercise 1.3, an example of a "bad" convex function proposed by Yu.E. Nesterov, on which I would like to compare the Nesterov method of the second and third order [1], the method from [2] of the second and third order and the usual fast gradient methods (of the first order). It is worth comparing both by the number of iterations and by the total running time.
  • References:
  1. https://alfresco.uclouvain.be/alfresco/service/guest/streamDownload/workspace/SpacesStore/aabc2323-0bc1-40d4-9653-1c29971e7bd8/coredp2018_05web.pdf?guest=true
  2. https://arxiv.org/pdf/1809.00382.pdf
  3. https://arxiv.org/pdf/1711.00394.pdf
  • Author: Evgenia Alekseevna Vorontsova (Associate Professor of Far Eastern Federal University, Vladivostok), Alexander Vladimirovich Gasnikov

Task 22

  • Name: Cutting plane methods for copositive optimization
  • Task: Conic program over the copositive cone (copositive program) min <C,X> : <A_i,X> = b_i, X \in \Pi_i C^k_i, k_i <= 5 A linear function is minimized over the intersection of an affine subspace with a product of copositive cones of orders k_i <= 5. Подробнее тут
  • Data: The algorithm will be tested on randomly generated instances
  • References:
    • [1] Peter J. C. Dickinson, Mirjam Dür, Luuk Gijben, Roland Hildebrand. Scaling relationship between the copositive cone and Parrilo’s first level approximation. Optim. Lett. 7(8), 1669—1679, 2013.
    • [2] Stefan Bundfuss, Mirjam Dür. Algorithmic copositivity detection by simplicial partition. Linear Alg. Appl. 428, 1511—1523, 2008.
    • [3] Mirjam Dür. Copositive programming — a Survey. In Recent advances in Optimization and its Applications in Engineering, Springer, pp. 3-20, 2010.
  • Basic algorithm: The reference algorithm is described in [4] Stefan Bundfuss, Mirjam Dür. An Adaptive Linear Approximation Algorithm for Copositive Programs. SIAM J. Optim., 20(1), 30-53, 2009.
  • Solution: The copositive program will be solved by a cutting plane algorithm. The cutting plane (in the case of an infeasible iterate) will be constructed from the semidefinite representation of the diagonal 1 section of the cone proposed in [1]. The algorithm will be compared to a simplicial division method proposed in [2], [4]. General information about copositive programs and their applications in optimization can be found in [3] .
  • Novelty: The proposed algorithm for optimization over copositive cones up to order 5 uses an exact semi-definite representation. In contrast to all other algorithms existing today the generation of cutting planes is non-iterative.
  • Автор: Roland Hildebrand

Task 23

  • Name: Fractal analysis and synthesis of optical images of sea waves
  • Task: A variety of physical processes and phenomena are studied with the help of images obtained remotely. An important task is to obtain adequate information about the processes and phenomena of interest by measuring certain image characteristics. Lines of equal brightness (isolines) on the images of many natural objects are fractal, that is, they are sets of points that cannot be represented by lines of finite length and occupy an intermediate position between lines and two-dimensional flat figures. Such sets are characterized by the fractal dimension D, which generalizes the classical concept of the dimension of a set and can take fractional values. For a solitary point on the image D=0, for a smooth curve D=1, for a flat figure D=2. The fractal isoline has the dimension 1<D<2. The algorithm for calculating D is given, for example, in [1]. The fractal dimension of the sea surface isolines can serve to estimate the spatial spectra of sea waves according to remote sensing data [1]. Task is as follows. It is necessary to conduct a numerical study of the relationship between the characteristics of the spatial spectra of sea waves and the fractal dimension of satellite images of the Earth in the solar glare region. For the study, the method of numerical synthesis of optical images of sea waves, described in [2], should be used. Numerical modeling should be done with different characteristics of sea waves, as well as with different positions of the Sun and spatial resolution of images.
  • References:
    1. Lupyan E. A., Murynin A. B. Possibilities of fractal analysis of optical images of the sea surface. // Preprint of the Space Research Institute of the Academy of Sciences of the USSR Pr.-1521, Moscow, 1989, 30 p.
    2. Murynin A. B. Reconstruction of the spatial spectra of the sea surface from optical images in a nonlinear model of the brightness field // Research of the Earth from Space, 1990. No. 6. P. 60-70.
  • Author: Ivan Alekseevich Matveev

Task 24

  • Name Entropy maximization for various types of image transformations
  • Task: Pansharpening is an algorithm for upscaling multispectral images using a reference image. The task of pansharpening is formulated as follows: having a panchromatic image of the required resolution and a multispectral image of reduced resolution, it is required to restore the multispectral image in the spatial resolution of the panchromatic one. From empirical observations based on a large number of high-resolution images, it is known that the spatial variability of the reflected radiation intensity for objects of the same nature is much greater than the variability of their spectrum. In other words, one can observe that the spectrum of reflected radiation is homogeneous within the boundaries of one object, while even within one object the intensity of reflected radiation varies. In practice, good results can be achieved using a simplified approach, in which it is assumed that if the intensity of neighboring regions differ significantly, then these regions probably belong to different objects with different reflected spectra. This is the basis for the developed probabilistic algorithm for increasing the resolution of multispectral images using a reference image [1]
  • It is necessary to conduct a study on maximizing the entropy for various types of transformations on the image. Show that entropy can serve as an indicator of the loss of information contained in the image during transformations over it. Formulation of the inverse problem for image restoration: Condition 1: Correspondence of the intensity (at each point) of the restored image with the intensity of the panchromatic image. Condition 2: Correspondence of the low-frequency component of the reconstructed image with the original multispectral image. Condition 3: Homogeneity (similarity) of the spectrum within one object and the assumption of an abrupt change in the spectrum at the border of two homogeneous regions. Condition 4: Under the first three conditions, the local entropy of the reconstructed image must be maximized.
  • References:
    1. Gorohovsky K. Yu., Ignatiev V. Yu., Murynin A. B., Rakova K. O. Search for optimal parameters of a probabilistic algorithm for increasing the spatial resolution of multispectral satellite images // Izvestiya RAN. Theory and control systems, 2017, No. 6.
  • Author: Ivan Alekseevich Matveev

Task 25

  • Name: Automatic detection and recognition of objects in images
  • Task: Automatic detection and recognition of objects in images and videos is one of the main tasks of computer vision. As a rule, these tasks are divided into several subtasks: preprocessing, extraction of the characteristic properties of the object image and classification. The pre-processing stage usually includes some operations on the image such as filtering, brightness equalization, geometric corrective transformations to facilitate robust feature extraction.

The characteristic properties of an image of an object are understood as a set of features that approximately describe the object of interest. Features can be divided into two classes: local and integral. The advantage of local features is their versatility, invariance with respect to uneven changes in brightness and illumination, but they are not unique. Integral features that characterize the image of the object as a whole are not resistant to changes in the structure of the object and difficult lighting conditions. There is a combined approach - the use of local features as elements of an integral description, when the desired object is modeled by a set of areas, each of which is characterized by its own set of features - a local texture descriptor. The totality of such descriptors characterizes the object as a whole. Classification is understood as determining whether an object belongs to a particular class by analyzing the feature vector obtained at the previous stage, dividing the feature space into subdomains indicating the corresponding class. There are many approaches to classification: neural network, statistical (Bayesian, regression, Fisher, etc.), decision trees and forests, metric (nearest K-neighbors, Parzen windows, etc.) and nuclear (SVM, RBF, method of potential functions), compositional (AdaBoost). For the task of detecting an object in an image, membership in two classes is evaluated - the class of images containing the object, and the class of images that do not contain the object (background images).

Task 29

  • Name: Cross-Language Document Extractive Summarization with Neural Sequence Model.
  • Task: It is proposed to solve the transfer learning problem for the text reduction model by extractive summarization and to investigate the dependence of the quality of text reduction on the quality of training of the translation model. Having data for training the abbreviation model in English and a parallel English-Russian corpus of texts, build a model for abbreviating the text in Russian. The solution of the problem is evaluated on a small set of data for testing the model in Russian, the quality of the solution to the problem is determined by the ratio of the values of the ROUGE criteria in English and Russian sets.
  • Data: Data for training the model in English (SummaRuNNer2016), OPUS parallel corpus, data for verification in Russian.
  • References: The article (SummaRuNNer2016) describes the basic text reduction algorithm, the work Neural machine translation by jointly learning to align and translate.(NMT2016) describes the translation model. The idea of sharing models is presented in Cross-Language Document Summarization Based on Machine Translation Quality Prediction (CrossSum2010).
  • Basic algorithm: One idea of the basic algorithm is presented in (CrossSum2010), a translation model is implemented (OpenNMT), an implementation of a text reduction model is provided (SummaRuNNer2016).
  • Solution: It is suggested to explore the solution idea proposed in the article (CrossSum2010) and options for combining reduction and translation models. Basic models and dataset preprocessing implemented (OpenNMT), PyTorch and Tensorflow libraries. Analysis of text reduction errors is performed as described in (SummaRuNNer2016), analysis of the quality of model training by standard library tools, .
  • Novelty: For the base model, the applicability was investigated on a couple of datasets, confirming the possibility of transferring training to a dataset in another language and specifying the conditions for this transfer will expand the scope of the model and indicate the necessary new refinements of the model or data preprocessing.
  • Authors: Alexey Romanov (consultant), Anton Khritankov (Expert).

Task 30

  • Name: Method for constructing an HG-LBP descriptor based on gradient histograms for pedestrian detection.
  • Task: It is proposed to develop a new descriptor that generalizes the LBP descriptor based on histograms of gradient modules, having HOG-LBP composition properties for the task of detecting pedestrians in an image. As an analysis of the quality of a new descriptor, it is proposed to use FAR/FRR detection error plots based on INRIA.
  • Data: INRIA pedestrian database: http://pascal.inrialpes.fr/data/human/
  • References:
    1. 1. T. Ojala and M. Pietikainen. Multiresolution Gray-Scale and Rotation Invariant Texture Classification with Local Binary Patterns, IEEE Trans on Pattern Analysis and Machine Intelligence, Vol. 24. No. 7, July, 2002.
    2. 2. T. Bouwmans, C. Silva, C. Marghes, M. Zitouni, H. Bhaskar, C. Frelicot, "On the Role and the Importance of Features for Background Modeling and Foreground Detection", https:// arxiv.org/pdf/1611.09099v1.pdf
    3. 3. N. Dalal and B. Triggs, Histograms of Oriented Gradients for Human Detection, Proc. IEEE Conference on Computer Vision and Pattern Recognition, 2005, pp.886-893
    4. 4. T. Ahonen, A. Hadid, M. Pietikainen Face Description with Local Binary Patterns: Application to Face Recognition \\ IEEE Transactions on Pattern Analysis and Machine Intelligence, Volume:28 , Issue: 121.
    5. 5. http://www.magicandlove.com/blog/2011/08/26/people-detection-in-opencv-again/
    6. 6. http://www.cse.oulu.fi/CMV/Downloads/LBPMatlab2.
    7. 7. http://www.mathworks.com/help/vision/ref/extractlbpfeatures.html3.
    8. 8. http://www.codeproject.com/Articles/741559/Uniform-LBP-Features-and-Spatial-Histogram-Computa4.
    9. 9. http://www.cse.oulu.fi/CMV/Research
  • Basic algorithm: Xiaoyu Wang, Tony X. Han, Shuicheng Yan. An HOG-LBP Human Detector with Partial Occlusion Handling \\ ICCV 2009
  • Solution: One of the options for generalizing LBP can be to use instead of histograms of distribution of points by LBP code, histograms of distribution of modules of point gradients in a block by LBP code (HG-LBP). It is proposed to use the OpenCV library for the basis of experiments, in which the HOG and LBP algorithms are implemented. It is necessary to modify the source code of the LBP implementation and insert the calculation of the modules of the gradient and the accumulation of the corresponding histogram over the LBP. It is necessary to write a program for reading the INRIA base, learning the linear SVM method on the original and modified descriptors, collecting detection statistics and plotting FAR/FRR DET plots.
  • Novelty: The development of computationally simple methods for extracting the most informative features in recognition tasks is relevant in the field of creating embedded systems with low computing resources. Replacing the composition of descriptors with one that is more informative than each individually can simplify the solution of the problem. The use of gradient values in LPB descriptor histograms is new.
  • Authors: Gneushev Alexander Nikolaevich

Task 31

  • Name: Using the HOG descriptor to train a neural network in a pedestrian detection task
  • Task: It is proposed to replace the linear SVM classifier in the classical HOG algorithm with a simple convolutional neural network of small depth, while the HOG descriptor should be represented by a three-dimensional tensor that preserves the spatial structure of local blocks. As an analysis of the quality of a new descriptor, it is proposed to use FAR/FRR detection error plots based on INRIA.
  • Data: INRIA pedestrian database: http://pascal.inrialpes.fr/data/human/
  • References:
    1. 1. N. Dalal and B. Triggs, Histograms of Oriented Gradients for Human Detection, Proc. IEEE Conference on Computer Vision and Pattern Recognition, 2005, pp.886-893
    2. 3. Q. Zhu, S. Avidan, M.-C. Yeh, and K.-T. Cheng. Fast human detection using a cascade of histograms of oriented gradients. In CVPR, pages 1491-1498, 2006 O. Tuzel, F. Porikli, and P. Meer. Human detection via classification on riemannian manifolds. In CVPR, 2007
    3. 4. P. Dollar, C. Wojek, B. Schiele and P. Perona Pedestrian Detection: An Evaluation of the State of the Art / IEEE Transactions on Pattern Analysis and Machine Intelligence (PAMI), Vol 34. Issue 4, pp . 743-761
    4. 5. Xiaoyu Wang, Tony X. Han, Shuicheng Yan, An HOG-LBP Human Detector with Partial Occlusion Handling, ICCV 2009 http://www.xiaoyumu.com/s/PDF/Wang_HOG_LBP.pdf
    5. 6. https://en.wikipedia.org/wiki/Pedestrian_detection
    6. 7. HOG person detector tutorial https://chrisjmccormick.wordpress.com/2013/05/09/hog-person-detector-tutorial/
    7. 8. NavneetDalalThesis.pdf Navneet Dalal. Finding People in Images and Videos. PhD Thesis. Institut National Polytechnique de Grenoble / INRIA Rhone-Alpes, Grenoble, July 2006)
    8. 9. People Detection in OpenCV http://www.magicandlove.com/blog/2011/08/26/people-detection-in-opencv-again/
    9. 10. Andrew G. Howard, Menglong Zhu, Bo Chen, Dmitry Kalenichenko, Weijun Wang, Tobias Weyand, Marco Andreetto, Hartwig Adam. MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications
  • Basic algorithm:
    1. 1. N. Dalal and B. Triggs, Histograms of Oriented Gradients for Human Detection, Proc. IEEE Conference on Computer Vision and Pattern Recognition, 2005, pp.886-893
    2. 2. Xiaoyu Wang, Tony X. Han, Shuicheng Yan, An HOG-LBP Human Detector with Partial Occlusion Handling, ICCV 2009
  • Solution: One of the options for generalizing the HOG algorithm can be to use another classifier instead of the linear SVM algorithm, for example, some kind of neural network. It is proposed to use the OpenCV library for the basis of experiments, which implements the HOG algorithm and the SVM classifier. It is necessary to analyze the source code of the HOG implementation, formalize the internal structure of the descriptor HOG vector in the form of a three-dimensional tensor — two spatial and one spectral dimensions. It is necessary to write a program for reading the INRIA base, learning the linear SVM method on HOG descriptors from it, collecting detection statistics and plotting FAR/FRR DET plots. Based on some neural network training system (for example, mxnet), it is necessary to assemble a shallow (no more than 2-3 convolutional layers) convolutional neural network of known architecture, train it on the basis of INRIA and on HOG tensor descriptors, build the corresponding FAR / FRR graphs.
  • Novelty: The development of computationally simple methods for extracting the most informative features in recognition tasks is relevant in the field of creating embedded systems with low computing resources. Using a small number of the most informative descriptors can reduce computational complexity compared to using a large composition of simple features, such as in a deep convolutional neural network. Typically, classifiers use the HOG descriptor as a vector as a whole, however, information about the local spatial structure and feature spectrum is lost. The novelty lies in the use of the block locality property in the HOG descriptor and the representation of the HOG as a 3D tensor. The use of this information makes it possible to achieve detection resistance to pedestrian overlap.
  • Authors: Gneushev Alexander Nikolaevich

YEAR

Author Topic Links Consultant Reviewer Report Letters \Sigma=3+13
Goncharov Alexey (example) Metric classification of time series code,

paper, slides

Maria Popova Zadayanchuk Andrey BMF AILSBRCVTDSWH>
Astakhov Anton Restoring the structure of a predictive model from a probabilistic representation folder

code paper

Alexander Katrutsa Kislinsky Vadim BHF A-I-L0S0B0R0C0V0T0 [A-I-L-S-B0R0C0V0T0E0D0W0S] + [AILSBRCBTEDWS] 2+4
Gavrilov Yuri Choice of Interpreted Multimodels in Credit Scoring Tasks folder

code paper video

Goncharov Alexey Ostroukhov Petr BF A+IL-S0B-R0 [A+ILSBRC-VT0E0D0W0S] + (W) 2+9+1
Gadaev Tamaz Estimating the optimal sample size folder

code paper slides video

Alexander Katrutsa Shulgin Egor BHF A-IL>SB-R-C0V0T0 [AILSBR0CVT0E-D0W0S] 2+9
Gladin Egor Accelerometer Battery Savings Based on Time Series Forecasting folder

code paper slides

Maria Vladimirova Kozlinsky Evgeny

review

.F AILS [A-I-L-SB0R0C000V0T0E0D0W0S] 1+4
Grabovoi Andrey Automatic determination of the relevance of neural network parameters. folder

code paper slides video

Oleg Bakhteev Kulkov Alexander BHMF A+ILS+BRC+VTE>D> [AILSBRCVTEDWS] [\emptyset] 3+13
Nurlanov Zhakshylyk Deep Learning for reliable detection of tandem repeats in 3D protein structures folder

code paper slides video

S. V. Grudinin, Guillaume Pages Pletnev Nikita

Review

BHF AILB [A-I-LS-BRC0V0T-E0D0W0S] 2+7
Rogozina Anna Deep learning for RNA secondary structure prediction folder

code paper slides video

Maria Popova Gadaev Tamaz BHMF AILSBR> [AILSBRC0V0T0E0D0W0S]+CW 3+9
Terekhov Oleg Generation of features using locally approximating models folder

code paper slides

S.D. Ivanychev, R.G. Neichev Gladin Egor

review

BHM AILSBRCVTDSW [AIL0SB0R0C0V0TE0D0W0S] 2+12
Shulgin Egor Generation of features that are invariant to changes in the frequency of the time series folder

code paper

R.G. Neichev Terekhov Oleg BHM AIL [AI-LS-BR0CV0T0E0D0W0S] 2+5
Malinovsky Grigory Graph Structure Prediction of a Neural Network Model folder

code paper slides video

Oleg Bakhteev Grabovoi Andrey

review

BHMF A+I+L+SBR>C>V>T>E>D> [AILSBRC0VTED0WS]+(C) 3+11
Kulkov Alexander Brain signal decoding and intention prediction folder

code paper slides video

R.V. Isachenko Malinovsky Grigory

review

BHMF AILSBR [AILSBRCVTED0W0S] 3+11
Pletnev Nikita Approximation of the boundaries of the iris paper

slides [ video]

Alexander Aduenko Nurlanov Zhakshylyk BF AILSB>R> [AILSTWS] 2+7
Ostroukhov Petr Selection of models superposition for identification of a person on the basis of a ballistocardiogram folder

paper code slides

Alexander Prozorov Gavrilov Yuri

review

BhF AIL>S?B?R? [AILSBRCVT-E0D0W0S] 2+10
Kislinsky Vadim Predicting user music playlists in a recommender system. folder

code slides paper video

Evgeny Frolov Astakhov Anton .F (AIL)------(SB)---(RCVT)-- [AILS-BRCVTED0W0S] 1+11
Kozlinsky Evgeny Analysis of banking transactional data of individuals to identify customer consumption patterns. folder

code paper slides video

Rosa Aisina Rogozina Anna

review

BHMF AILSBR>CV> [AILSBR0C0V0TE0D0WS]+(С) 3+8+1

Task 1

  • Name: Approximation of the boundaries of the iris
  • Task: Based on the image of the human eye, determine the circles approximating the inner and outer border of the iris.
  • Data: Bitmap monochrome images, typical size 640*480 pixels (however other sizes are possible)[52], [53].
  • References::
    • Aduenko A.A. Selection of multi-models in Tasks classification (supervisor Strizhov V.V.). Moscow Institute of Physics and Technology, 2017. [54]
    • K.A. Gankin, A.N. Gneushev, I.A. Matveev Segmentation of the iris image based on approximate methods with subsequent refinements // Izvestiya RAN. Theory and control systems, 2014, no. 2, p. 78–92.
    • Duda, R. O. Use of the Hough transformation to detect lines and curves in pictures / R. O. Duda, P. E. Hart // Communications of the ACM. 1972 Vol. 15, no. 1.Pp.
  • Basic algorithm: Efimov Yury. Search for the outer and inner boundaries of the iris in the eye image using the paired gradient method, 2015.
  • Solution: See iris_circle_problem.pdf
  • Novelty: A fast non-enumerative algorithm for approximating boundaries using linear multimodels is proposed.
  • consultant: Alexander Aduenko (by Strizhov V.V., Expert Matveev I.A.)

Task 2

  • Name: Estimated optimal sample size
  • Task: In conditions of an insufficient number of expensive measurements, it is required to predict the optimal size of the replenished sample.
  • Data: Samples of measurements in medical diagnostics, in particular, a sample of immunological markers.
  • References::
  • Basic algorithm: Sample size estimation algorithms for .
  • Solution: Investigation of the properties of the parameter space when replenishing the sample.
  • Novelty: A new methodology for sample size forecasting is proposed, justified in terms of classical and Bayesian statistics.
  • Authors: A.M. Katrutsa, Strizhov V.V., Expert A.P. Motrenko

Task 3

  • Name: Restoring the structure of the prognostic model from a probabilistic representation
  • Task: It is required to reconstruct the superposition tree from the generated connection probability graph.
  • Data: Segments of time series, spatio-temporal series (and text collections).
  • References::
    • Works by Tommy Yakkola and others at LinkReview [56].
  • Basic algorithm: Branch and bound method, dynamic programming when building a fully connected graph.
  • Solution: Building a model in the form of GAN, VAE generates a weighted graph, NN approximates a tree structure.
  • Novelty: Suggested a way to penalize a graph for not being a tree. A method for predicting the structures of prognostic models is proposed.
  • Authors: A.M. Katrutsa, Strizhov V.V.

Task 4

  • Name: Text recognition based on skeletal representation of thick lines and convolutional networks
  • Task: It is required to build two CNNs, one recognizes a bitmap representation of an image, the other a vector one.
  • Data: Bitmap fonts.
  • References:: List of works [57], in particular arXiv:1611.03199 and
  • Basic algorithm: Convolution network for bitmap.
  • Solution: It is required to propose a method for collapsing graph structures, which allows generating an informative description of the skeleton of a thick line.
  • Novelty: A way to improve the quality of recognition of thick lines due to a new way of generating their descriptions is proposed.
  • Authors: L.M. Mestetsky, I.A. Reyer, Strizhov V.V.

Task 5

  • Name: Generation of features using locally approximating models
  • Task: It is required to test the feasibility of the hypothesis of simplicity of sampling for the generated features. Features are the optimal parameters of approximating models. Moreover, the entire sample is not simple and requires a mixture of models to approximate it. Explore the information content of the generated features - the parameters of the approximating models trained on the segments of the original time series.
  • Data:
    • WISDM (Kwapisz, J.R., G.M. Weiss, and S.A. Moore. 2011. Activity recognition using cell phone accelerometers. ACM SigKDD Explorations Newsletter. 12(2):74–82.), USC-HAD or higher. Accelerometer data (Human activity recognition using smart phone embedded sensors: A Linear Dynamical Systems method, W Wang, H Liu, L Yu, F Sun - Neural Networks (IJCNN), 2014)
    • (Time series (examples library), Accelerometry section).
  • References::
    • Kuznetsov M.P., Ivkin N.P. Algorithm for Classifying Accelerometer Time Series by Combined Feature Description // Machine Learning and Data Analysis. 2015. V. 1, No. 11. C. 1471-1483. [58]
    • Karasikov M.E., Strizhov V.V. Classification of time series in the space of parameters of generating models // Informatics and its applications, 2016.URL
    • Kuznetsov M.P., Ivkin N.P. Algorithm for Classifying Accelerometer Time Series by Combined Feature Description // Machine Learning and Data Analysis. 2015. V. 1, No. 11. C. 1471 - 1483. URL
    • Isachenko R.V., Strizhov V.V. Metric learning in Taskx multiclass classification of time series // Informatics and its applications, 2016, 10(2) : 48-57. URL
    • Zadayanchuk A.I., Popova M.S., Strizhov V.V. Choosing the optimal model for classifying physical activity based on accelerometer measurements // Information technologies, 2016. URL
    • Motrenko A.P., Strijov V.V. Extracting fundamental periods to segment human motion time series // Journal of Biomedical and Health Informatics, 2016, Vol. 20, no. 6, 1466 - 1476.
    • Ignatov A., Strijov V. Human activity recognition using quasiperiodic time series collected from a single triaxial accelerometer // Multimedia Tools and Applications, 2015, 17.05.2015 : 1-14. URL
  • Basic algorithm: Described by Kuznetsov, Ivkin.
  • Solution: It is required to build a set of locally approximating models and choose the most adequate ones.
  • Novelty: A standard for building locally approximating models has been created.
  • Authors: S.D. Ivanychev, R.G. Neichev, Strizhov V.V.

Task 6

  • Name: Brain signal decoding and intention prediction
  • Task: It is required to build a model that restores the movement of the limbs from the corticogram.
  • Data: neurotycho.org [59]
  • References::
    • Neichev R.G., Katrutsa A.M., Strizhov V.V. Selection of the optimal set of features from a multicorrelated set in the forecasting problem. Zavodskaya Lab. Diagnostics of materials, 2016, 82(3) : 68-74. [60]
    • MLAlgorithms: Motrenko, Isachenko (submitted)
  • Basic algorithm: Partial Least Squares[61]
  • Solution: Create a feature selection algorithm alternative to PLS and taking into account the non-orthogonal structure of feature interdependence.
  • Novelty: A feature selection method is proposed that takes into account the regularities of both the and independent variable and the dependent variable.
  • Authors: R.V. Isachenko, Strizhov V.V.

Task 7

  • Name: Automatic determination of the relevance of neural network parameters.
  • Task: The Task of finding a stable (and not redundant in terms of parameters) neural network structure is considered. To cut off redundant parameters, it is proposed to introduce a priori probabilistic assumptions about the distribution of parameters and remove non-informative parameters from the neural network using the Belsley method. To adjust the prior distribution, it is proposed to use gradient methods.
  • Data: A selection of handwritten MNIST digits
  • Basic algorithm: Optimal Brain Damage, decimation based on variance inference. The structure of the final model is proposed to be compared with the model obtained by the AdaNet algorithm.
  • References::
    • [62] Gradient hyperparameter optimization methods.
    • [63] Gradient hyperparameter optimization methods.
    • [64] Optimal Brain Damage.
    • [65] AdaNet
    • [66] Belsley Method
  • Authors: Oleg Bakhteev, Strizhov V.V.

Task 8

  • Name: Prediction of the graph structure of the neural network model.
  • Task: The Task is considered to find a stable (and non-redundant in terms of parameters) structure of a convolutional neural network. It is proposed to predict the structure of a neural network using doubly-recurrent neural networks. As a training sample, it is proposed to use the structures of models that have shown good quality on subsamples of small power.
  • Data: Samples MNIST, CIFAR-10
  • Basic algorithm: random search. Comparison with work on reinforcement learning is possible.
  • References::
    • [67] doubly-recurrent neural networks.
    • [68] Similar approach using reinforcement learning.
  • Authors: Oleg Bakhteev. Strijov V.V.

Task 9

  • Name: Deep Learning for reliable detection of tandem repeats in 3D protein structures more in PDF
  • Task: Deep learning algorithms pushed computer vision to a level of accuracy comparable or higher than a human vision. Similarly, we believe that it is possible to recognize the symmetry of a 3D object with a very high reliability, when the object is represented as a density map. The optimization problem includes i) multiclass classification of 3D data. The output is the order of symmetry. The number of classes is ~10-20 ii) multioutput regression of 3D data. The output is the symmetry axis (a 3-vector). The input data are typically 24x24x24 meshes. The total amount of these meshes is of order a million. Biological motivation : Symmetry is an important feature of protein tertiary and quaternary structures that has been associated with protein folding, function, evolution, and stability. Its emergence and ensuing prevalence has been attributed to gene duplications, fusion events, and subsequent evolutionary drift in sequence. Methods to detect these symmetries exist, either based on the structure or the sequence of the proteins, however, we believe that they can be vastly improved.
  • Data: Synthetic data are obtained by ‘symmetrizing’ folds from top8000 library (http://kinemage.biochem.duke.edu/databases/top8000.php).
  • References:: Our previous 3D CNN: [69] Invariance of CNNs (and references therein): 01630265/document, [70]
  • Basic algorithm: A prototype has already been created using the Tensorflow framework [4], which is capable of detecting the order of cyclic structures with about 93% accuracy. The main goal of this internship is to optimize the topology of the current neural network prototype and make it rotational and translational invariant with respect to input data. [4] [71]
  • Solution: The network architecture needs to be modified according to the invariance properties (most importantly, rotational invariance). Please see the links below [72],

[73] The code is written using the Tensorflow library, and the current model is trained on a single GPU (Nvidia Quadro 4000)of a desktop machine.

  • Novelty: Applications of convolutional networks to 3D data are still very challenging due to large amount of data and specific requirements to the network architecture. More specifically, the models need to be rotationally and transnationally invariant, which makes classical 2D augmentation tricks loosely applicable here. Thus, new models need to be developed for 3D data.
  • Authors: Expert Sergei Grudinin, consultants Guillaume Pages, Strizhov V.V.

Task 10

  • Name: Semi-supervised representation learning with attention
  • Task: training of vector representations using the attention mechanism, thanks to which the quality of machine translation has increased significantly. It is proposed to use it in the encoder-decoder architecture network to obtain vectors of text fragments of arbitrary length.
  • Data: It is proposed to consider two samples: Microsoft Paraphrase Corpus (a small set of proposals, https://www.microsoft.com/en-us/download/details.aspx?id=52398) and PPDB (a set of short segments, not always correct markup. http://sitem.herts.ac.uk/aeru/ppdb/en/)
  • References::

1. Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, Illia Polosukhin. Attention Is All You Need (https://arxiv.org/abs/1706.03762). 2. John Wieting, Mohit Bansal, Kevin Gimpel, Karen Livescu. Towards Universal Paraphrastic Sentence Embeddings (https://arxiv.org/abs/1511.08198). 3. Ryan Kiros, Yukun Zhu, Ruslan Salakhutdinov, Richard S. Zemel, Antonio Torralba, Raquel Urtasun, Sanja Fidler. Skip Thought Vectors (https://arxiv.org/abs/1506.06726). 4. Keras seq2seq (https://github.com/farizrahman4u/seq2seq).

  • Basic algorithm: solution [3] or vector representations obtained using seq2seq[].
  • Solution: in the task it is proposed to train vector representations for phrases using the attention and partial learning mechanism. As an internal quality functional, it is proposed to use the improved error function from [2]. As an applied problem, we can consider the problem of detecting paraphrases and sentiment analysis. Moreover, based on the results obtained in [1], it can be assumed that the attention mechanism has a greater influence on obtaining universal vectors for phrases than the network architecture. It is proposed to test this hypothesis using two different architectures - a standard recurrent and feed-forward network.
  • Novelty: new method.
  • Authors: Rita Kuznetsova, consultant

Task 11

  • Name: Selection of Interpreted Multi-Models in Credit Scoring Tasks
  • Task: The task of credit scoring is to determine the level of creditworthiness of the borrower. For this, a borrower's questionnaire is used, containing both numerical (age, income) and categorical features (gender, profession). It is required, having historical information about the repayment of loans by other borrowers, to determine whether the borrower will return the loan. The data can be heterogeneous (example, if there are different income regions in a country), and several models will be needed to adequately classify. It is necessary to determine the optimal number of models. Based on the set of model parameters, it is necessary to draw up a portrait of the borrower.
  • Data: It is proposed to consider five samples from the UCI and Kaggle repositories, with a capacity of 50,000 objects or more.
  • References:: A.A. Aduenko \MLAlgorithms\PhDThesis; C. Bishop, Pattern recognition and machine learning, final chapter; 20 years of Mixture experts.
  • Basic algorithm: Clustering and building independent logistic regression models, Adaboost, Decision Forest (with restrictions on complexity), Blend of Experts.
  • Solution: An algorithm is proposed for selecting a multi-model (a mixture of models or a mixture of Experts) and determining the optimal number of models.
  • Novelty: Proposed function of distance between models in which parameter distributions are given on different media.
  • Authors: Goncharov Alexey, Strizhov V.V.

Task 12

  • Name: Generation of features that are invariant to changes in the frequency of the time series.
  • Task: Informally: there is a set of time series of a certain frequency (s1), and the information we are interested in is distinguishable and at a lower sampling rate (in the example, the samples occur every millisecond, and the events of interest to us occur at an interval of 0.1 s). These series are integrated reducing the frequency by a factor of 10 (i.e. every 10 values are simply summed) and a set of time series s2 is obtained. be described in the same way. Formally: Given a set of time series s1, .., sNS with a high sampling rate 1. Target information (example, hand movement/daily price fluctuation/…) is distinguishable and at a lower sampling rate 2 < 1. It is necessary to find such a mapping f: S G, - the frequency of the series, that it will generate similar feature descriptions for series of different frequencies. Those.

f* = argminf E(f1(s1) -f2(s2)) , where E is some error function.

  • Data: Sets of time series of people's physical activity from accelerometers; human EEG time series; time series of energy consumption of cities/industrial facilities. Sample link: UCI repository, our EEG and accelerometer samples.
  • References:: See above for Accelerometers
  • Basic algorithm: Fourier transform.
  • Solution: Building an autoencoder with a partially fixed internal representation as the same time series with a lower frequency.
  • Novelty: For time series, there is no “common approach” to analysis, in contrast, in the example, to image analysis. If you look at the problem abstractly, now the cat is defined as well as and the cat, which takes up half the space in the image. An analogy with time series suggests itself. Moreover, the nature of data in pictures and in time series is similar: in pictures there is a hierarchy between values along two axes (x and y), and in time series - one at a time - along the time axis. The hypothesis is that methods similar to image analysis will provide qualitative results. The resulting feature representation can be further used for classification and prediction of time series.
  • Authors: R. G. Neichev, Strizhov V.V.

Task 18

YEAR

Group 594

Author Topic Link Consultant Reviewer Report Letters \Sigma=3+13
Goncharov Alexey (example) Metric classification of time series code,

paper, slides

Maria Popova Zadayanchuk Andrey BMF AILSBRCVTDSWH>
Belykh Evgeny Proskurin Alexander Classification of superpositions of movements of physical activity paper

slides code

Maria Vladimirova, Alexandra Malkova Romanenko Ilya, Popovkin Andrey, review

video

MF AILSBRC>V> [AILSBRC0VT0E0D0WS] CTD 2+9
Zueva Nadezhda Style Change Detection paper

slides video

Rita Kuznetsova Igashov Ilya, review BHMF AIL-S-B-R- [AILSBRCV0TE0D0WS] 3+10
Igashov Ilya Formulation and solution of an optimization problem combining classification and regression to estimate the binding energy of a protein and small molecules. paper

slides video

Sergei Grudinin, Maria Kadukova Manucharyan Vardan, review, correction BHMF AILBS+BRHC>V> [AILSBRCVTE0D0WS] 3+11
Kalugin Dmitry Graph Structure Prediction of a Neural Network Model paper

slides

Oleg Bakhteev Zueva Nadezhda review BHM AI-L-S--B0R0C0V0 [A-ILSBR0CVT0ED0WS] 2+11
Manucharyan Vardan Prediction of properties and types of atoms in molecular graphs using convolutional networks paper,

slides, code video

Sergei Grudinin, Maria Kadukova Fattakhov Artur review BMF AILS>B> [AILSB0R0CV0TE0D0WS] VED 3+7
Muraviev Kirill Determination of neural network parameters to be optimized. paper,

slides, code video

Oleg Bakhteev Kalugin Dmitry review BHMF A+IL-S-B-RCVTED [AILSBRCV0TE0DWS] 3+12
Murzin Dmitry Danilov Andrey Text recognition based on skeletal representation of thick lines and convolutional networks paper, slides, code

[video]

L. M. Mestetsky, Ivan Reyer, Zharikov I. N. Muraviev Kirill review BHMF A+IL> [AILSB0R0CV0TE0D0WS] 3+8
Popovkin Andrey Romanenko Ilya Creation of ranking models for information retrieval systems. Algorithm for Predicting the Structure of Locally Optimal Models paper

slides code video

Kulunchakov Andrey, Strizhov V.V. Proskurin Alexander, Belykh Evgeny, review BHMF AILS0BC>V> [AILSBRC0VTED0WS] 3+11
Fattakhov Artur Style Change Detection paper

slides code video

Rita Kuznetsova Danilov Andrey, Murzin Dmitry, review BMF AIL-S-B-R-CVTDSWH [AILSBRCVTE0D0WS] 3+11

Task 1 (1-2)

  • Name: Classification of superpositions of movements of physical activity
  • Task: Human behavior analysis by mobile phone sensor measurements: detect human movements from accelerometer data. The accelerometer data is a signal without precise periodicity, which contains an unknown superposition of physical models. We will consider the superposition of models: body + arm/bag/backpack.

Classification of human activities according to measurements of fitness bracelets. According to the measurements of the accelerometer and gyroscope, it is required to determine the type of activity of the worker. It is assumed that the time series of measurements contain elementary movements that form clusters in the space of time series descriptions. (Development: The characteristic duration of movement is seconds. Time series are marked with activity type marks: work, rest. The characteristic duration of activity is minutes. It is required to restore the type of activity by the description of the time series and cluster.)

  • Data:
  • References::
    • Karasikov M. E., Strizhov V. V. Classification of time series in the space of parameters of generating models // Informatics and its applications, 2016. [URL]
    • Kuznetsov M.P., Ivkin N.P. Algorithm for classification of accelerometer time series by combined feature description // Machine Learning and Data Analysis. 2015. T. 1, No. 11. C. 1471-1483. [URL]
    • Isachenko R. V., Strizhov V. V. Metric learning in Tasks of multiclass classification of time series // Informatics and its applications, 2016, 10(2): 48-57. [URL]
    • Zadayanchuk A.I., Popova M.S., Strizhov V.V. Choice of the optimal model for classifying physical activity based on accelerometer measurements // Information technologies, 2016. [URL]
    • Motrenko A.P., Strijov V.V. Extracting fundamental periods to segment human motion time series // Journal of Biomedical and Health Informatics, 2016, Vol. 20, no. 6, 1466-1476. [URL]
    • Ignatov A., Strijov V. Human activity recognition using quasiperiodic time series collected from a single triaxial accelerometer // Multimedia Tools and Applications, 2015, 17.05.2015 : 1-14. [URL]
  • Basic algorithm: Basic algorithm is described in [Karasikov, Strizhov: 2016] and [Kuznetsov, Ivkin: 2014].
  • Solution: Find the optimal segmentation method and optimal description of the time series. Construct a metric space of descriptions of elementary motions.
  • Novelty: A method for classifying and analyzing complex movements is proposed (Development: Connection of two characteristic times of a description of a person's life, combined problem statement.)
  • Authors: Alexandra Malkova, Maria Vladimirova, R. G. Neichev, Strizhov V.V.

Task 2 (1)

Task 3 (1-2)

  • Name: Text recognition based on skeletal representation of thick lines and convolutional networks
  • Task: It is required to build two CNNs, one recognizes a bitmap representation of an image, the other a vector one. (Development: generation of thick lines by neural networks)
  • Data: Bitmap fonts.
  • References:: List of works [74], in particular arXiv:1611.03199 and
  • Basic algorithm: Convolution network for bitmap.
  • Solution: It is required to propose a method for collapsing graph structures, which allows generating an informative description of the skeleton of a thick line.
  • Novelty: A way to improve the quality of recognition of thick lines due to a new way of generating their descriptions is proposed.
  • Authors: L. M. Mestetsky, I. A. Reyer, Strizhov V.V.

Task 4 (1-2)

  • Name: Creation of ranking models for information retrieval systems. Algorithm for Predicting the Structure of Locally Optimal Models
  • Task: It is required to predict a time series using some parametric superposition of algebraic functions. It is proposed not to cost the prognostic model, but to predict it, that is, to predict the structure of the approximating superposition. A class of considered superpositions is introduced, and on the set of such structural descriptions, a search is made for a locally optimal model for the problem under consideration. The task consists in 1) searching for a suitable structural description of the model 2) describing the search algorithm for the structure that will correspond to the optimal model 3) describing the algorithm for inverse construction of the model according to its structural description. For an already existing example of the answer to questions 1-3, see the works of A. A. Varfolomeeva.
  • Data:
    • Collection of text documents TREC (!)
    • A set of time series, which implies the restoration of functional dependencies. It is proposed to first use synthetic data or immediately apply the algorithm to forecasting time series 1) electricity consumption 2) physical activity with subsequent analysis of the resulting structures.
  • References::
    • (!) Kulunchakov A.S., Strijov V.V. Generation of simple structured Information Retrieval functions by genetic algorithm without stagnation // Expert Systems with Applications, 2017, 85: 221–230.
    • A. A. Varfolomeeva Selection of features when marking up bibliographic lists using structural learning methods, 2013, [75]
    • Bin Cao, Ying Li and Jianwei Yin Measuring Similarity between Graphs Based on the Levenshtein Distance, 2012, [76]
  • Basic algorithm: Specifically, there is no basic algorithm for the proposed problem. It is proposed to try to repeat the experiment of A.A. Varfolomeeva for a different structural description in order to understand what is happening.
  • Solution: The superposition of algebraic functions defines an ortree, on the vertices of which the labels of the corresponding algebraic functions or variables are given. Therefore, the structural description of such a superposition can be its DFS-code. This is a string consisting of vertex labels, written in the order in which the tree is traversed by depth-first search. Knowing the arities of the corresponding algebraic functions, we can restore any such DFS-code in O(n) and get back the superposition of functions. On the set of similar string descriptions, it is proposed to search for the string description that will correspond to the optimal model.
  • Authors: Kulunchakov Andrey, Strizhov V.V.

Task 5 (1)

  • Name: Definition of neural network parameters to be optimized.
  • Task: The task of neural network optimization is considered. It is required to divide the model parameters into two groups:
    • a) Model parameters to be optimized
    • b) Model parameters whose optimization has been completed. Further optimization of these parameters will not improve the quality of the model.

It is proposed to consider the optimization of parameters as a stochastic process. Based on the history of the process, we find those parameters whose optimization is no longer required.

  • Data: A selection of handwritten MNIST digits
  • Basic algorithm: Random choice of parameters.
  • References::
    • [77] SGD as a stochastic process.
    • [78] Variational inference in neural networks.
  • Novelty: The resulting algorithm will significantly reduce the computational cost of optimizing neural networks. A possible further development of the method is to obtain estimates for the parameters of the network obtained from the original operations of expansion, compression, adding and removing layers.
  • Authors: Oleg Bakhteev, Strizhov V.V.

Task 6 (1)

  • Name: Prediction of the graph structure of the neural network model.
  • Task: The Task is considered to find a stable (and non-redundant in terms of parameters) structure of a convolutional neural network. It is proposed to predict the structure of a neural network using doubly-recurrent neural networks. As a training sample, it is proposed to use the structures of models that have shown good quality on subsamples of small power.
  • Data: Samples MNIST, CIFAR-10
  • Basic algorithm: random search. Comparison with work on reinforcement learning is possible.
  • References::
    • [79] doubly-recurrent neural networks.
    • [80] Similar approach using reinforcement learning.
  • Authors: Oleg Bakhteev, Strizhov V.V.

Task 7 (1)

PAN 2017 (http://pan.webis.de/clef17/pan17-web/author-identification.html) PAN 2016 (http://pan.webis.de/clef16/pan16-web/author-identification.html)

  • References::

1. Ian Goodfellow. NIPS 2016 Tutorial: Generative Adversarial Networks (https://arxiv.org/pdf/1701.06547.pdf) 2. Jiwei Li, Will Monroe, Tianlin Shi, Sebastien Jean, Alan Ritter and Dan Jurafsky. Adversarial Learning for Neural Dialogue Generation(https://arxiv.org/pdf/1701.06547.pdf) 3. M. Kuznetsov, A. Motrenko, R. Kuznetsova, V. Strijov. Methods for Intrinsic Plagiarism Detection and Author Diarization 4. K. Safin, R. Kuznetsova. Style Breach Detection with Neural Sentence Embeddings (https://pdfs.semanticscholar.org/c70e/7f8fbc561520accda7eea2f9bbf254edb255.pdf)

  • Basic algorithm: solution described in [3, 4].
  • Solution: is proposed to solve the problem using generative adversarial networks — the generative model generates texts in the same author's style, the discriminative model — a binary classifier.
  • Novelty: it is assumed that the solution of this problem by the proposed method can give an increase in quality compared to typical methods for solving this problem, as well as related clustering problems of the authors.
  • Authors: Rita Kuznetsova (consultant), Strizhov V.V.

Task 8 (1)

  • Name: Obtaining likelihood estimates using autoencoders
  • Task: it is assumed that the objects under consideration obey the manifold hypothesis (manifold learning) - high-dimensional vectors are concentrated around some subspace of lower dimension. Works [1, 2] show that some modifications of autoencoders are looking for a k-dimensional manifold in the object space, which most fully conveys the data structure. In [2], an estimate of the probability density of data is derived using an autoencoder. It is required to obtain this estimate for the plausibility of the model.
  • Data: it is proposed to experiment on short text fragments of Google ngrams (http://storage.googleapis.com/books/ngrams/books/datasetsv2.html)
  • References::
    1. Pascal Vincent, Hugo Larochelle, Isabelle Lajoie, Yoshua Bengio, Pierre-Antoine Manzagol. Stacked Denoising Autoencoders: Learning Useful Representations in a Deep Network with a Local Denoising Criterion (http://www.jmlr.org/papers/volume11/vincent10a/vincent10a.pdf).
    2. Guillaume Alain, Yoshua Bengio. What Regularized Auto-Encoders Learn from the Data Generating Distribution (https://arxiv.org/pdf/1211.4246.pdf)
    3. Hanna Kamyshanska, Roland Memisevic. The Potential Energy of an Autoencoder (https://www.iro.umontreal.ca/~memisevr/pubs/AEenergy.pdf)
  • Basic algorithm:
  • Solution: in the problem it is proposed to train vector representations for phrases (n-grams) using an autoencoder, using Theorem 2 in [2] to obtain an estimate for the likelihood of the sample and, using this estimate, derive the likelihood of the model . Using the estimates obtained, one can also consider the sampling process.
  • Novelty: obtaining data and model likelihood estimates, generating texts using the resulting estimates.
  • Authors: Rita Kuznetsova (consultant).

Task 9 (1)

  • Name: Predict properties and types of atoms in molecular graphs using convolutional networks.
  • Task: Multilabel classification using convolutional neural networks (CNN) on graphs.

To predict the interaction of molecules with each other, it is often necessary to correctly describe their constituent atoms by assigning certain types to them. For small molecules, not many descriptors are available: the coordinates and chemical elements of atoms, the lengths of bonds and the magnitude of the angles between them. Using these features, we successfully predict atomic hybridizations and bond types. In this approach, each atom is considered "individually", the information about neighboring atoms necessary to determine the type of an atom is practically not used, and the types of atoms are determined by checking a large number of conditions. At the same time, molecules are represented as 3D molecular graphs, and it would be interesting to use this to predict their types with machine learning methods, for example, using CNNs. It is necessary to predict the types of vertices and edges of molecular graphs:

    • atom type (graph vertex type, about 150 classes),
    • atom hybridization (auxiliary feature, vertex type, 4 classes),
    • connection type (auxiliary feature, edge type, 5 classes).

The type of an atom (graph vertex) is based on information about its hybridization and the properties of neighboring atoms. Therefore, in the case of a successful solution of the classification problem, clustering can be carried out to find other ways to determine the types of atoms.

  • Data: About 15 thousand molecules represented as molecular graphs. For each vertex (atom), 3D coordinates and a chemical element are known. Additionally, bond lengths, angles and dihedral angles between atoms (3D graph coordinates), binary signs reflecting whether an atom is included in the cycle and whether it is terminal are calculated. The sample is labeled, but the labeled data may contain ~5% errors.

If there is not enough data, it is possible to increase the sample (up to 200 thousand molecules), associated with an increase in inaccuracies in labeling.

  • References::
  • Basic algorithm: Prediction of hybridizations and link orders using a multiclass non-linear SVM with a small number of descriptors. https://hal.inria.fr/hal-01381010/document
  • Solution: Proposed solution to the problem and ways of conducting research.

Methods for presenting and visualizing data and conducting error analysis, analyzing the quality of the algorithm. At the first stage, it will be necessary to determine the operations on the graphs necessary to build the network architecture. Next, you will need to train the network for multi-class classification of the types of vertices (and edges) of the input graph. To assess the quality of the algorithm, it is supposed to evaluate the accuracy using cross-validation. For the final publication (in a specialized journal), it will be necessary to make a specific test for the quality of predictions: based on the predicted bond types, the molecule is written as a string (in SMILES format) and compared with a sample. In this case, for each molecule, the prediction will be considered correct only if the types of all bonds in it were predicted without errors.

  • Novelty: The proposed molecular graphs have a 3D structure and internal hierarchy, making them an ideal CNN application.
  • Authors: Sergei Grudinin, Maria Kadukova, Strizhov V.V.

Task 10 (1)

  • Name: Formulation and solution of an optimization problem combining classification and regression to estimate the binding energy of a protein and small molecules. Task description [84]
  • Task:

From the point of view of bioinformatics, the Task is to estimate the free energy of protein binding to a small molecule (ligand): the best ligand in its best position has the \textbf{lowest free energy} of interaction with the protein. (Following a large text, see the file at the link above.)

  • Data:
    • Data for binary classification.

Approximately 12,000 protein-ligand complexes: for each of them there is 1 native position and 18 non-native ones. The main descriptors are histograms of distributions of distances between different atoms of the protein and ligand, the dimension of the vector of descriptors is ~ 20,000. In the case of continued research and publication in a specialized journal, the set of descriptors can be expanded. The data will be provided as binary files with a python script to read.

    • Data for regression.

For each of the presented complexes, the value of the quantity is known, which can be interpreted as the binding energy.

  • References::
  • Basic algorithm: [88]

In the classification problem, we used an algorithm similar to linear SVM, whose relationship with the energy estimate, which is outside the scope of the classification problem, is described in the above article. Various loss functions can be used in a regression problem.

  • Solution: It is necessary to connect the previously used optimization problem with the regression problem and solve it using standard methods. Cross-validation will be used to check the operation of the algorithm.

There is a separate test set consisting of (1) 195 complexes of proteins and ligands, for which it is necessary to find the best ligand pose (the algorithm for obtaining ligand positions differs from that used in training), (2) complexes of proteins and ligands, for which native poses it is necessary to predict the energy binding, and (3) 65 proteins for which the most strongly binding ligand is to be found.

  • Novelty:' First of all, the interest is combining classification and regression problems.

The correct assessment of the quality of protein and ligand binding is used in drug development to search for molecules that interact most strongly with the protein under study. Using the classification problem described above to predict the binding energy results in an insufficiently high correlation of predictions with experimental values, while using the regression problem alone leads to overfitting.

  • Authors Sergei Grudinin, Maria Kadukova, Strizhov V.V.

2017

Author Topic Link Consultant Reviewer Report Letters
Goncharov Alexey (example) Metric classification of time series code,

paper, slides

Maria Popova Zadayanchuk Andrey BMF AILSBRCVTDSWH>
Alekseev Vasily Intratext coherence as a measure of interpretability of thematic models of text collections code

data paper slides video

Viktor Bulatov Zakharenkov Anton BMF AILSB+RC+V+TDHW
Anikeev Dmitry Local approximation of time series for building predictive metamodels code

paper slides

Strizhov V.V. Смердов Антон BMF AILS>B0R0C0V0T0D0H0W0
Gasanov Elnur Construction of an approximating description of a scalogram in the problem of predicting movements using an electrocorticogram code paper

slides

Anastasia Motrenko Kovalev Dmitry BMF AILSBRCVTDH0W0
Zakharenkov Anton Massively multitask deep learning for drug discovery code

paper slides video

Maria Popova Alekseev Vasily BMF AILSBRCVT>D>H0W0
Kovalev Dmitry Unsupervised representation for molecules code

paper slides

Maria Popova Gasanov Elnur BMF AILSBRCVT>D>H0W0
Novitsky Vasily Feature Selection in Problems of Autoregressive Prediction of Biomedical Signals paper

code slides

Alexander Katrutsa B - F AILS>B0R0C0V0T0D0H0W0
Selezneva Maria Aggregation of heterogeneous text collections in a hierarchical thematic model of Russian-language popular science content paper

code slides video

Irina Efimova Шолохов Алексей BMF A+IL+SBRCVTDHW
Smerdov Anton Choosing the optimal recurrent network model in the Paraphrase Search Tasks paper

code slides video

Oleg Bakhteev Dmitry Anikeev BMF AIL+SB+RC>V+M-T>D0H0W0
Uvarov Nikita Optimal Algorithm for Reconstruction of Dynamic Models paper

slides code video

Yuri Maksimov BMF AILS0B0R0C0V0T0D0H0W0
Usmanova Karina Multiple Manifold Learning (Joint diagonalization for 3D shapes - AJD on Hessian matrices) paper

slides code video

Mikhail Karasikov Innokenty Shibaev BMF AILSBRC+VT+EDH>W
Innokenty Shibaev Convex relaxations for multiple structure alignment (synchronization problem for SO(3)) paper

slides code video

Mikhail Karasikov Usmanova Karina BMF AILS-BRCVT>D>H>W
Sholokhov Alexey Noise immunity of methods for informational analysis of ECG signals

paper code slides video

Vlada Bunakova Selezneva Maria BMF AILSBRCVTDHW

Академ или новые

Author Topic Link Consultant Reviewer Report Letters
Kulkov Alexander Adaptive relaxations of NP hard problems through machine learning paper Yuri Maksimov академ A>I>L>B0R0C0V0T0D0H0W0
Kaloshin Pavel Using deep learning networks to transfer classification models in case of insufficient data.

paper code data

Anton Khritankov - MF AIL-SBRC-VT+D>H>W0
Malinovsky Grigory Choice of Interpreted Multimodels in Credit Scoring Tasks paper

code

Alexander Aduenko академ B - - AILS-B>R>C>V>T0D0H0W0
Pletnev Nikita Internal plagiarism detection paper Rita Kuznetsova академ - - - A-I-L-S>B0R0C0V0T0D0H0W0
Grevtsev Alexander Parallel Algorithms for Parametric Identification of the Tersoff Potential for AlN

paper

Karine Abgaryan
Zaitsev Nikita Automatic classification of scientific articles on crystallography

paper readme

Evgeny Gavrilov
Diligul Alexander Determination of the optimal potential parameters for the Rosato-Guillope-Legrand (RGL) model from experimental data and the results of quantum mechanical calculations

paper

Karine Abgaryan
Daria Fokina Selection of Candidates in the Problem of Finding Text Borrowings with Paraphrasing Based on the Vectorization of Text Fragments Alexey Romanov AILSB0R0C0V0T0D0H0W0

Task 1

  • Name: Classification of human activities according to fitness bracelet measurements.
  • Task: According to the accelerometer and gyroscope measurements, it is required to determine the type of worker's activity. It is assumed that the time series of measurements contain elementary movements that form clusters in the space of time series descriptions. The characteristic duration of the movement is seconds. Time series are labeled with activity type labels: work, leisure. The typical duration of activity is minutes. It is required to restore the type of activity according to the description of the time series and cluster.
  • Data: WISDM accelerometer time series (Time series (examples library), Accelerometry section).
  • References::
    • Karasikov M.E., Strizhov V.V. Classification of time series in the space of parameters of generating models // Informatics and its applications, 2016. [URL]
    • Kuznetsov M.P., Ivkin N.P. Algorithm for Classifying Accelerometer Time Series by Combined Feature Description // Machine Learning and Data Analysis. 2015. V. 1, No. 11. C. 1471 - 1483. [URL]
    • Isachenko R.V., Strizhov V.V. Metric learning in Taskx multiclass classification of time series // Informatics and its applications, 2016, 10(2) : 48-57. [URL]
    • Zadayanchuk A.I., Popova M.S., Strizhov V.V. Choosing the optimal model for classifying physical activity based on accelerometer measurements // Information technologies, 2016. [URL]
    • Motrenko A.P., Strijov V.V. Extracting fundamental periods to segment human motion time series // Journal of Biomedical and Health Informatics, 2016, Vol. 20, no. 6, 1466 - 1476.
    • Ignatov A., Strijov V. Human activity recognition using quasiperiodic time series collected from a single triaxial accelerometer // Multimedia Tools and Applications, 2015, 17.05.2015 : 1-14. [URL]
  • Basic algorithm: Basic algorithm is described in [Karasikov, Strizhov: 2016] and [Kuznetsov, Ivkin: 2014].
  • Solution: Find the optimal segmentation method and optimal description of the time series. Construct a metric space of descriptions of elementary motions.
  • Novelty:: Connection of two characteristic times of the description of a person's life, combined statement of the problem.
  • Authors: Strizhov V.V., M.P. Kuznetsov, P.V. Levdik.

Task 2

  • Name: Construction of an approximating description of a scalogram in the problem of predicting movements using an electrocorticogram.
  • Task: As part of solving the problem of decoding ECoG signals, the Task of classifying movements by time series of electrode readings is solved. The tools for extracting features from ECoG time series are the coefficients of the wavelet transform of the signal under study [Makarchuk 2016], on the basis of which a scalogram is built for each electrode - a two-dimensional array of features in frequency-time space. Combining scalograms for each electrode gives signs of a time series in the spatio-frequency-time domain. The feature description constructed in this way obviously contains multicorrelated features and is redundant. It is required to propose a method for reducing the dimension of the feature space.
  • Data: Measurements of the positions of the fingers when performing simple gestures. Description of experiments data.
  • References::
    • Makarchuk G.I., Zadayanchuk A.I. Strijov V.V. 2016. Using partial least squares to decode hand movement using ECoG cues in monkeys. pdf
    • Karasikov M.E., Strizhov V.V. Classification of time series in the space of parameters of generating models // Informatics and its applications, 2016. [URL]
    • Kuznetsov M.P., Ivkin N.P. Algorithm for Classifying Accelerometer Time Series by Combined Feature Description // Machine Learning and Data Analysis. 2015. T. 1, No. 11. C. 1471 - 1483.
  • Basic algorithm: PLS

Chen C, Shin D, Watanabe H, Nakanishi Y, Kambara H, et al. (2013) Prediction of Hand Trajectory from Electrocorticography Signals in Primary Motor Cortex. PLoS ONE 8(12): e83534.

  • Solution: To reduce the dimension, it is proposed to use the local approximation method proposed in [Kuznetsov 2015] used to classify accelerometric time series [Karasikov 2016].
  • Novelty: A new method of movement recovery based on electrocorticograms is proposed.
  • Authors: Strizhov V.V., A.P. Motrenko

Task 3

  • Name: Multiple Manifold Learning (Joint diagonalization for 3D shapes - AJD on Hessian matrices).
  • Task: Building an optimal algorithm for the Multiple Manifold Learning task. Two protein conformations (two tertiary structures) are given. In the vicinity of each state, a model of an elastic body is specified (oscillations of the structure in the vicinity of these states). The task is to build a general model of an elastic body to find intermediate states with the maximum match with these models in the vicinity of given conformations. The space of motion of an elastic body is given by the Hessian eigenvectors. It is required to find a common low-rank approximation of the space of motions of two elastic bodies.
  • Data: Protein structures in double conformations from PDB, about 100 sets from the article https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4677049/
  • References:: A list of scientific papers, supplemented by 1) the statement of the problem being solved, 2) links to new results (a recent article that is close in results), 3) basic information about the problem under study.

Tirion, M. M. (1996). Large amplitude elastic motions in proteins from a single-parameter, atomic analysis. Physical Review Letters, 77(9), 1905. Moal, I. H., & Bates, P. A. (2010). {SwarmDock} and the Use of Normal Modes in Protein-Protein Docking. IJMS, 11(10), 3623–3648. https://doi.org/10.3390/ijms11103623

  • Basic algorithm: AJD algorithm: http://perso.telecom-paristech.fr/~cardoso/jointdiag.html, AJD algorithms implemented as part of Shogun ML toolbox http://shogun-toolbox.org , http://shogun-toolbox.org/api/latest/classshogun_1_1CApproxJointDiagonalizer.html.
  • Solution: Computing Hessians (C++ code from Sergey), learning and running standard joint diagonalization algorithms for the first n non-trivial eigenvectors, analyzing loss functions, adapting the standard algorithm to solve the original problem.
  • Novelty: Using simple elasticity models with one or more free parameters, thermal fluctuations in proteins can be described. However, such models do not describe transitions between several stable conformations in proteins. The purpose of this work is to refine the elastic model so that it also describes the space of conformational changes.
  • Authors: Sergey Grudinin, consultant: Mikhail Karasikov / Yury Maksimov.

Task 4

  • Name: Convex relaxations for multiple structure alignment (synchronization problem for SO(3)).
  • Task: Find transformations to align protein tertiary structures simultaneously (in simple words: find orthogonal transformations that align data in R^3 molecules that have the same chemical formula). If the structures are the same (the RMSD is equal to zero after alignment, the structures are aligned exactly), then you can align in pairs. However, if this is not the case, then the Basic algorithm, generally speaking, does not find the optimum of the original problem with a loss function for simultaneous equalization.
  • Data: Protein structures in PDB format in various states and coordinate systems.
  • References::
    • Multiple structural alignment:
      1. Kearsley.S.K. (1990)7. Comput. Chem., 11, 1187-1192.
      2. Shapiro., BothaJ.D., PastorA and Lesk.A.M. (1992) Acta Crystallogr., A48, 11-14.
      3. Diamond,R. (1992) Protein Sci., 1, 1279-1287.
      4. May AC, Johnson MS, Improved genetic algorithm-based protein structure comparisons: pairwise and multiple superpositions. ProteinEng. 1995 Sep;8(9):873-82.
    • Synchronization problem:
      1. O. Özyeşil, N. Sharon, A. Singer, ``Synchronization over Cartan motion groups via contraction”, Available at arXiv.
      2. L. Wang, A. Singer, `ʻExact and Stable Recovery of Rotations for Robust Synchronization”, Information and Inference: A Journal of the IMA, 2(2), pp. 145--193 (2013).
      3. Semidefinite relaxations for optimization problems over rotation matrices J Saunderson, PA Parrilo… - Decision and Control ( …, 2014 - ieeexplore.ieee.org
      4. Spectral synchronization of multiple views in SE (3) F Arrigoni, B Rossi, A Fusiello - SIAM Journal on Imaging Sciences, 2016 - SIAM
      5. Robust Rotation Synchronization via Low-rank and Sparse Matrix Decomposition, F Arrigoni, A Fusiello, B Rossi, P Fragneto - arXiv preprint arXiv: …, 2015 - arxiv.org
    • Spectral relaxation for SO(2)
      1. A. Singer, Angular synchronization by eigenvectors and semidefinite programming, Applied and Computational Harmonic Analysis 30 (1) (2011) 20 – 36.
    • Spectral relaxation for SO(3)
      1. M.Arie-Nachimson,S.Z.Kovalsky,I.Kemelmacher-Shlizerman,A.Singer,R.Basri,Global motion estimation from point matches, in: International Conference on 3D Imaging, Modeling, Processing, Visualization and Transmission, 2012 , pp. 81–88.
      2. A. Singer, Y. Shkolnisky, Three-dimensional structure determination from common lines in cryo-em by eigenvectors and semidefinite programming, SIAM Journal on Imaging Sciences 4 (2) (2011) 543–572.
  • Basic algorithm: Local (pairwise) alignment algorithm. Kearsley S.K. (1989) Acta Crystallogr., A45, 208-210; Rapid determination of RMSDs corresponding to macromolecular rigid body motions

Petr Popov, Sergei Grudinin, Journal of Computational Chemistry, Wiley, 2014, 35(12), pp.950-956. <10.1002/jcc.23569> DOI: 10.1002/jcc.23569

  • Solution: Two options for setting optimization problems (through rotation matrices and through quaternions). Relaxation of the obtained problems by convex ones, comparison of the solutions of the problem by the basic algorithm and relaxations (spectral relaxation, SDP).
  • Novelty: A method that flattens structures by minimizing the loss function, taking into account all pairwise losses.
  • Authors: Sergey Grudinin, consultant: Mikhail Karasikov.

Task 5

  • Name: Local approximation of time series for building predictive metamodels.
  • Task: The physical activity of a person is investigated by time series - accelerometer measurements. The aim of the project is to create a tool for analyzing the problem of creating models for predicting models - metamodels. The segment of the time series is investigated. It is required to predict the class of the segment. (Option: predict the end of the segment, the next segment, its class. In this case, the class of the next segment may differ from the class of the previous one).
  • Data: Based on a Santa Fe or WISDM sample (samples consist of segments with many elementary movements and class labels corresponding to the segments), a variant of the OPPORTUNITY Activity Recognition Challenge.
  • References::
    • Karasikov M.E., Strizhov V.V. Classification of time series in the space of parameters of generating models // Informatics and its applications, 2016. [URL]
    • Kuznetsov M.P., Ivkin N.P. Algorithm for Classifying Accelerometer Time Series by Combined Feature Description // Machine Learning and Data Analysis. 2015. V. 1, No. 11. C. 1471 - 1483. [URL]
  • Basic algorithm: [Karasikov 2016]
  • Solution: See task description.
  • Novelty: When creating meta-prognostic models (predictive models of predictive models), the problem of using the values of parameters of local models when creating meta-models remains open. The purpose of the project below is to create a tool to analyze this problem.
  • Authors: Strijov V.V.

Task 6

  • Name: Choosing the optimal recurrent network model in the Paraphrase Search Tasks
  • Task: Given a selection of pairs of sentences labeled <<similar>> and <<dissimilar>>. It is required to build a recurrent network of low complexity (that is, with a small number of parameters) that delivers a minimum error in the classification of pairs of sentences.
  • Data: It is proposed to consider two samples: Microsoft Paraphrase Corpus (a small set of sentences) and [http ://sitem.herts.ac.uk/aeru/ppdb/en/ PPDB] (set of short segments, markup not always correct)
  • References::
    • [1] Step by step description of the implementation of the LSTM recurrent network
    • [2] Thinning algorithm based on building a network with a minimum description length
    • [3] Optimal Brain Damage
  • Basic algorithm: The basic algorithm can be:
    1. Solution without thinning
    2. Solution described in [3]
    3. Optimal Brain Damage
  • Solution: It is proposed to consider the thinning method described in [3] with a block covariance matrix: either neurons or parameters grouped by input features act as blocks.
  • Novelty: The proposed method will effectively reduce the complexity of the recurrent network, taking into account the relationship between neurons or input features.
  • Authors: Oleg Bakhteev, consultant

Task 7

  • Name: Internal plagiarism detection
  • Task: Solved by Task to identify internal borrowings in text. It is required to test the hypothesis that the given text was written by a single author, and if it is not fulfilled, highlight the borrowed parts of the text. A borrowing is a part of the text, presumably written by another author and containing characteristic differences from the style of the main author. It is required to develop such a style function that allows to distinguish with a high degree of certainty the style of the main author of the text from borrowings.
  • Data: It is proposed to consider the corpus PAN-2011, PAN-2016
  • References::
    • [1] Step by step description of the implementation of the LSTM recurrent network
    • [2] Author clustering algorithm
    • [3] Statistical Language Models Based on Neural Networks
    • [4] Methods for intrinsic plagiarism detection and author diarization
  • Basic algorithm: The solution described in [4] can be used as the basic algorithm.
  • Solution: It is proposed to consider the method described in [2] and build a style function based on the neural network outputs.
  • Novelty: It is assumed that the construction of a style function by the proposed method can give an increase in quality compared to typical solutions to this problem.
  • Authors: Rita Kuznetsova, consultant

Task 8

  • Name: Adaptive relaxations of NP hard problems through machine learning
  • Task: Modern problems of optimizing power flows in power networks lead to non-convex optimization tasks with a large number of restrictions. Statements similar in structure also arise in a number of other engineering problems and in classical tasks of combinatorial optimization. The traditional approach to solving such NP hard problems is to write their convex relaxations (semidefinite/SDP, second order conic/SOCP, etc), which usually have a much larger set of feasible solutions than in the original problem. and by the subsequent projection of the obtained solution into the region where the constraints of the original problem are satisfied. In many practical cases, the quality of the solution obtained in this way is not high. Alternative approaches, for example MILP (mixed integer linear programming) relaxation, are substantially more time consuming but result in a more accurate answer.

The main problem is the impossibility of using known methods for solving large-scale problems (networks of 1000 nodes and more). One of the key obstacles is not so much the dimension of the problem as a large number of restrictions. At the same time, in real Tasks it is possible to single out a small set of restrictions such that the sets of admissible points in the selected set and in the original one are very close. This will allow us to replace the task with another one with fewer restrictions, which will increase the speed of the algorithms used. It is proposed to use machine learning methods to build the indicated set of the most important constraints.

  • References:: Sampling/machine learning methods:
    1. Beygelzimer, A., Dasgupta, S., & Langford, J. (2009, June). Importance weighted active learning. In Proceedings of the 26th annual international conference on machine learning (pp. 49-56). ACM.
    2. Tong, S., & Koller, D. (2001). Support vector machine active learning with applications to text classification. Journal of machine learning research, 2(Nov), 45-66.
    3. Owen, A., & Zhou, Y. (2000). Safe and effective importance sampling. Journal of the American Statistical Association, 95(449), 135-143.

Relaxations: Nagarajan, H., Lu, M., Yamangil, E., & Bent, R. (2016). Tightening McCormick Relaxations for Nonlinear Programs via Dynamic Multivariate Partitioning. arXiv preprint arXiv:1606.05806.

  • Data: ieee + matpower data containing descriptions of energy networks and their modes of operation.
  • Novelty: This approach seems to be the first application of applied statistics/machine learning methods to solve difficult optimization problems. We expect substantial gains in labor-intensive style methods
  • Author: consultant: Yuri Maksimov, Expert: Mikhail Chertkov

Task 9

  • Name: Optimal Algorithm for Reconstruction of Dynamic Models.
  • Task: A standard machine learning problem statement in the context of unsupervised learning assumes that the examples are independent and come from the same probability distribution. However, often observed data are of dynamic origin and are correlated. The task is to develop an efficient method for restoring a dynamic graphical model (graph and model parameters) from observed correlated dynamic configurations. This Task is theoretically important and has many applications. The basis of the algorithm will be the adaptation of a new optimal method of screening interactions (interaction screening), developed for the Ising model. The solution process will combine familiarity with computer science/machine learning theoretical methods and numerical experiments.
  • Data: Simulated dynamic configurations of spins in the kinetic Ising model.
  • References::
    1. Lokhov et al., "Optimal structure and parameter learning of Ising models", arXiv:1612.05024 (2016) {https://arxiv.org/abs/1612.05024}
    2. Vuffray et al., "Interaction screening: efficient and sample-optimal learning of Ising models", NIPS 2016 {https://arxiv.org/abs/1605.07252}
    3. Decelle and Zhang, "Inference of the sparse kinetic Ising model using the decimation method", Phys. Rev. E 2016 {https://arxiv.org/abs/1502.01660}
    4. Bresler et al., "Learning graphical models from the Glauber dynamics", Allerton 2014 {https://arxiv.org/abs/1410.7659}
    5. Zeng et al., "Maximum likelihood reconstruction for Ising models with asynchronous updates", Phys. Rev. Lett. 2013
  • Basic algorithm: Dynamic method for shielding interactions. Comparison with the maximum likelihood method.
  • Novelty: Currently, the optimal (ie using the minimum possible number of examples) algorithm for this problem is unknown. The dynamic method of interaction screening has a good chance of finally "closing" this task, because is optimal for a static problem.
  • Author: consultants Andrey Lokhov, Yuri Maksimov. Expert Mikhail Chertkov

Task 10

  • Name: Choice of Interpreted Multimodels in Credit Scoring Tasks
  • Task: The task of credit scoring is to determine the level of creditworthiness of the borrower. For this, a borrower's questionnaire is used, containing both numerical (age, income) and categorical features (gender, profession). It is required, having historical information about the repayment of loans by other borrowers, to determine whether the borrower will return the loan. The data can be heterogeneous (example, if there are different income regions in a country), and several models will be needed to adequately classify. It is necessary to determine the optimal number of models. Based on the set of model parameters, it is necessary to draw up a portrait of the borrower.
  • Data: It is proposed to consider five samples from the UCI and Kaggle repositories, with a capacity of 50,000 objects or more.
  • References:: A.A. Aduenko \MLAlgorithms\PhDThesis; C. Bishop, Pattern recognition and machine learning, final chapter; 20 years of Mixture experts.
  • Basic algorithm: Clustering and building independent logistic regression models, Adaboost, Decision Forest (with restrictions on complexity), Blend of Experts.
  • Solution: An algorithm is proposed for selecting a multi-model (a mixture of models or a mixture of Experts) and determining the optimal number of models.
  • Novelty: Proposed function of distance between models in which parameter distributions are given on different media.
  • Authors: A.A. Aduenko, Strizhov V.V.

Task 11

  • Name: Feature Selection in Problems of Autoregressive Prediction of Biomedical Signals.
  • Task: The task of predicting biomedical signals and IoT signals is being solved. It is required to predict the vector - the next few signal samples. It is assumed that the proper dimension of the space of both the predicted variable and the independent variable can be significantly reduced, thereby increasing the stability of the forecast without significant loss of accuracy. For this, the Partial Least Squares approach in autoregressive forecasting is used.
  • Data: SantaFe biomedical time series sample, IoT signal sample.
  • References:: Katrutsa A.M., Strijov V.V. Stresstest procedure for feature selection algorithms // Chemometrics and Intelligent Laboratory Systems, 2015, 142 : 172-183; : Katrutsa A.M., Strijov V.V. Comprehensive study of feature selection methods to solve multicollinearity problem according to evaluation criteria // Expert Systems with applications, 2017; Kee Siong Ng A Simple Explanation of Partial Least Squares keesiong.ng@gopivotal.com Draft, April 27, 2013, http://users.cecs.anu.edu.au/~kee/pls.pdf
  • Basic algorithm: PLS, quadratic optimization algorithm for feature selection.
  • Solution: build a design matrix with a suboptimal set of objects and features, propose a quadratic optimization error function (if possible, develop it for the case of a tensor representation of the design matrix).
  • Novelty: Generalized feature selection algorithm (published two weeks ago) for the PLS case.
  • Authors: A.M. Katrutsa, Strizhov V.V.

Task 12

  • Name: Massively multitask deep learning for drug discovery
  • Task: Develop a multi-task recurrent neural network to predict biological activity. For each molecule-protein pair, it is required to predict the binary value 0/1, which means that the molecule binds/does not bind to the protein.
  • Data: sparse biological activity data for ~100K molecules versus ~1000 proteins. Molecules are represented as SMILES strings (sequence of characters encoding a molecule)
  • References:: https://arxiv.org/pdf/1502.02072
  • Basic algorithm: multi-task neural network that predicts activity by numerical features, single-task recurrent neural network
  • Solution: Multitasking means that you need to build a model that is obtained for the input of a molecule and predicts its biological activity against all proteins in the sample.
  • Novelty: Existing methods did not show a significant improvement in the quality of the DL model compared to standard ML models
  • Authors: Expert -- Alexander Isaev, consultant -- Maria Popova

Task 13

  • Name: Unsupervised representation for molecules
  • Task: Develop an unsupervised method for representing molecules
  • Data: ~1.5M molecules in SMILES string format (character sequence encoding the molecule)
  • References:: https://www.cs.toronto.edu/~hinton/science.pdf
  • Basic algorithm: currently hand-selected numerical features are used as such representation. The quality of the resulting representations can be compared with the tox21 dataset (10K molecules versus 12 proteins)
  • Solution: use convolutional or recurrent networks to build an autoencoder.
  • Novelty: building an end-to-end model to get informative features
  • Authors: Expert -- Alexander Isaev, consultant -- Maria Popova

Task 14

  • Name: Intratext coherence as a measure of interpretability of thematic models of text collections.
  • Task: Interpretability is a subjective measure of the quality of topic models, as measured by Expert Scores. Coherence is a measure of the occurrence of thematic words, calculated automatically from the text and correlates well with interpretability, as shown in the Newman and Mimno series. The first Task is to evaluate the representativeness of the sequence of words in the text, according to which the coherence is estimated. The second Task is to compare several new methods for measuring interpretability and coherence based on the selection of the most representative sequence of words in the source text.
  • Data: A collection of popular science content PostNauka, a collection of news content.
  • References::
    1. Vorontsov K. V. Review of probabilistic thematic models, 2017.
    2. N.Aletras, M.Stevenson. Evaluating Topic Coherence Using Distributional Semantics, 2013.
    3. D. Newman et al. Automatic evaluation of topic coherence, 2010
    4. D.Mimno et al. Optimizing semantic coherence in topic models, 2011
    5. http://palmetto.aksw.org/palmetto-webapp/
  • Basic algorithm: Standard methods for estimating the interpretability and coherence of topics in topic models.
  • Solution: A new method for measuring interpretability and coherence, experiments to find the most correlated measures of interpretability and coherence, similar to [D.Newman, 2010].
  • Novelty: inline measures of interpretability and coherence were not previously proposed.
  • Authors: Vorontsov K. V.. consultants: Viktor Bulatov, Anna Potapenko, Artyom Popov.

Task 15

  • Name: Aggregation of heterogeneous text collections in a hierarchical thematic model of Russian-language popular science content.
  • Task: Implement and compare multiple ways of combining text collections from different sources into one hierarchical topic model. Build a classifier that determines the presence of a topic in the source.
  • Data: Collection of popular science content PostNauka, Wikipedia collection.
  • References::
    1. Vorontsov K. V. Review of probabilistic thematic models, 2017.
    2. Chirkova N. A, Vorontsov K. V. Additive regularization of multimodal hierarchical topic models // Machine Learning and Data Analysis, 2016. T. 2. No. 2.
  • Basic algorithm: An algorithm for constructing a thematic hierarchy in BigARTM, implemented by Nadezhda Chirkova. Marking tool
  • Solution: Build a topic model with source modalities and highlight topics specific to only one of the sources. Prepare a sample for training a classifier that determines the presence of a topic in the source.
  • Novelty: Additive regularization of topic models has not been applied to this problem before.
  • Authors: Vorontsov K. V.. consultants: Alexander Romanenko, Irina Efimova, Nadezhda Chirkova.

Task 16

  • Name: Application of the methods of symbolic dynamics in the technology of informational analysis of electrocardiosignals.
  • Task: The technology of informational analysis of electrocardiosignals, proposed by V.M.Uspensky, involves converting a raw signal into a character sequence and searching for disease patterns in this sequence. So far, symbolic n-grams have been predominantly used to search for patterns. In the framework of this work, it is proposed to expand the class of templates in which the search for diagnostic signs of diseases is performed. Quality criterion -- AUC and MAP ranking of diagnoses.
  • Data: A selection of electrocardiograms with known diagnoses.
  • References::
    1. Uspensky V.M. Informational function of the heart. Theory and practice of diagnosing diseases of internal organs by the method of information analysis of electrocardiosignals. - M .: "Economics and Information", 2008. - 116s
    2. Technology of information analysis of electrocardiosignals.
  • Basic algorithm: Classification methods .
  • Solution: Search for logical patterns in character strings, methods of character dynamics, comparison of algorithms according to the quality criteria AUC and MAP (diagnosis ranking).
  • Novelty: So far, character n-grams have been used predominantly to search for patterns.
  • Authors: Vorontsov K. V.. consultants: Vlada Tselykh.

Vorontsov tasks +

  • Title: Dynamic hierarchical thematic model of the news flow.
  • Task: Develop an algorithm for classifying topics in news flows into new and ongoing ones. Apply the obtained criteria for creating new topics at all levels of the topic model hierarchy when adding the next piece of data to the text collection (for example, all news for one day).
  • Data: Collection of news in Russian. A subsample of news classified into two classes: new and ongoing topics.
  • Literature:
    1. Vorontsov K.V. Review of probabilistic thematic models, 2017.
    2. Chirkova N. A, Vorontsov K. V. Additive regularization of multimodal hierarchical topic models // Machine Learning and Data Analysis , 2016 T. 2. No. 2.
  • Basic Algorithm: An algorithm for constructing a thematic hierarchy in BigARTM, implemented by Nadezhda Chirkova. Known Topic Detection & Tracking algorithms.
  • Solution: Using BigARTM, selecting regularizers and their parameters, using the topic selection regularizer. Building an algorithm for classifying topics into new and ongoing.
  • Novelty: Additive regularization of topic models has not been applied to this problem before.
  • Authors: KV Vorontsov. Consultants: Alexander Romanenko, Artyom Popov.

Task Antiplagiarism +

  • Name: Selection of Candidates in the Problem of Finding Text Borrowings with Paraphrasing Based on the Vectorization of Text Fragments.
  • Task: Searching for text borrowings in a collection of documents involves selecting a small set of candidates for subsequent detailed analysis. The Candidate Selection Task is formulated as finding the optimal ranking of documents in a collection for a query with respect to some function that is an estimate for the total length of borrows from a collection document to a query document.
  • Data: PAN
  • References::
    1. Romanov A.V., Khritankov A.S. Selection of candidates when searching for borrowings in a collection of documents in a foreign language .pdf
  • Basic algorithm: shingles method with reverse index construction.
  • Solution: Vectorization of text fragments (word embeddings + convolutional / recurrent neural networks) and subsequent search for nearest objects in a multidimensional metric space.
  • Novelty: a new approach to solving the problem.
  • Authors: Alexey Romanov (consultant)

Additional tasks

Vorontsov tasks +

  • Name: Thematic modeling of an economic sector based on bank transaction data.
  • Task: Test the hypothesis that a large sample of transactions between firms is adequately described by a relatively small set of economic activities (aka topics). Task is reduced to decomposing the matrix of transactional data "buyers × sellers" into the product of three non-negative matrices "buyers × topics", "topics × topics", "topics × sellers", while the middle matrix describes a directed graph of financial flows in the industry. It is required to compare several methods for constructing such expansions and find the number of topics for which the observed set of transactions is modeled with sufficient accuracy.
  • Data: selection of transactions between firms, such as "buyer, seller, volume".
  • References::
    1. Vorontsov K. V. Review of probabilistic thematic models, 2017.
  • Basic algorithm: Standard methods for non-negative matrix expansions.
  • Solution: Regularized EM-algorithm for sparse non-negative matrix expansions. Visualization of the graph of financial flows. Testing the algorithm on synthetic data, testing the hypothesis about the stability of sparse solutions.
  • Novelty: Thematic modeling has not previously been applied to the analysis of financial transactional data.
  • Authors: Vorontsov K. V.. consultants: Viktor Safronov, Rosa Aisina.

Task scoring +

  • Name: Generating and selecting features when building a credit scoring model.
  • Task: Credit scoring models are built step by step. In particular, a number of independent transformations of individual features are performed, and new features are generated. Each step uses its own quality criterion. It is required to build a scoring model that adequately describes the sample. Maximizing the quality of the model at each step does not guarantee the maximum quality of the resulting model. It is proposed to abandon the step-by-step construction of the scoring model. To do this, the quality criterion must include all the optimized parameters of the model.
  • Data: The computational experiment will be performed on 5-7 samples to be found. It is desirable that the samples be of the same nature, for example, the samples of consumer credit questionnaires.
  • References:: Siddique N. Constructing scoring models, SAS. Hosmer D., Lemeshow S., Applied logistic regression, Wiley. Katrutsa A.M., Strijov V.V. Comprehensive study of feature selection methods to solve multicollinearity problem according to evaluation criteria // Expert Systems with applications, 2017.
  • Basic algorithm: The scoring model construction algorithm recommended by SAS.
  • Solution: Each step of the procedure is represented as an optimization problem. The parameters to be optimized are combined, and the Feature Selection Task is included as a Mixed Optimization Task.
  • Novelty: An error function is proposed, when using which the generation and selection of features, as well as the optimization of model parameters, are performed together.
  • Authors: T.V. Voznesenskaya, Strizhov V.V.

Task Popova +

  • Name: Representation of molecules in 3D
  • Task: Develop representations of the 3D structure of molecules that would have the property of rotational and translational invariance.
  • Data: Millions of molecules given by 3D coordinates
  • References:: https://arxiv.org/abs/1610.08935, http://journals.aps.org/prl/abstract/10.1103/PhysRevLett.98.146401
  • Basic algorithm: low rank matrix/tensor factorization
  • Solution: Molecules have a different number of atoms, and therefore the matrix of their 3D coordinates is Nx3. We need to find a mathematical transformation that would be independent of N (N is the number of atoms).
  • Novelty: existing algorithms depend on the number of atoms in the molecule
  • Authors: Expert -- Alexander Isaev, consultant -- Maria Popova

Task Maksimov +

  • Name: Optimal algorithm for recovering block Hamiltonians (XY and Heisenberg models).
  • Task: The task is to reconstruct block Hamiltonians with continuous spins (a generalization of the Ising model to two- and three-dimensional spins) from the observed data. This setting is a special case of a field of machine learning known as unsupervised learning. Reconstruction of a graphical spin model from observational data is an important problem in physics. The basis of the algorithm will be the adaptation of a new optimal method of screening interactions (interaction screening), developed for the Ising model. The solution process will combine familiarity with computer science/machine learning theoretical methods and numerical experiments.
  • Data: Simulated block spin model configurations.
  • References::
    1. Lokhov et al., "Optimal structure and parameter learning of Ising models", arXiv:1612.05024 (2016) {https://arxiv.org/abs/1612.05024}
    2. Vuffray et al., "Interaction screening: efficient and sample-optimal learning of Ising models", NIPS 2016 {https://arxiv.org/abs/1605.07252}
    3. Tyagi et al., "Regularization and decimation pseudolikelihood approaches to statistical inference in XY spin models", Phys. Rev. B 2016 {https://arxiv.org/abs/1603.05101}
  • Basic algorithm: Dynamic method for shielding interactions. Comparison with the method of maximum pseudo-likelihood (pseudolikelihood).
  • Novelty: An algorithm based on the dynamic interaction shielding method has a good chance of being optimal for this problem, because the corresponding method is optimal for the inverse Ising problem.
  • Author: consultants Andrey Lokhov, Yuri Maksimov. Expert Mikhail Chertkov

Task Khritankova (Transfer Learning)

  • Name: Using deep learning networks to transfer classification models in case of insufficient data.
  • Task:
    1. Develop an algorithm for calculating a set of latent features in the symmetric homogeneous transfer learning problem, the solution of the classification problem in which does not depend on the original area, and which is no worse than when solving for each area separately (transfer error) for the case of small sample sizes with errors in markup
    2. Develop an algorithm for transitioning to a hidden set of features without using markup (unsupervised domain adaptation)
  • Data: teraPromise-CK (33 datasets with the same features but different distributions).
  • References:: Base article: Xavier Glorot , Antoine Bordes , Yoshua Bengio. (2011) Domain Adaptation for Large-Scale sentiment classification: A Deep Learning approach / In Proceedings of the Twenty-eight International Conference on Machine Learning, ICML.

Articles with ideas for improving the algorithm will be handed out (several).

  • Basic algorithm: SDA (Stacked Denoising Autoencoder) – described in the Glorot et al.
  • Solution: Take the Basic algorithm, a) try to improve it for application to small datasets of 100-1000 objects (when transfer learning is applied) by applying regularizers, adjusting the architecture of the autoencoder, adjusting the learning algorithm (for example, bootstrapping) b ) investigate the model for resistance to markup errors (label corruption / noisy labels) and propose improvements to increase stability (robustness).
  • Novelty: Obtaining a stable algorithm for transferring classification models on small amounts of data with markup errors.
  • Authors: Khritankov

Task INRIA-МТФИ +

  • Name: Estimated binding energy of protein and small molecules.
  • Task: Modeling the binding of a protein and a small molecule (hereinafter referred to as a ligand) is based on the fact that the best ligand in its best position has the lowest free energy of interaction with the protein. It is necessary to estimate the free energy of protein and ligand binding. Complexes of proteins with ligands can be used for training, and for each protein there are several positions of the ligand: 1 correct, "native", for which the energy is minimal, and several generated incorrect ones. For a third of the data set, values are known that are proportional to the desired binding energy of ligands in native positions with the protein. There is a separate test set consisting of 1) complexes of proteins and ligands, for which it is necessary to find the best ligand position (the algorithm for obtaining ligand positions differs from that used in training), 2) complexes of proteins and ligands, for whose native positions it is necessary to predict the binding energy, and 3) proteins for which it is necessary to find the most strongly binding ligand.
  • Data: About 10000 complexes: for each of them there is 1 native pose and 18 (more can be generated) non-native ones. The main descriptors are histograms of distributions of distances between different atoms of the protein and ligand, the dimension of the vector of descriptors is ~ 20,000. The set of descriptors can be extended (you can generate poses with different deviations and use it as a descriptor, you can add the properties of small molecules: the number of bonds around which rotation is possible in a molecule, its surface area, its surface division by a Voronoi diagram. The data will be provided in the form of binary files with a python script to read.
  • References:: PEPSI-Dock: a detailed data-driven protein–protein interaction potential accelerated by polar Fourier correlation Predicting Binding Poses and Affinities in the CSAR 2013―2014 Docking Exercises Using the Knowledge-Based Convex-PL Potential
  • Basic algorithm: We used a linear SVM (these are just lecture notes, I see no reason to give Vapnik here, especially since all this, including these lecture notes, is googled), the connection of which with an energy estimate that goes beyond scope of the classification task is described in the articles listed above. To take into account experimentally known values proportional to energy, it is proposed to use linear regression SVR .
  • Solution: It is necessary to reduce the previously used SVM problem to a regression problem and solve it using standard methods. To check the operation of the algorithm, both the test described above and several other test sets with similar Tasks but different data will be used.
  • Novelty: Proper assessment of the quality of protein and ligand binding is used in drug development to find molecules that interact most strongly with the protein under study.

Of particular importance is the assessment of the values of the binding energy of the protein with the ligand: the coefficient of correlation (Pearson) of the energy with its experimental values determined by different groups on the proposed test does not exceed 0.7. Prediction of the most strongly binding ligand from a large number of non-protein-binding molecules is also difficult. The aim of this work is to obtain a method that allows a fairly accurate assessment of protein binding to ligands. From the point of view of machine learning and optimization, it is of interest to combine classification and regression problems.

  • Appendix Given several data sets describing an atom in a molecule or a bond between atoms, with a small feature vector (usually 3-10 descriptors) and several classes corresponding to the atom's hybridization or bond order. The data itself can be from ~100 to 20,000 vectors depending on the type of atom. You need to test some kind of multiclass machine learning on this (random forests, neural network, something else), you can do anything with descriptors. We are currently using SVM. Not only the accuracy is important, but also the computational complexity of the prediction.
  • Authors: Sergei Grudinin, Maria Kadukova

Task Strizhov and Kulunchakov +

  • Name: Creation of delay-operators for multiscale forecasting by means of symbolic regression
  • Task: Suppose that one needs to build a forecasting machine for a response variable. Given a large set of time series, one can advance a hypothesis that they are related to this variable. Relying upon this hypothesis, we can use given time series as features for the forecasting machine. However, the values of time series could be produced with different frequencies. Therefore, we should take into account not only the values, but the delays as well. The simplest model for forecast is a linear one. In the presence of large set of features this model can approximate the response quite well. To avoid the problem of multiscaling, we introduce a definition of delay-operators. Each delay-operator corresponds to one time series and represents continuous correlation function. This correlation function shows a dependence between the response variable and corresponding time series. Therefore, each delay-operator put weights on the values of corresponding time series depending on the greatness of the delay. Having these delay-operators, we avoid the problem of multiscaling. To find them, we use genetic programming and symbolic regression. If the resulted weighted linear regression model would produce poor approximation, we can use a nonlinear one instead. To find good nonlinear function, we would use symbolic regression as well.
  • Data: Any data from the domain of multiscalse forecating of time series. See the full version of this introduction.
  • References:: to be handed by V.V.Strijov
  • Basic algorithm: to be handed by V.V.Strijov
  • Solution: Use genetic algorithms applied to symbolic regression to create and test delay-operators in multiscale forecasting.
  • Novelty: to be handed by V.V.Strijov
  • Authors: supervisor: V.V.Strijov, consultant: A.S. Kulunchakov

2016

Author Topic Link Consultant Reviewer Report Letters Grade Magazine
Goncharov Alexey (example) Metric classification of time series code,

paper, slides

Maria Popova Zadayanchuk Andrey BMF AILSBRCVTDSWH> 10 ИИП
Bayandina Anastasia Thematic models of distributive semantics for highlighting ethno-relevant topics in social networks paper

slides video

Anna Potapenko Oleg Gorodnitsky BF AILSB++RCVTDEWHS 10
Belozerova Anastasia Coordination of logical and linear classification models in the information analysis of electrocardiosignals code

paper slides video

Vlada Tselykh Malygin Vitaly BF AILSB+RC+VTD>E0WH>S 10
Maria Vladimirova Bagging of neural networks in the problem of predicting the biological activity of cell receptors code

paper slides vido

Maria Popova Volodin Sergey BMF AILSBRCVTD>E>WHS 10
Volodin Sergey A probabilistic approach to the problem of predicting the biological activity of nuclear receptors code paper slides

video, itis

Maria Popova Maria Vladimirova BMF AILSBRCVTDEWHS 10
Gorodnitsky Oleg An Adaptive Nonlinear Method for Recovering a Matrix from Partial Observations code

paper slides, itis

Mikhail Trofimov Bayandina Anastasia M A++I++L++S+B+R+C++VTDE+WH 10
Ivanychev Sergey Synergy of classification algorithms (SVM Multimodelling) code

paper slides

Alexander Aduenko BM A+I+L++S+BRCVTDEW+H 10
Kovaleva Valeria Regular structure of rare macromolecular clusters code

paper slides video, itis

Olga Valba, Yuri Maksimov Dmitry Fedoryaka BM A+IL+SBRCVTD0E0WH 10
Makarchuk Gleb Time series transformations for hand motion decoding using ECoG signals (electrocorticographic signals) of monkeys code,

paper slides video

Andrey Zadayanchuk BF AI+L+S+BRС>V>T+D>E0WH>S 10
Malygin Vitaly Application of combinatorial estimates of retraining of threshold decision rules for feature selection in the problem of medical diagnostics by the method of V. M. Uspensky code,

paper, slides

Shaura Ishkina Belozerova Anastasia B AILSBRCVTDEWH 10
Molibog Igor Using Dimension Reduction Methods When Building a Feature Space in the Problem of Internal Plagiarism Detection

paper, doc, slides, itis

Anastasia Motrenko Safin Kamil BMF AILSBRCVTDEWHS 10
Pogodin Roman Determining the position of proteins using an electronic map code, paper, slides

video, itis

Alexander Katrutsa Andrey Ryazanov BMF AILSBRСVTDEWHS 10
Andrey Ryazanov Restoration of the primary structure of a protein according to the geometry of its main chain folder

paper slides video, itis

Mikhail Karasikov Roman Pogodin BMF AIL+SBRC++VTD+EWHS 10
Safin Kamil Definition of borrowings in the text without indicating the source code, paper

slides video

Mikhail Kuznetsov Molibog Igor BMF AIL+SBRC>V>T>D>E0WHS 10
Dmitry Fedoryaka Mixtures of vector autoregression models in the problem of time series forecasting code,

slides, paper

Radoslav Neichev Kovaleva Valeria BM AILSBRCV-T>D0E0WH> 10
Tsvetkova Olga Building scoring models in the SAS system code,

paper slides

Raisa Jamtyrova Chygrynskiy Viktor BF A+I+L+S+B+R+C+V0T0D0E0WH>S 10
Chygrynskiy Viktor Approximation of the boundaries of the iris code paper

slides video

Yuri Efimov B AI+L+SBRCV+TDEHFS 10

Task 1

  • Data: Synergy of classification algorithms. Data from the UCI repository so that it can be compared directly with other works, in particular the work of Vapnik.
  • References:: There are different approaches to combining SVMs: on example, bagging (http://www.ecse.rpiscrews.us/~cvrl/FaceProject/Homepage/Publication/ICPR04_final_cameraready_v4.pdf), also try and boosting (http://www.researchgate.net/profile/Hong-Mo_Je/publication/3974309_Pattern_classification_using_support_vector_machine_ensemble/links/09e415091bdc559051000000.pdf).
  • Basic algorithm: Described in the problem statement
  • Solution: a modification of the basic algorithm, or simply the Basic algorithm itself. The main thing is to compare with other methods and draw conclusions, in particular, about the relationship between the presence of an improvement in the quality and diversity of sets of reference objects built by different SVMs.
  • Novelty: It is known (for example, from Konstantin Vyacheslavovich's lectures) that it is not possible to build short compositions from strong classifiers (for example, SVM) using boosting (although they still try (see literature)). Therefore, it is proposed to build a nonlinear combination instead of a linear one. It is assumed that such a composition can give an increase in quality compared to a single SVM.
  • consultant: Alexander Aduenko

Task 2

  • Name: Temporal theme model of the press release collection.
  • Task: Development of methods for analyzing the thematic structure of a large text collection and its dynamics over time. The problem is the assessment of the quality of the constructed structure. It is required to implement the criteria of stability and completeness of the temporal thematic model using manual selection of the found topics according to their interpretability, difference and eventfulness.
  • Data: A collection of press releases from the foreign ministries of a number of countries over 10 years, in English.
  • References::
    1. Doikov N.V. Adaptive regularization of probabilistic topic models. VKR bachelor, VMK MSU. 2015.
  • Basic algorithm: Blay's classic LDA with post-hoc time analysis.
  • Solution: Implementation of an additively regularized topic model using the BigARTM library. Building a series of thematic models. Evaluation of their interpretability, stability and completeness.
  • Novelty: Criteria for sustainability and completeness of thematic models are new.
  • consultant: Nikita Doikov, problem author Vorontsov K. V.

Task 3

  • Name: Coordination of logical and linear classification models in the information analysis of electrocardiosignals.
  • Task: There are logical classifiers based on the identification of diagnostic standards for each disease and built by the Expert in semi-manual mode. For these classifiers, estimates of disease activities are determined, which have been used in the diagnostic system for many years and satisfy physician users. We build linear classifiers that are trained completely automatically and are ahead of logical classifiers in terms of classification quality. However, a direct transfer of the activity estimation technique to linear classifiers turned out to be impossible. It is required to build a linear activity model, setting it to reproduce the known activity estimates of the logical classifier.
  • Data: A selection of more than 10 thousand electrocardiograms with diagnoses for 32 diseases.
  • References:: will issue :)
  • Basic algorithm: Linear classifier.
  • Solution: Methods of linear regression, linear classification, feature selection.
  • Novelty: The task of matching two models of different nature can be considered as learning with privileged information - a promising direction proposed by the machine learning classic VN Vapnik several years ago.
  • consultant: Vlada Tselykh, problem author Vorontsov K. V.

Task 4

  • Name: Thematic classification model for diagnosing diseases by electrocardiogram.
  • Task: Technology of information analysis of electrocardiosignals according to V.M.Uspensky is based on ECG conversion into a character string and selection of informative sets of words - diagnostic standards for each disease. The linear classifier builds one diagnostic standard for each disease. The Screenfax screening diagnostic system now uses four standards for each disease, built in a semi-manual mode. It is required to fully automate the process of constructing diagnostic standards and to determine their optimal number for each disease. To do this, it is supposed to finalize the thematic classification model of S. Tsyganova, to perform a new implementation under BigARTM, to expand computational experiments, to improve the quality of classification.
  • Data: A selection of more than 10 thousand electrocardiograms with diagnoses for 32 diseases.
  • References:: will issue :)
  • Basic algorithm: Classification models by V.Tselykh, thematic model by S.Tsyganova.
  • Solution: Topic model implemented using the BigARTM library.
  • Novelty: Topic models have not previously been used to classify sampled biomedical signals.
  • consultant: Svetlana Tsyganova, problem author Vorontsov K. V.

Task 5

  • Name: Thematic models of distributive semantics for highlighting ethno-relevant topics in social networks.
  • Task: Thematic modeling of social media text collections faces the problem of ultra-short documents. It is not always clear where to draw the boundaries between documents (possible options: a single post, a user's wall, all posts by a given user, all posts for a given day in a given region, and so on). Topic models give interpretable vector representations of words and documents, but their quality depends on the distribution of document lengths. The word2vec model is independent of document lengths, since it takes into account only the local contexts of words, but the coordinates of vector representations do not allow thematic interpretation. The objective of the project is to build a hybrid model that combines the advantages and is free from the disadvantages of both models.
  • Data: Collections of social networks LJ and VK.
  • References:: will issue :)
  • Basic algorithm: Topic models previously built on this data.
  • Solution: Implementation of a distributive semantics regularizer similar to the vord2vec language model in the BigARTM library.
  • Novelty: So far, there are no language models in the literature that combine the main advantages of probabilistic topic models and the word2vec model.
  • consultant: Anna Potapenko, on technical issues Murat Apishev, problem author Vorontsov K. V.

Task 7

  • Name: Determining the position of proteins using an electronic map
  • Task: informally --- there are sets of experimentally determined maps of the location of proteins in complexes, some of them are known in high resolution, it is necessary to restore the entire map in high resolution; formally --- there are matrices and energy vectors corresponding to each map of the protein complex, it is necessary to determine which set of proteins minimizes the quadratic form formed by the matrix and vector.
  • Data: experimental data from the site http://www.emdatabank.org/ will be converted into matrices into energy vectors. Understanding the biophysical nature is not necessary.
  • References:: articles on methods for solving quadratic programming problems and various relaxations
  • Basic algorithm: quadratic programming methods with various relaxations
  • Solution: minimizing the total energy of the protein complex
  • Novelty: the application of quadratic programming methods and the study of their accuracy in the tasks of restoring electronic maps
  • consultant: Alexander Katrutsa, problem author: Sergei Grudinin.
  • Desirable skills: understanding and interest in optimization methods, working with CVX package

Task 8

  • Name: Classification of Physical Activity: Investigation of Parameter Space Variation in Retraining and Modification of Deep Learning Models
  • Task: Given a classification model for a sample of time segments recorded from a mobile phone's accelerometer. The model is a multilayer neural network. It is required 1) to investigate the variance and covariance matrix of the neural network parameters under different optimization schedules (i.e., under different approaches to staged learning). 2) based on the obtained parameter covariance matrix, propose an effective way to modify the deep learning model.
  • Data: WISDM Sample http://www.cis.fordham.edu/wisdm/dataset.php.
  • References::
    • Zadayanchuk A.I., Popova M.S., Strizhov V.V. Choosing the optimal physical activity classification model based on accelerometer measurements http://strijov.com/papers/Zadayanchuk2015OptimalNN4.pdf
    • Popova M.S., Strizhov V.V. Building Deep Learning Networks for Time Series Classification - http://strijov.com/papers/PopovaStrijov2015DeepLearning.pdf
    • Oleg Bakhteev Yu., Popova M.S., Strizhov V.V. Deep Learning Systems and Tools in Task Classification
    • LeCun Y. Optimal Brain Damage - yann.lecun.com/exdb/publis/pdf/lecun-90b.pdf
    • Works on pre-training (pre-training) and additional training (fine-tuning)
  • Basic algorithm: The basic model is described in the article "Building Deep Learning Networks for Time Series Classification". The algorithm can be implemented either using the PyLearn library or keras (other libraries and programming languages are also acceptable).
  • Solution: Analysis of the covariance matrix, building an add-del method based on the received data.
  • Novelty: The technique for studying a high-dimensional covariance matrix, as well as the resulting model modification algorithm, are important and will be used in the future when analyzing deep learning models.
  • consultant: Oleg Bakhteev

Task 9

  • Name: Restoration of the primary structure of a protein according to the geometry of its main chain
  • Task: on the basis of the main chain of the protein, that is, in essence its geometry, it is necessary to restore the primary structure of the protein, that is, which sequence of amino acids corresponds to the given geometry of the main chain. It is proposed to do this on the basis of minimizing the total energy of the protein, expressed by a quadratic form, most likely not positive definite.
  • Data: at the choice of the student: collected energy matrices for various proteins based on their descriptions in the PDB format or the PDB files themselves; in the latter case, it will be necessary to collect matrices for further work
  • References:: articles on methods for solving quadratic programming problems and various relaxations
  • Basic algorithm: quadratic programming methods with various relaxations
  • Solution: minimizing the total protein energy
  • Novelty: application of quadratic programming methods and study of their accuracy
  • consultant: Mikhail Karasikov, problem author: Sergei Grudinin.
  • Desirable skills: understanding and interest in optimization methods, working with CVX package

Task 10

  • Name: Multi-task learning approach for the task of predicting the biological activity of nuclear receptors
  • Task: In the task it is necessary to build a multi-task model that predicts the interaction of two types of molecules: receptors and proteins. The solution of this problem is necessary for the development of new drugs (drug design).
  • Data: description of 8500+ proteins and labels for 12 receptors
  • References:: will be sent to the student
  • Basic algorithm: multi-task lasso regression from scikit-learn python library
  • Solution: generalization of linear regression to the multi-task case in probabilistic interpretation
  • Novelty: Multi-task learning approach is pioneering in drug design
  • consultant: Maria Popova
  • Desired skills: understanding of and interest in probability theory, willingness to quickly understand various approaches to regression, knowledge or willingness to learn Python

Task 11

  • Name: Bagging of neural networks in the task of predicting the biological activity of nuclear receptors.
  • Task: In the task, it is necessary to implement bagging (bootstrap aggregating) for a two-layer neural network. Such a model will be multitasking and predict the interaction of two types of molecules: receptors and proteins. The solution of this problem is necessary for the development of new drugs (drug design).
  • Data: description of 8500+ proteins and labels for 12 receptors
  • References:: will be sent to the student
  • Basic algorithm: two-layer neural network
  • Solution: Composition of base classifiers bagging
  • Novelty: This approach is innovative in the field of drug design
  • consultant: Maria Popova

Task 12

  • Name: Mixtures of models in vector autoregression in the problem of predicting (large) time series.
  • Task: There is a set of time series of length T containing the readings of various sensors that reflect the state of the device. It is necessary to predict the next t sensor readings. Practical significance: before a breakdown, the state of the device changes, the prediction of "abnormal" behavior will help to take timely measures and avoid breakdowns or minimize losses.
  • Data: Multivariate time series with indications of various server sensors (CPU, memory, temperature)
  • References:: Keywords: mixture models, boosting, Adaboost, vector autoregression.
    • Alexander Tsyplakov. Introduction to forecasting in classical time series models. [89]
    • Neichev R.G., Katrutsa A.M., Strizhov V.V. Selection of the optimal set of features from a multicorrelated set in the forecasting problem[90]
    • Christopher M. Bishop. Pattern Recognition and Machine Learning. Page 667
  • Basic algorithm: Boosting, Adaboost algorithm.
  • Solution: Use a mixture of several linear models instead of one complex one to build pronosis.
  • Novelty: Improved parameter space for mixture of models in vector autoregression.
  • consultant: Radoslav Neichev

Task 13

  • Name: Selection of multicorrelated features in the problem of vector autoregression.
  • Task: There is a set of time series containing the readings of various sensors that reflect the state of the device. The readings of the sensors correlate with each other. It is necessary to select the optimal set of features for solving the forecasting problem.
  • Data: Multivariate time series with indications of various server sensors (CPU, memory, temperature)
  • References:: Keywords: bootstrap aggregation, Belsley method, vector autoregression.
    • Neichev R.G., Katrutsa A.M., Strizhov V.V. Selection of the optimal set of features from a multicorrelated set in the forecasting problem[91]
  • Basic algorithm: Belsley's method for univariate autoregression (see bibliography article).
  • Solution: Apply the Belsley method to detect correlated features.
  • Novelty: The Belsley method is used for vector autoregression.
  • consultant: Radoslav Neichev

Task 14

  • Name: Generation of features in the prediction problem.
  • Task: There is a set of time series containing the readings of various sensors that reflect the state of the device. It is necessary to expand the feature space with the help of non-linear parametric generating functions.
  • Data: Multivariate time series with indications of various server sensors (CPU, memory, temperature)
  • References:: Keywords: curvilinear regression, feature generation, non-linear regression, time series approximation.
    • M.P. Kuznetsov, Strizhov V.V., M.M. Medvednikov. Algorithm for multiclass classification of objects described in rank scales.[92]
  • Basic algorithm: Non-parametric generating functions.
  • Solution: Apply quasi-linear and non-linear parameter dependent transformations to features.
  • Novelty: A new set of features for solving autoregressive problems is proposed.
  • consultant: Roman Isachenko

Task 15

  • Name: Time series transformations for hand motion decoding using ECoG signals (electrocorticographic signals) in monkeys.
  • Task: There is a set of time series records of ECoG signals. It is necessary to extract the features using time series transformations (for example, the windowed Fourier transform).
  • Data: Multivariate time series with ECOG readings and monkey movement data [93]
  • References:: Keywords: feature extraction, time series transformations, ECoG signal processing
    • Zenas C. Chao, Yasuo Nagasaka and Naotaka Fujii. Long-term asynchronous decoding of arm motion using electrocorticographic signals in monkeys
  • Basic algorithm: Wavelet transform
  • Solution: Feature extraction from ECoG by various methods.
  • Novelty: Wavelet Transform Optimality Analysis in ECoG Signal Processing Tasks
  • consultant: Zadayanchuk Andrey

Task 16

  • Name: An adaptive nonlinear method for recovering a matrix from partial observations
  • Task: Let there be an unknown (possibly multidimensional) matrix A, the position of an element in it is described by an integer vector p. The values of the matrix on some subset of its elements are known. It is required to find a parametrization and parameters such that the quadratic deviation is minimized on some subset of elements. More detailed description at the link [94]
  • Data: model data, Netflix Prize Data Set, MovieLens 20M Dataset, Criteo Display Advertising Challenge Dataset
  • References::
    • "ACCAMS: Additive Co-Clustering to Approximate Matrices Succinctly" (Beutel, Amr Ahmed, Smola)
    • "Non-linear Matrix Factorization with Gaussian Processes" (Neil D. Lawrence)
    • "Low-rank matrix completion using alternating minimization" (Prateek Jain, Praneeth Netrapalli, Sujay Sanghavi)
  • Basic algorithm: Low-rank approximation
  • Solution: and parameters, and search for parametrization from the data.
  • Novelty: A summary of works in this area; a new model is proposed, the effectiveness of which is proposed to be tested
  • consultant: Mikhail Trofimov
  • Desirable Skills: python

Task 17

  • Name: Building scoring models in the SAS system (or MATLAB).
  • Task: Describe the main steps in building scoring models. At the stage of data preparation, the Task of filtering choices (removing noise objects) is solved. Since the sample contains a significant number of features that do not correlate with solvency, it is necessary to solve the problem of feature selection. In addition, due to the heterogeneity of the data (by example, by region), it is proposed to build a mixture of models, in which each model describes its own subset of the sample. At the same time, different sets of features can correspond to different components of the mixture.
  • Data: Credit Story/Potential Borrower Questionnaires [95], .uci.edu/ml/datasets/Statlog+%28Australian+Credit+Approval%29/.
  • References::
  • Basic algorithm: Logistic regression
  • Solution: Mix of models
  • Novelty: A method for constructing scoring maps is described, in which both feature generation and multi-modeling are included in the optimization problem.
  • consultant: Raisa Jamtyrova
  • Desirable Skills: SAS

Task 18

  • Name: Approximation of the boundaries of the iris.
  • Task: Based on the image of the human eye, determine the circles approximating the inner and outer border of the iris.
  • Data: Raster monochrome images, typical size 640*480 pixels (however, other sizes are also possible)

[96], [97].

  • References::
    • K.A. Gankin, A.N. Gneushev, I.A. Matveev Segmentation of the iris image based on approximate methods with subsequent refinements // Izvestiya RAN. Theory and control systems, 2014, no. 2, p. 78–92.
    • Duda, R. O. Use of the Hough transformation to detect lines and curves in pictures / R. O. Duda, P. E. Hart // Communications of the ACM. 1972 Vol. 15, no. 1.Pp.
  • Basic algorithm: Efimov Yury. Search for the outer and inner boundaries of the iris in the eye image using the paired gradient method, 2015.
  • Solution: See iris_circle_problem.pdf
  • Novelty: A fast non-enumerative algorithm for approximating boundaries using linear multimodels is proposed.
  • consultant: Yuri Efimov (by Strizhov V.V., Expert Matveev)

Task 19

  • Name: Approximation of combinatorial overfitting estimates for feature selection in the problem of medical diagnostics.
  • Task: Technology of information analysis of electrocardiosignals according to V. M. Uspensky is used to diagnose diseases of internal organs by electrocardiogram. The linear naive bayesian classifier with feature selection performs well in this task. However, only very simple greedy strategies have been used so far for feature selection. It is proposed to use more intensive enumeration strategies to find better and shorter diagnostic feature sets. However, the more intense the search, the higher the probability of overfitting. To reduce overfitting, it is proposed to use combinatorial estimates of overfitting of threshold decision rules. For efficient calculation of these estimates, it is proposed to use surrogate modeling.
  • Data: Samples of vectors of ECG feature descriptions obtained using the Screenfax screening diagnostics system. Will be issued.
  • References::
  • Basic algorithm: linear naive bayes classifier with feature selection.
  • Solution: Exact combinatorial formulas are used to evaluate overfitting. For approximation (surrogate modeling) of these formulas, MVR Composer is used. Heuristic semi-greedy combinatorial optimization algorithms are used for feature selection.
  • Novelty: Previously, combinatorial retraining estimates were not used for feature selection. This method makes it possible to reduce diagnostic sets of features and improve the quality of classification.
  • consultant: Ishkina Shaura, Kulunchakov Andrey (MVR Composer), problem author: Vorontsov K. V.

Task 20

  • Name: Object generation model in the problem of time series forecasting
  • Task: Build an object generation model for the prediction task, which will create a high-quality sample for the subsequent solution of the prediction task.
  • Data: Electricity consumption time series, mobile phone accelerometer time series
  • References::
    • Keogh E. J., Pazzani M. J. Scaling up dynamic time warping to massive datasets
    • Salvador S., Chan P. Fastdtw: Toward accurate dynamic time warping in linear time and space
    • Kuznetsov M.P., Ivkin N.P. Algorithm for classification of accelerometer time series by combined feature description
    • Karasikov M. E. Classification of time series in the space of parameters of generating models
  • Basic algorithm: Various heuristics
  • Problem Statement: The formulation and detailed description of the problem is given at [98]
  • Novelty: consideration of the data generation model in a similar task
  • consultant: Alexey Goncharov

Task 21

  • Name: Algorithm for predicting the structure of locally optimal models
  • Task: It is required to predict a time series using some parametric superposition of algebraic functions. It is proposed not to cost the prognostic model, but to predict it, that is, to predict the structure of the approximating superposition. A class of considered superpositions is introduced, and on the set of such structural descriptions, a search is made for a locally optimal model for the problem under consideration. The task consists in 1) searching for a suitable structural description of the model 2) describing the search algorithm for the structure that will correspond to the optimal model 3) describing the algorithm for inverse construction of the model according to its structural description. For an already existing example of the answer to questions 1-3, see the work of A. A. Varfolomeeva.
  • Data: A set of time series, which implies the restoration of functional dependencies. It is proposed to first use synthetic data or immediately apply the algorithm to forecasting time series 1) electricity consumption 2) physical activity with subsequent analysis of the resulting structures.
  • References::
    • A. A. Varfolomeeva Selection of features when marking up bibliographic lists using structural learning methods, 2013, [99]
    • Bin Cao, Ying Li and Jianwei Yin Measuring Similarity between Graphs Based on the Levenshtein Distance, 2012, [100]
  • Basic algorithm: Specifically, there is no basic algorithm for the proposed problem. It is proposed to try to repeat the experiment of A. A. Varfolomeeva for a different structural description in order to understand what is happening.
  • Solution: The superposition of algebraic functions defines an ortree, on the vertices of which the labels of the corresponding algebraic functions or variables are given. Therefore, the structural description of such a superposition can be its DFS-code. This is a string consisting of vertex labels, written in the order in which the tree is traversed by depth-first search. Knowing the arities of the corresponding algebraic functions, we can restore any such DFS-code in O(n) and get back the superposition of functions. On the set of similar string descriptions, it is proposed to search for the string description that will correspond to the optimal model.
  • consultant: Kulunchakov Andrey

Task 22

  • Name: Definition of borrowings in the text without indicating the source
  • Task: The Task is solved to detect internal borrowings in the text. It is required to test the hypothesis that the given text was written by a single author, and if it is not fulfilled, highlight the borrowed parts of the text. A borrowing is a part of the text, presumably written by another author and containing characteristic differences from the style of the main author. It is required to develop such a style function that allows to distinguish with a high degree of certainty the style of the main author of the text from borrowings.
  • Data: PAN-2011 contest collection.
  • References::
    1. Oberreuter, G., L'Huillier, G., Rıos, S. A., & Velásquez, J. D. (2011). Approaches for intrinsic and external plagiarism detection. Proceedings of the PAN.
  • Basic algorithm, solution: At the moment, a basic method for identifying dependencies is implemented, based on the analysis of the frequencies of words and symbolic n-grams in a sentence. For each text, a dictionary is formed, in which each word (n-gram) is assigned the value of its occurrence in the text. Based on the occurrence values, an indicative description of each segment-offer is formed. Classification of text segments is performed on the basis of Expert markup of borrowings. The quality of the base algorithm is 0.29 in F1-measure (Pladget 0.21) on the PAN-2011 collection, while the quality of the best algorithm that participated in the 2011 competition [Oberreuter] is 0.32 in F1-measure (Pladget 0.32). It is proposed to implement this algorithm and compare it with the base method.
  • consultant: Mikhail Kuznetsov

Task 23

  • Name: Using Dimension Reduction Methods When Building a Feature Space in the Problem of Internal Plagiarism Detection
  • Task: For a more efficient solution to the task of detecting internal plagiarism, use dimensionality reduction methods that preserve the distance between objects. It is required to refine the tSNE method [2] by including in the model information about data markup and the possibility of adding previously unconsidered objects to the space of reduced dimension. For details see [1]
  • Data: PAN-2011 contest collection.
  • References::
    1. Problem_statement_dim_reduce.pdf‎
    2. Laurens van der Maaten. Visualizing Data using t-SNE Journal of Machine Learning Research, 9 (2008) 2579-2605.
    3. Julian Brooke and Graeme Hirst. Paragraph Clustering for Intrinsic Plagiarism Detection using a Stylistic Vector-Space Model with Extrinsic Features, 2012.
  • Basic algorithm, solution: See [1]
  • consultant: Anastasia Motrenko

Task 26

  • Name: Construction of mappings with minimal deformation to compare images with the standard.
  • Task: Apply the variational method of constructing quasi-isometric mappings to solve the classical problem of geometric morphology and image registration - constructing a two-dimensional or three-dimensional deformation for comparison with the standard.
  • Data: Images in bmp format. At the first stage, simple bodies can be defined by means of a b/w coloring of the Cartesian lattice.
  • References::
    1. Michael I. Miller, Alain Trouve, Laurent Younes. ON THE METRICS AND EULER-LAGRANGE EQUATIONS OF COMPUTATIONAL ANATOMY. Annu. Rev. Biomed. Eng. 2002. 4:375–405
    2. Beg MF, Miller MI, Trouve A, Younes L. Computing large deformation metric mappings via geodesics flows of diffeomorphisms. International Journal of Computer Vision. 2005; V.61(2):139-157.
    3. Trouve A. An approach of pattern recognition through infinite dimensional group action. Research report LMENS-95-9. 1995.
    4. Garanzha VA. Maximum norm optimization of quasi-isometric mappings. Num. Linear Algebra Appl. 2002; V.9(6-7):493-510.
    5. Garanzha V.A., Kudryavtseva L.N., Utyzhnikov S.V. Untangling and optimization of spatial meshes // Journal of Computational and Applied Mathematics. -- 2014. -- October. -- V. 269 -- P. 24--41.
  • Basic algorithm: Use the variational method for constructing mappings, which was previously proposed for constructing spatial mappings with a given boundary mapping [4], [5], in the case when a measure of proximity of functions describing geometric bodies is given on example , as an rms measure of the proximity of brightness functions.
  • Solution: For the existing code that implements the variational method for constructing two-dimensional mappings with minimal distortion, it is necessary to add a module that implements an additive to the functional, which is a measure of the proximity of geometric bodies. This includes calculating the functional itself, its gradient, and adjusting the preconditioner.
  • Novelty: Compare the obtained method with the method of geodesic flow of diffeomorphisms proposed in the works of Alain Trouvé (see references [1]-[3]). Estimate the quality of the approximation and the performance of the resulting algorithm.
  • consultant: Vladimir Anatolyevich Garanzha (CC RAS).

Task 27

  • Name: Cross-language thematic search for scientific publications.
  • Task: Creation of a prototype search service that accepts the text of a scientific article in Russian as a request and returns thematically related articles in English from the arXiv.org collection as a search result.
  • Data: The arXiv.org text collection, Wikipedia's bilingual text collection.
  • References:: will issue.
  • Basic algorithm: Topic model built from the combined collection of the English-language arXiv and the bilingual English-Russian Wikipedia.
  • Solution: Building a regularized topic model using the BigARTM library. Application of standard means of constructing inverted indexes.
  • Novelty: There is no such service on the Russian Internet yet.
  • consultant: Marina Suvorova.

Task 28

  • Name: Search for resonant frequencies in polymer solutions.
  • Task: Mathematically, Task comes down to finding the spectral density of random graphs in the vicinity of the percolation point.
  • Data: Simulation data (Erdos-Rényi graphs around the percolation point).
  • References:: Nazarov L. I. et al. A statistical model of intra-chromosome contact maps //Soft matter. - 2015. - T. 11. - No. 5. - S. 1019-1025.
  • Basic algorithm: Monte Carlo.
  • Novelty: At present, an algorithm for estimating the spectral density of linear chains is known, the issue with estimating the spectral density of tree ensembles is open.
  • consultant: Olga Valba, Yuri Maksimov, Problem Author: Nechaev Sergey.

YEAR

Author Topic Link Consultant Reviewer Report Letters Grade Magazine
Goncharov Alexey (example) Metric classification of time series code,

paper, slides

Maria Popova Zadayanchuk Andrey BMF AILSBRCVTDSW 10 ИИП
Akhtyamov Pavel Selection of multicorrelating features in the problem of vector autoregression code,

paper, slides

Radoslav Neichev Medvedeva Anna BF AI+LSB++R+CVTDEH 10
Bataev Vladislav Thematic classification model for diagnosing diseases by electrocardiogram code,

paper

Svetlana Tsyganova B AIL-S++B>R>C0V0T0D0E0W0H> >26.05 (7)
Ivanov Ilya Classification of physical activity: study of parameter space change during retraining and modification of deep learning models code,

paper, slides

Oleg Bakhteev BF A+ILS+B+R++C+VT+DEW0H 10
Medvedeva Anna Object generation model in the problem of time series forecasting code

paper slides

Goncharov Alexey Akhtyamov Pavel BF AILS-BRCVTD0EWS 10
Persianov Dmitry Temporal theme model of press release collection code

paper slides

Nikita Doikov BF A+I+L+S++B+R+C+V+T0DEW0H 10
Semenenko Denis Algorithm for Predicting the Structure of Locally Optimal Models code

paper

Kulunchakov Andrey B AI+L+SB0R0C0V0T0D0E0W0H0
Sofienko Alexander Coordination of logical and linear classification models in the information analysis of electrocardiosignals code,

paper

Vlada Tselykh B A-I-L-S-C0V0T0D0E0W0H> >26.05
Yaronskaya Lyubov Sparse Regularized Regression on Protein Complex Data code

paper slides

Alexander Katrutsa A-I-L-SB-R-CVT--D-EW0H> >26.05
Aksenov Sergey Cross-language thematic search for scientific publications. code

paper slides

Marina Suvorova AILS0B0R0C0V0T0D0E0W0H> >26.05 (7)
Khismatullin Timur Analysis and classification of the DNA-protein complex interface code

paper slides

Vladimir Garanzha F AILSBRCVT>H> >26.05 (7)

Task 6

  • Name: Sparse Regularized Regression on Protein Complex Data
  • Task: find the best regression model on protein complex binding data
  • Data: feature description of protein complexes and binding constants for them
  • References:: articles on regression and comparing methods on similar data
  • Basic algorithm: regularized linear regression (Lasso, Ridge, ..), SVR, kernel methods, etc.
  • Solution: comparison of various regression algorithms on data, selection of the optimal model and parameter optimization
  • Novelty: getting the best regression model for protein complex binding data
  • consultant: Alexander Katrutsa, problem author: Sergei Grudinin.
  • Desirable Skills: willingness to quickly understand various approaches to regression, knowledge or willingness to master C++ at an intermediate level (for a more complete study, you will need to try C++ libraries)

Task 8

  • Name: Classification of physical activity: study of parameter space change during retraining and modification of deep learning models
  • Task: Given a classification model for a sample of time segments recorded from a mobile phone's accelerometer. The model is a multilayer neural network. It is required 1) to investigate the variance and covariance matrix of the neural network parameters under different optimization schedules (i.e., under different approaches to staged learning). 2) based on the obtained parameter covariance matrix, propose an effective way to modify the deep learning model.
  • Data: WISDM Sample http://www.cis.fordham.edu/wisdm/dataset.php.
  • References::
    • Zadayanchuk A.I., Popova M.S., Strizhov V.V. Choosing the optimal physical activity classification model based on accelerometer measurements http://strijov.com/papers/Zadayanchuk2015OptimalNN4.pdf
    • Popova M.S., Strizhov V.V. Building Deep Learning Networks for Time Series Classification - http://strijov.com/papers/PopovaStrijov2015DeepLearning.pdf
    • Oleg Bakhteev Yu., Popova M.S., Strizhov V.V. Deep Learning Systems and Tools in Task Classification
    • LeCun Y. Optimal Brain Damage - yann.lecun.com/exdb/publis/pdf/lecun-90b.pdf
    • Works on pre-training (pre-training) and additional training (fine-tuning)
  • Basic algorithm: The basic model is described in the article "Building Deep Learning Networks for Time Series Classification". The algorithm can be implemented either using the PyLearn library or keras (other libraries and programming languages are also acceptable).
  • Solution: Analysis of the covariance matrix, building an add-del method based on the received data.
  • Novelty: The technique for studying a high-dimensional covariance matrix, as well as the resulting model modification algorithm, are important and will be used in the future when analyzing deep learning models.
  • consultant: Oleg Bakhteev

Task 25

  • Name: Stability of sampling of electrocardiosignals relative to frequency filtering.
  • Task: Technology of information analysis of electrocardiosignals according to V.M.Uspensky is based on the transformation of the electrocardiogram into a character string (codogram) and the selection of informative sets of words - diagnostic standards for each disease. The problem is that for discretization it is necessary to accurately determine the amplitude of the R-peaks. The amplitude can be affected by the frequency filtering of the signal, which is performed by the electrocardiograph at the hardware or software level. The task is to evaluate how much different frequency filters (example, 50.4Hz mains suppression filter, high-pass filter) can affect the word frequencies in the codegram and the quality of the classification.
  • Data: electrocardiograms in KDM format.
  • References:: will issue :)
  • Basic algorithm: Linear classifier.
  • Solution: Direct and inverse Fourier transform, algorithm for detecting R-peaks on an electrocardiogram, algorithm for determining the amplitude of R-peaks.
  • Novelty: The study of the stability of codograms in relation to frequency filtering with different parameters has not previously been carried out in the information analysis of electrocardiosignals.
  • consultant: Victor Safronov (Scientific Center named after V.I.Kulakov)

2015

Author Topic Link Consultant Reviewer DZ-1 DZ-2 (Problem number) Letters Sum Grade
Bernstein Julia Methods for characterizing fibrinolysis by in vitro blood imaging sequence Matveev I. A. Solomatin 1 3 (8) AILSBRCVTDE 11 10
Bochkarev Artem Structural learning when generating models [101] (no code), paper, slides Varfolomeeva Anna, Oleg Bakhteev Isachenko 2 2 (7) A+I++LS+BRCVT+DS 9.25 10 Goncharov Alexey Metric classification of time series code,

paper, slides

Maria Popova Задаянчук 1.5 1 (4) AILSBRCVTDSW 12 10
Dvinskikh Darina Improving the quality of forecasting using product groups code,

paper, slides

Kanevsky D. Yu. Smirnov 0.5 3 (7) AILSBRCVTDEHS 14 10
Efimov Yuri Search for the outer and inner boundaries of the iris in the eye image using the paired gradient method code,

paper, slides

Matveev I. A. Neichev AILSBRCVTDEW 12 10
Zharikov Ilya Checking the compliance of the electrocardiograph with the requirements of the diagnostic system "Screenfax" and assessing the quality of electrocardiograms. code, paper, slides Shaura Ishkina Bochkarev 3.5 3 (5) AIL+SBRCVTDEHSW 14.25 10
Zadayanchuk Andrey Choosing the optimal physical activity classification model code,

paper, slides

Maria Popova Goncharov 2 0 (17) AI-LSB+RCVTD 10 10
Zlatov Alexander Building a hierarchical model of a large conference code,

paper, slides

Arsenty Kuzmin Dvinskyh 1.5 3 (14) AI+L+SBRC++V+TDESW 14.25 10
Isachenko Roman Metric Learning and Space Dimension Reduction in Tasks of Time Series Clustering code, paper, slides Alexander Katrutsa Zharikov 3.5 3 (14) A-I+L+S-BR+CVTDEHSW 14.25 10
Radoslav Neichev Feature Selection in Time Series Forecasting Using Exogenous Factors code, paper, slides Alexander Katrutsa Efimov 1 3 (9) AI-L-SBRCVTDEHSW 13.5 10
Podkopaev Alexander Prediction of Quaternary Structures of Proteins code,

paper, slides

Maksimov Yu. V. Reshetov 3.5 3 (11) AILS+B+RCVTDEHS 13.5 10
Reshetova Daria Multiclass Classification Methods with Improved Convergence Estimators in Partial Learning Tasks code,

paper, slides

Maksimov Yu. V. Kamzolov 2.5 3 (10) AIL++SB+RCVT++DEHS- 14 10
Smirnov Евгений Thematic model of interests of permanent users of the mobile application code, paper, slides Victor Safronov Zlatov 1 1 (4) AILSBRCVTWDE 11.25 10
Solomatin Ivan Determination of the iris shading area by the classifier of local textural features code, paper, slides Matveev I. A. Bernstein 3 (9) AILSBRCVTDE 11 10
Chernykh Vladimir Testing nonparametric algorithms for time series forecasting under nonstationary conditions code,

paper, slides

Stenina Maria Shishkovets Svetlana 3.5 3 (4) A+I+LSBRCVT+DE++H++ 13.75 10
Shishkovets Svetlana Regularization of a linear naive bayes classifier. code,

paper, slides

Михаил Усков, Vorontsov K. V. Chernykh Vladimir 3.5 2 (9) A+I+L+SBR+CV+TD+E+H+S 15 10
Kamzolov Dmitri New algorithms for the problem of ranking web pages Alexander Gasnikov, Yuri Maksimov Podkopaev AILSB+RCVT+DEHS-- 13 8
Sukhareva Angelica Classification of scientific texts by branches of knowledge code,

paper, slides

Sergei Tsarkov 0.5 AILSBRCVTDEH 9

Task 1

  • Name: Improving the quality of demand forecasting using product groups
  • Task:

Given:

    1. Time series of sales for several product groups in one hypermarket. Also, for each product, periods of shortage, periods of influence on the demand of calendar holidays and periods of holding are known. marketing promotions. A product classifier is also known: a tree of product groups, where the products themselves are leaves.
    2. Forecasting algorithm that is used to generate demand forecasts for these products: self-adaptive exponential smoothing (Trigg-Leach model, see [1])
    3. Loss function by which the quality of forecasts is measured: MAPE.
    4. Requirements for building forecasts: forecasts must be built weekly for 4 weeks ahead (at the beginning of the current week, you need to build a forecast of total demand for the next week, a week in one, two, and 3).

Hypothesis: Demand for individual goods is too volatile to reveal their characteristic seasonality. It is proposed to use data on product groups in order to more accurately determine the parameters of seasonality. Note: there are other options for improving the quality of forecasting by working with groups of goods. Task is to improve the quality of forecasting within the framework of the task by taking into account the effect of the interchangeability of goods, in comparison with the basic algorithm. The result can be considered achieved if a statistically significant increase in quality is shown when building a series of forecasts (at least 20) for each time series using a sliding control.

  • Data:
    1. Data on sales of several product groups in a hypermarket of a large retail chain: https://drive.google.com/file/d/0B5YjPespcL83X3pHaE1aRzBUaDg/view?usp=sharing
  • References:
    1. Lukashin Yu. P. Adaptive methods of short-term forecasting of time series. - M .: Finance and statistics, 2003.
    2. http://www.machinelearning.ru/wiki/index.php?title=%D0%9C%D0%BE%D0%B4%D0%B5%D0%BB%D1%8C_%D0%A2%D1 %80%D0%B8%D0%B3%D0%B3%D0%B0-%D0%9B%D0%B8%D1%87%D0%B0
    3. Nitin Patel, Mahesh Kumar, Rama Ramakrishnan. Clustering models to improve forecasts in retail merchandising. http://www.cytel.com/Papers/INFORMS_Prac_%2004.pdf
    4. Kumar M., Error-based Clustering and Its Application to Sales Forecasting in Retail Merchandising. PhD Thesis. http://books.google.ru/books/about/Error_based_Clustering_and_Its_Applicati.html?id=6252NwAACAAJ&redir_esc=y
  • Basic algorithm: It is proposed to use the seasonality model [3] in combination with the Trigg-Leach model as a non-seasonal series prediction algorithm ([1] and [2]). In this case, 3 variants of the algorithm are possible, depending on the method of assessing seasonality:
    1. Seasonality is estimated by the very series of sales. For products with a "short" history, seasonality is not assessed.
    2. Seasonality is estimated for a group of goods, based on the classifier of commodity groups (lower level of the classifier)
    3. Seasonality is estimated by clusters, based on the methodology [3], [4].
  • Solution: It is required to implement the combination of the seasonality model [3] and the Trigg-Leach model as a non-seasonal series prediction algorithm ([1] and [2]), with the 3 variants of seasonality analysis described above. When constructing seasonal profiles, it is necessary to exclude periods of marketing campaigns (otherwise, there may be a significant distortion of seasonality). Next, you need a series of experiments with quality analysis on real data. When analyzing quality, you can exclude periods of holidays and marketing campaigns. Based on the results of the experiments, it may be necessary to adapt the clustering algorithm.
  • Novelty: Building a self-adaptive forecasting algorithm taking into account seasonality, identified by cluster analysis.
  • consultant: Kanevsky D.Yu.

Task 2

Task 3

  • Name: Obtaining an estimate of the sparse covariance matrix for nonlinear models (neural networks).
  • Task: Suggest a method for estimating the covariance matrix of parameters of a general model for the case of linear regression, logistic regression, general non-linear models, including neural networks. Suggest a way to take into account the structure of the matrix (sparseness, dependencies between coefficients, etc.)
  • Data: Synthetic data and tests.
  • References::
  • Basic algorithm: Diagonal matrix estimation, see MLAlgorithms/HyperOptimization folder.
  • Solution:
  • Novelty: A fast algorithm for obtaining estimates of the general covariance matrix for nonlinear models is proposed, the properties of sparse matrices are investigated.
  • consultant: Alexander Aduenko.

Task 4

  • Name: Feature selection in time series forecasting using exogenous factors
  • Task: task statement from [102] formula (32)
  • Data: time series with electricity prices.
  • References::
    • Keywords: Hourly Price Forward Curve, short-term time series forecasting, feature selection, Add-Del method, (non)linear regression.
    • Main Articles:
    1. [103] - study of the influence of prices in one country on the price in another and how to take this into account when forecasting .
    2. [104] - overview of terms and processes emerging in HPFC forecasting + motivation
    3. [105] - also about price forecasting, but here about spot prices
  • Basic algorithm:
    1. LAD-Lasso estimation from [106]
    2. Sanduleanu's article about the Add-Del modification: [107].
  • Solution: apply the modified Add-Del method as a feature selection method.
  • Novelty: comparison of basic and proposed methods, analysis of properties of the proposed method.
  • consultant: Alexander Katrutsa.

Task 5

  • Name: Development of an image recognition algorithm for the search for fibrinolysis parameters.
  • Task: A set of images of fibrin clot growth obtained during the study of thrombodynamics and 80%D0%B8%D0%BD%D0%BE%D0%BB%D0%B8%D0%B7|fibrinolysis. It is required to develop an algorithm for finding the coordinates of the segment and the angle of inclination of the activator line from a series of images. Test the developed algorithm on different types of fibrinolysis and examples where this process is absent.
  • Data: An array of images for each study in tiff format 16 bits with time points from the beginning in seconds.
  • References:
    • Description of the applied task and terms of reference: on request.
  • Basic algorithm: Hough Transform [108], discussed.
  • consultant: I.A. Matveev

Task 6

  • Name: Prediction of Quaternary Structures of Proteins: нивелирование
  • Task: The task is to predict the packing of protein molecules into a multimeric complex in the rigid body approximation. One of the formulations of the problem is written as a non-convex optimization.

It is necessary to study this formulation and propose a solution algorithm. Suppose we have N proteins in an assembly, such that each protein i can be located in one of P positions x_{p}^{i}. N is ~ 10, P ~ 100. To each two vectors x_{i}^{p} and x_{j}^{q}, we can assign an energy function q_{0}, which is the overlap integral in the simplest approximation. Each protein position also has an associated score b_{0}. Thus, the optimal packing problem can be formulated as 
\begin{align}
x^{T}Q_{0}x+b_{0}^{T}x	&\rightarrow&	\textrm{min}\\
\textrm{w.r.t}.		&&\left\Vert x^{k}\right\Vert _{\infty}=1\;\forall k \\
	&&	x_{i}^{k}\geq0\;\forall i,k
\end{align}

  • Data: Collected using one of the standard complexes resolved using electron microscopy. The energy values and overlap integrals are calculated by modifying one of the standard packages, on example, HermiteFit. Data is generated in ~1 minute, code modification and data preparation will take ~1 week.
  • References: Yu.E. Nesterov Introduction to Convex Optimization (available at PreMoLab website)
  • Code notes: Implementation notes
  • Basic algorithm: I would like to try convex relaxations.
  • Novelty: Convex relaxations have not been used before in such Tasks on these proteins
  • consultant: Yu.V. Maksimov

Task 7

  • Name: Metric learning and space dimensionality reduction in Time Series Classification Tasks
  • Task: task statement from the base article, some modification of the error function is possible due to the specifics of the time series
  • Data: electricity price time series
  • References::
    1. [109] - basic article
    2. [110] - excellent overview of Metric Learning methods
    3. [111] - more overview
  • Basic algorithm: Frank-Wolf algorithm (conditional gradient descent)
  • Solution: apply target matrix decimation with Belsley method to remove multicollinearity
  • Novelty: application of Metric Learning methods in the problem of time series clustering, analysis of the properties of the proposed method
  • consultant: Alexander Katrutsa

Task 8

  • Name: Structural learning when generating models
  • Task: Solved by Task search ranking function in Information Search Tasks. The search is carried out among non-parametric functions (structures) generated by a grammar of the form G: g---> B(g, g) | U(g) | S, where B is a set of binary operations {+, -, *, /}, U - unary operations {-(), sqrt, log, exp}, S - variables and parameters {x, y, k}. It is proposed to solve the problem of generating a ranking model in two stages, using the history of restoring the structure of the model as a training sample.
  • Data: TREC subcollections.
  • Description of the collection of data used to evaluate the features, and the evaluation procedure. [112]
  • References:
  • Basic algorithm: Parantap, BM25 - models for comparison.
  • Solution: It is proposed to cluster the collection and generate models for document clusters. Then, using the structural learning method, find models that generalize the unions of clusters up to the collection itself.
  • Novelty: Ranking functions found that are as good as those used in practice.
  • * consultant: Anna Varfolomeeva, Oleg Bakhteev

Task 9

  • Name: Checking the compliance of the electrocardiograph with the requirements of the diagnostic system "Screenfax" and assessing the quality of electrocardiograms.
  • Task: The task of checking the compliance of an arbitrary electrocardiograph with the requirements of the "Screenfax" diagnostic system [1—4] is solved based on a comparison of electrocardiograms (ECG) of the same and the same patients recorded by both devices according to the ABAB scheme, where A is the first device, B - the second. The Task of automatic detection of low-quality electrocardiograms that do not meet the requirements of the diagnostic system is also solved.
  • Data: The selection consists of records with ECG values recorded by the device for which the test is being carried out, and by the device used in the Screenfax diagnostic system (data with a detailed description of the recording format will be provided to the person who selected the task). You can use http://www.physionet.org/physiobank/database/ptbdb/ to test algorithms for R-peak detection and noise level estimation.
  • References:
    1. Information portal of the Diagnostic system "Screenfax". URL: http://skrinfax.ru/method-author/
    2. Technology for information analysis of electrocardiosignals
    3. Uspensky V.M. Information function of the heart. Theory and practice of diagnosing diseases of internal organs by the method of information analysis of electrocardiosignals. M.: Economics and informatics, 2008. 116p.
    4. Uspensky V.M. Information function of the heart. // Clinical medicine. 2008. V.86. No. 5. pp.4–13.
    5. Naseri H., Homainezhad M.R. Electrocardiogram signal quality assessment using an artificially reconstructed target lead // Computer Methods in Biomechanics and Biomedical Engineering. 2015. Vol.18, No. 10.Pp. 1126-1141.
    6. Zidelmal Z., Amirou A., Ould-Abdeslam D., Moukadem A., Dieterlen A. QRS detection using S-Transform and Shannon energy. // Comput Methods Programs Biomed. 2014. Vol. 116, no. 1.Pp. 1-9. URL: https://yadi.sk/i/-kD00y1VepB3q
    7. Sarfraz M., Li F. F., Khan A. A. Independent Component Analysis Methods to Improve Electrocardiogram Patterns Recognition in the Presence of Non-Trivial Artifacts // Journal of Medical and Bioengineering. 2015. Vol. 4, no. 3.Pp. 221-226. URL: https://yadi.sk/i/-kD00y1VepB3q
    8. Meziane N. et al. Simultaneous comparison of 1 gel with 4 dry electrode types for electrocardiography // Physiol. Meas. 2015. Vol. 36, no. 513.
    9. Allana S., Aversa J., Varghese C., et al. Poor quality electrocardiograms negatively affect the diagnostic accuracy of ST segment elevation myocardial infarction. // J Am Call Cardiol. 2014. Vol. 63, no. 12_S. doi:10.1016/S0735-1097(14)60172-8.
  • Basic algorithm: ECG quality estimation – [4], R-peak detection – [5], noise level estimation in data – [6].
  • Solution: The task of checking the compliance of an arbitrary electrocardiograph with the requirements of the "Screenfax" diagnostic system is proposed to be solved by constructing permutation statistical tests by comparing the values of RR-intervals and R-amplitudes and detected code sequences (calculated by amplitudes and intervals) for each diseases. This is where the task of detecting R peaks comes in. In the task of detecting low-quality electrocardiograms, the Task of estimating the noise level arises. In addition, it is necessary to learn how to filter out ECG with non-informative amplitude values or a large spread of interval values, since the method of analyzing electrocardiographic signals is not applicable to the diagnosis of arrhythmia.
  • Novelty: The task of checking the compliance of the electrocardiograph with the requirements of the diagnostic system can be considered as the task of comparing ECG recording devices that arise, for example, when comparing different types of electrodes, and the noise level in the values of electrocardiosignals, the presence of baseline drift are selected as criteria and some other features [7].
  • consultant: Shaura Ishkina

Task 10

  • Name: Simplification of the IR models structure
  • Task: To achieve the acceptable quality of the information retrieval models, modern search engines use models of very complex structure. In current research we propose to simplify the model structure and make it interpretable without decreasing the model accuracy. To do this, we follow the idea from (Goswami et al., 2014) of constructing the set of nonlinear IR functions of simple structure and admissible accuracy. However, each of these functions is expected to have lower accuracy while comparing with the best IR model of complex structure. Thus, we propose to approximate this complex model with the linear combination of simple nonlinear functions and expect to obtain the comparable quality of solution.
  • Data: TREC collections.
  • References:
    • P. Goswami et Al. Exploring the Space of IR Functions // Advances in Information Retrieval. Lecture Notes in Computer Science. 8416:372-384, 2014.
    • problem statement
  • Basic algorithm: Gradient boosting machine for constructing a model of high complexity. Exaustive search of superpositions from a set of elementary functions for approximation and simplification.
  • Solution: The optimal functions for the linear combination can be found by the greedy algorithm.
  • Novelty: A new ranking function of simple structure competitive with traditional ones.
  • consultant: Mikhail Kuznetsov.

Task 11

  • Name: Testing non-parametric time series forecasting algorithms under non-stationary conditions
  • Task: One of the key assumptions about the distribution of data in non-parametric is the assumption that the time series is stationary. The adequacy of forecasts if this requirement is not met is not guaranteed. It is required to develop a method for determining the fulfillment of the condition of local stationarity of the time series to study the applicability of the main algorithms of nonparametric forecasting in the absence of stationarity. Consider the main methods of nonparametric regression, such as kernel smoothing, spline smoothing, autoregression, moving average, etc.
  • Data: Data on freight rail transportation (RZD)
  • References::
    • Valkov A.S., Kozhanov E.M., Medvednikova M.M., Khusainov F.I. Nonparametric forecasting of railway junction system load based on historical data // Machine Learning and Data Analysis. - 2012. - No. 4.
    • Dickey D. A. and Fuller W. A. Distribution of the Estimators for Autoregressive Time Series with a Unit Root / Journal of the American Statistical Association. - 74. - 1979. - p. 427--431.
  • Basic algorithm: ARMA, Hist.
  • Solution: Use the Dickey-Fuller test as a basic method for checking series for non-stationarity. It is also proposed to consider such sources of non-stationarity as trend and seasonality.
  • Novelty: A method for determining the fulfillment of the condition of local stationarity of a time series has been developed and substantiated.
  • consultant: Stenina Maria

Task 12

  • Name: Learning metrics in Full and Partial Learning Tasks
  • Task: is a software implementation of a complex of convex and DC-optimization methods for the problem of choosing the optimal metric in Tasks of recognition. In other words, in constructing a metric such that the nearest neighbor classification gives high accuracy.
  • Data: Birds and Fungus ImageNet collection with Deep features extracted (provided by consultant). Primary tests can be done on the data provided by here
  • References: References and a detailed description of the problem are given in file
  • Code notes: Implementation notes
  • Basic algorithm: 1) convex relaxation of the problem solved by an internal point through CVX 2) SVM on a modified sample consisting of pairs of objects
  • consultant: Yu.V. Maksimov

Task 13

  • Name: Building a hierarchical topic model of a large conference
  • Task: Every year, the program committee of a major EURO conference (more than 2000 reports) is faced with the task of building a hierarchical model of conference abstracts. Due to the fact that the structure of the conference changes little from year to year, it is proposed to build a thematic model of the future conference using expert models of conferences of previous years. This raises the following subtasks:
  1. Classification of abstracts of the new conference.
  2. Predicting changes in the structure of the conference.
  • Data: Abstracts and expert models of EURO 2010, 2012, 2013 conferences.
  • References:: Alexander A. Aduenko, Arsentii A. Kuzmin, Vadim V. Strijov. Adaptive thematic forecasting of major conference proceedings text of the article
  • Basic algorithm:
  • Solution: For solving subtasks
  1. it is proposed to combine the expert models of conferences of previous years into one, and for each thesis of a new conference to find the most suitable cluster in the resulting combined model, on example, using a weighted cosine measure of proximity.
  2. explore changes in the structure of conferences from year to year and determine the threshold of intra-cluster similarity values at which, for a certain set of abstracts, Experts create a new cluster, rather than adding these abstracts to existing clusters.
  • Novelty: A weighted cosine proximity measure that takes into account the hierarchical structure of clusters. Forecasting changes in the hierarchical structure/topics of the conference
  • consultant: Arsenty Kuzmin

Task 14

  • Name: Regularization of the linear naive bayes classifier.
  • Task: Building a linear classifier is one of the classic and most well studied machine learning tasks. A linear naive bayesian (LNB) classifier has the strong advantage that it builds in time that is linear in sample length, and the strong limitation that it assumes that the features are independent in its derivation. On some data, LNB performs surprisingly well, despite a clear violation of the feature independence hypothesis. The Linear Support Vector Machine (SVM) is considered to be a very successful method, but takes a long time on large samples. Both of these methods work in the same space of linear classifiers. The idea of the study is to bring LNB closer to SVM in terms of quality, but without loss of efficiency, by means of minor corrections.
  • Data: One of the three data sets, optional: classification of texts into scientific and non-scientific, classification of abstracts by fields of science, classification of ECG codograms for sick and healthy.
  • References::
    1. Larsen (2005) Generalized Naive Bayes Classifiers.
    2. Abraham, Simha, Iyengar (2009) Effective Discretization and Hybrid feature selection using Naïve Bayesian classifier for Medical datamining.
    3. Lutu (2013) Fast Feature Selection for Naive Bayes Classification in Data Stream Mining.
    4. Zaidi, Carman, Cerquides, Webb (2014) Naive-Bayes Inspired Effective Pre-Conditioner for Speeding-up Logistic Regression.
    5. + ask Vorontsov K. V.а.
  • Basic algorithm: any ready-made LNB and SVM implementations. Plus naive feature selection for LNB.
  • Solution: Derive correction formulas for LNB weights when using a margin-maximization regularizer similar to SVM. We build an iterative process in which a correction is calculated at each step, bringing the LNB closer to the SVM a little more. ROC-curves and dependences of Hold-out AUC on the iteration number are built.
  • Novelty: The ML community still hasn't realized that any linear classifier is equivalent to some kind of Naive Bayesian classifier.
  • consultant: Mikhail Uskov. Hyperconsultant: Vorontsov K. V..

Task 15

  • Name: Thematic model of the interests of regular users of the mobile application.
  • Task: The mobile app for learning English words offers the user words one by one. The user can either add a word to the studied ones, or discard it. To start learning words, you need to type at least 10 words. It is required to build a probabilistic word generation model that adapts to the interests of the user.
  • Data: There are lists of added and dropped words for each user. In addition, it is intended to use a large external collection of texts, for example, Wikipedia, for sustainable topic definition.
  • References::
    1. Vorontsov K. V., Potapenko A. A. Additive Regularization of Topic Models // Machine Learning. Special Issue “Data Analysis and Intelligent Optimization with Applications”. 2014. Russian translation
    2. + ask Vorontsov K.V.
  • Basic algorithm: Random word selection algorithm.
  • Solution: The topic model for each user determines the topic profile of his interests p(t|u). To generate words, word distributions from the distributions p(w|t) of the topics of the given user are used. Dependences of the quality functionals of the thematic model on the iteration number are constructed. The main functionality of quality is the ability of the model to predict which words the user will leave and which ones they will discard.
  • Novelty: A feature of the model is the presence of discarded words. The developed methods can also be applied in recommender systems with likes and dislikes.
  • consultant: Viktor Safronov. Hyperconsultant: Vorontsov K. V..

2015

Author Topic Link Consultant Reviewer DZ-1 DZ-2 (Problem number) Letters Sum Grade
Bernstein Julia Methods for Characterizing Fibrinolysis from Blood Image Sequence in vitro Matveev I. A. Solomatin 1 3 (8) AILSBRCVTDE 11 10
Bochkarev Artem Structural learning in model generation [113] (no code), paper, slides Varfolomeeva Anna, Oleg Bakhteev Isachenko 2 2 (7) A+I++LS+BRCVT+DS 9.25 10
Goncharov Alexey Metric classification of time series code,

paper, slides

Maria Popova Zadayanchuk 1.5 1 (4) AILSBRCVTDSW 12 10
Dvinskikh Darina Improving the quality of forecasting using product groups code,

paper, slides

Kanevsky D. Yu. Smirnov 0.5 3 (7) AILSBRCVTDEHS 14 10
Efimov Yuri Search for the outer and inner boundaries of the iris in the eye image using the paired gradient method code,

paper, slides

Matveev I. A. Neichev AILSBRCVTDEW 12 10
Zharikov Ilya Checking the compliance of the electrocardiograph with the requirements of the diagnostic system "Screenfax" and assessing the quality of electrocardiograms. code, paper, slides Shaura Ishkina Bochkarev 3.5 3 (5) AIL+SBRCVTDEHSW 14.25 10
Zadayanchuk Andrey Choosing the optimal physical activity classification model code,

paper, slides

Maria Popova Goncharov 2 0 (17) AI-LSB+RCVTD 10 10
Zlatov Alexander Building a hierarchical model of a large conference code,

paper, slides

Arsenty Kuzmin Dvinskyh 1.5 3 (14) AI+L+SBRC++V+TDESW 14.25 10
Isachenko Roman Metric Learning and Space Dimension Reduction in Tasks of Time Series Clustering code, paper, slides Alexander Katrutsa Zharikov 3.5 3 (14) A-I+L+S-BR+CVTDEHSW 14.25 10
Radoslav Neichev Feature Selection in Time Series Forecasting Using Exogenous Factors code, paper, slides Alexander Katrutsa Efimov 1 3 (9) AI-L-SBRCVTDEHSW 13.5 10
Podkopaev Alexander Prediction of Quaternary Structures of Proteins code,

paper, slides

Maksimov Yu. V. Reshetov 3.5 3 (11) AILS+B+RCVTDEHS 13.5 10
Reshetova Daria Multiclass Classification Methods with Improved Convergence Estimators in Partial Learning Tasks code,

paper, slides

Maksimov Yu. V. Kamzolov 2.5 3 (10) AIL++SB+RCVT++DEHS- 14 10
Smirnov Evgeniy Thematic model of interests of permanent users of the mobile application code, paper, slides Victor Safronov Zlatov 1 1 (4) AILSBRCVTWDE 11.25 10
Solomatin Ivan Determination of the iris shading area by the classifier of local textural features code, paper, slides Matveev I. A. Bernstein Julia 3 (9) AILSBRCVTDE 11 10
Chernykh Vladimir Testing nonparametric algorithms for time series forecasting under nonstationary conditions code,

paper, slides

Stenina Maria Shishkovets Svetlana 3.5 3 (4) A+I+LSBRCVT+DE++H++ 13.75 10
Shishkovets Svetlana Regularization of a linear naive bayes classifier. code,

paper, slides

Михаил Усков, Константин Воронцов Chernykh Vladimir 3.5 2 (9) A+I+L+SBR+CV+TD+E+H+S 15 10
Kamzolov Dmitry New algorithms for the problem of ranking web pages Alexander Gasnikov, Yuri Maksimov Podkopaev AILSB+RCVT+DEHS-- 13 8
Sukhareva Anzhelika Classification of scientific texts by branches of knowledge code,

paper, slides

Sergei Tsarkov 0.5 AILSBRCVTDEH 9

Task 1

  • Name: Improving the quality of demand forecasting using product groups
  • Task:

Given:

    1. Time series of sales for several product groups in one hypermarket. Also, for each product, periods of shortage, periods of influence on the demand of calendar holidays and periods of holding are known. marketing promotions. A product classifier is also known: a tree of product groups, where the products themselves are leaves.
    2. Forecasting algorithm that is used to generate demand forecasts for these products: self-adaptive exponential smoothing (Trigg-Leach model, see [1])
    3. Loss function by which the quality of forecasts is measured: MAPE.
    4. Requirements for building forecasts: forecasts must be built weekly for 4 weeks ahead (at the beginning of the current week, you need to build a forecast of total demand for the next week, a week in one, two, and 3).

Hypothesis: Demand for individual goods is too volatile to reveal their characteristic seasonality. It is proposed to use data on product groups in order to more accurately determine the parameters of seasonality. Note: there are other options for improving the quality of forecasting by working with groups of goods. Task is to improve the quality of forecasting within the framework of the task by taking into account the effect of the interchangeability of goods, in comparison with the basic algorithm. The result can be considered achieved if a statistically significant increase in quality is shown when building a series of forecasts (at least 20) for each time series using a sliding control.

  • Data:
    1. Data on sales of several product groups in a hypermarket of a large retail chain: https://drive.google.com/file/d/0B5YjPespcL83X3pHaE1aRzBUaDg/view?usp=sharing
  • References:
    1. Lukashin Yu. P. Adaptive methods of short-term forecasting of time series. - M .: Finance and statistics, 2003.
    2. http://www.machinelearning.ru/wiki/index.php?title=%D0%9C%D0%BE%D0%B4%D0%B5%D0%BB%D1%8C_%D0%A2%D1 %80%D0%B8%D0%B3%D0%B3%D0%B0-%D0%9B%D0%B8%D1%87%D0%B0
    3. Nitin Patel, Mahesh Kumar, Rama Ramakrishnan. Clustering models to improve forecasts in retail merchandising. http://www.cytel.com/Papers/INFORMS_Prac_%2004.pdf
    4. Kumar M., Error-based Clustering and Its Application to Sales Forecasting in Retail Merchandising. PhD Thesis. http://books.google.ru/books/about/Error_based_Clustering_and_Its_Applicati.html?id=6252NwAACAAJ&redir_esc=y
  • Basic algorithm: It is proposed to use the seasonality model [3] in combination with the Trigg-Leach model as a non-seasonal series prediction algorithm ([1] and [2]). In this case, 3 variants of the algorithm are possible, depending on the method of assessing seasonality:
    1. Seasonality is estimated by the very series of sales. For products with a "short" history, seasonality is not assessed.
    2. Seasonality is estimated for a group of goods, based on the classifier of commodity groups (lower level of the classifier)
    3. Seasonality is estimated by clusters, based on the methodology [3], [4].
  • Solution: It is required to implement the combination of the seasonality model [3] and the Trigg-Leach model as a non-seasonal series prediction algorithm ([1] and [2]), with the 3 variants of seasonality analysis described above. When constructing seasonal profiles, it is necessary to exclude periods of marketing campaigns (otherwise, there may be a significant distortion of seasonality). Next, you need a series of experiments with quality analysis on real data. When analyzing quality, you can exclude periods of holidays and marketing campaigns. Based on the results of the experiments, it may be necessary to adapt the clustering algorithm.
  • Novelty: Building a self-adaptive forecasting algorithm taking into account seasonality, identified by cluster analysis.
  • consultant: Kanevsky D.Yu.

Task 2

Task 3

  • Name: Obtaining an estimate of the sparse covariance matrix for nonlinear models (neural networks).
  • Task: Suggest a method for estimating the covariance matrix of parameters of a general model for the case of linear regression, logistic regression, general non-linear models, including neural networks. Suggest a way to take into account the structure of the matrix (sparseness, dependencies between coefficients, etc.)
  • Data: Synthetic data and tests.
  • References::
  • Basic algorithm: Diagonal matrix estimation, see MLAlgorithms/HyperOptimization folder.
  • Solution:
  • Novelty: A fast algorithm for obtaining estimates of the general covariance matrix for nonlinear models is proposed, the properties of sparse matrices are investigated.
  • consultant: Alexander Aduenko.

Task 6

  • Name: Prediction of Quaternary Structures of Proteins: leveling
  • Task: The task is to predict the packing of protein molecules into a multimeric complex in the rigid body approximation. One of the formulations of the problem is written as a non-convex optimization.

It is necessary to study this formulation and propose a solution algorithm. Suppose we have N proteins in an assembly, such that each protein i can be located in one of P positions x_{p} ^{i}. N is ~ 10, P ~ 100. To each two vectors x_{i}^{p} and x_{j} ^{q}, we can assign an energy function q_{0}, which is the overlap integral in the simplest approximation. Each protein position also has an associated score b_{0}. Thus, the optimal packing problem can be formulated as 
\begin{align}
x^{T}Q_{0}x+b_{0}^{T}x	&\rightarrow&	\textrm{min}\\
\textrm{w.r.t}.		&&\left\Vert x^{k}\right\Vert _{\infty}=1\;\forall k \\
	&&	x_{i}^{k}\geq0\;\forall i,k
\end{align}

  • Data: Collected using one of the standard complexes resolved using electron microscopy. The energy values and overlap integrals are calculated by modifying one of the standard packages, on example, HermiteFit. Data is generated in ~1 minute, code modification and data preparation will take ~1 week.
  • References: Yu.E. Nesterov Introduction to Convex Optimization (available at PreMoLab website)
  • Code notes: Implementation notes
  • Basic algorithm: I would like to try convex relaxations.
  • Novelty: Convex relaxations have not been used before in such Tasks on these proteins
  • consultant: Yu.V. Maksimov

Task 8

  • Name: Structural learning when generating models
  • Task: Solved by Task search ranking function in Information Search Tasks. The search is carried out among non-parametric functions (structures) generated by a grammar of the form G: g---> B(g, g) | U(g) | S, where B is a set of binary operations {+, -, *, /}, U - unary operations {-(), sqrt, log, exp}, S - variables and parameters {x, y, k}. It is proposed to solve the problem of generating a ranking model in two stages, using the history of restoring the structure of the model as a training sample.
  • Data: TREC subcollections.
  • Description of the collection of data used to evaluate the features, and the evaluation procedure. [114]
  • References:
  • Basic algorithm: Parantap, BM25 - models for comparison.
  • Solution: It is proposed to cluster the collection and generate models for document clusters. Then, using the structural learning method, find models that generalize the unions of clusters up to the collection itself.
  • Novelty: Ranking functions found that are as good as those used in practice.
  • * consultant: Anna Varfolomeeva, Oleg Bakhteev

Task 9

  • Name: Checking the compliance of the electrocardiograph with the requirements of the diagnostic system "Screenfax" and assessing the quality of electrocardiograms.
  • Task: The task of checking the compliance of an arbitrary electrocardiograph with the requirements of the "Screenfax" diagnostic system [1—4] is solved based on a comparison of electrocardiograms (ECG) of the same and the same patients recorded by both devices according to the ABAB scheme, where A is the first device, B - the second. The Task of automatic detection of low-quality electrocardiograms that do not meet the requirements of the diagnostic system is also solved.
  • Data: The selection consists of records with ECG values recorded by the device for which the test is being carried out, and by the device used in the Screenfax diagnostic system (data with a detailed description of the recording format will be provided to the person who selected the task). You can use http://www.physionet.org/physiobank/database/ptbdb/ to test algorithms for R-peak detection and noise level estimation.
  • References:
    1. Information portal of the Diagnostic system "Screenfax". URL: http://skrinfax.ru/method-author/
    2. Technology for information analysis of electrocardiosignals
    3. Uspensky V.M. Information function of the heart. Theory and practice of diagnosing diseases of internal organs by the method of information analysis of electrocardiosignals. M.: Economics and informatics, 2008. 116p.
    4. Uspensky V.M. Information function of the heart. // Clinical medicine. 2008. V.86. No. 5. pp.4–13.
    5. Naseri H., Homainezhad M.R. Electrocardiogram signal quality assessment using an artificially reconstructed target lead // Computer Methods in Biomechanics and Biomedical Engineering. 2015. Vol.18, No. 10.Pp. 1126-1141.
    6. Zidelmal Z., Amirou A., Ould-Abdeslam D., Moukadem A., Dieterlen A. QRS detection using S-Transform and Shannon energy. // Comput Methods Programs Biomed. 2014. Vol. 116, no. 1.Pp. 1-9. URL: https://yadi.sk/i/-kD00y1VepB3q
    7. Sarfraz M., Li F. F., Khan A. A. Independent Component Analysis Methods to Improve Electrocardiogram Patterns Recognition in the Presence of Non-Trivial Artifacts // Journal of Medical and Bioengineering. 2015. Vol. 4, no. 3.Pp. 221-226. URL: https://yadi.sk/i/-kD00y1VepB3q
    8. Meziane N. et al. Simultaneous comparison of 1 gel with 4 dry electrode types for electrocardiography // Physiol. Meas. 2015. Vol. 36, no. 513.
    9. Allana S., Aversa J., Varghese C., et al. Poor quality electrocardiograms negatively affect the diagnostic accuracy of ST segment elevation myocardial infarction. // J Am Call Cardiol. 2014. Vol. 63, no. 12_S. doi:10.1016/S0735-1097(14)60172-8.
  • Basic algorithm: ECG quality estimation – [4], R-peak detection – [5], noise level estimation in data – [6].
  • Solution: The task of checking the compliance of an arbitrary electrocardiograph with the requirements of the "Screenfax" diagnostic system is proposed to be solved by constructing permutation statistical tests by comparing the values of RR-intervals and R-amplitudes and detected code sequences (calculated by amplitudes and intervals) for each diseases. This is where the task of detecting R peaks comes in. In the task of detecting low-quality electrocardiograms, the Task of estimating the noise level arises. In addition, it is necessary to learn how to filter out ECG with non-informative amplitude values or a large spread of interval values, since the method of analyzing electrocardiographic signals is not applicable to the diagnosis of arrhythmia.
  • Novelty: The task of checking the compliance of the electrocardiograph with the requirements of the diagnostic system can be considered as the task of comparing ECG recording devices that arise, for example, when comparing different types of electrodes, and the noise level in the values of electrocardiosignals, the presence of baseline drift are selected as criteria and some other features [7].
  • consultant: Shaura Ishkina

Task 12

  • Name: Learning metrics in Full and Partial Learning Tasks
  • Task: is a software implementation of a complex of convex and DC-optimization methods for the problem of choosing the optimal metric in Tasks of recognition. In other words, in constructing a metric such that the nearest neighbor classification gives high accuracy.
  • Data: Birds and Fungus ImageNet collection with Deep features extracted (provided by consultant). Primary tests can be done on the data provided by here
  • References: References and a detailed description of the problem are given in file
  • Code notes: Implementation notes
  • Basic algorithm: 1) convex relaxation of the problem solved by an internal point through CVX 2) SVM on a modified sample consisting of pairs of objects
  • consultant: Yu.V. Maksimov

Plans for next year:

  1. Expand the matlab test and give it along with the trial programming as the first task.

2014

Author Topic Link Consultant DZ-1 Letters Sum Grade
Gazizullina Rimma Forecasting the volume of rail freight traffic by pairs of branches [115], pdf Stenina Maria \frac{15}{15}+\frac{10}{16} [MF]TAI+L+SBR+CV+T>DEH(J) 16 10
Grinchuk Alexey Selection of Optimal Structures of Predictive Models by Structural Learning Methods [116], pdf Varfolomeeva Anna \frac{7}{15}+\frac{2}{16} [F]TA+I+LSBR+СV+T+D+E(F) 14,5 9
Gushchin Alexander Sequential Generation of Essentially Nonlinear Models in tasks of Document Ranking [117], pdf Kuznetsov Mikhail \frac{5}{15}+\frac{2}{16} [F]TAI+L+SBRCVTDEHS(F) 15,5 9
Efimova Irina Differential diagnosis of diseases by electrocardiogram [118], pdf Vlada Tselykh \frac{15}{15}+\frac{12}{16} [MF]T+A+I+L+SB++R+CV+TDE+H(J ed) 17,25 10
Zhukov Andrey Building University Rankings: Panel Analysis and Sustainability Assessment [119], pdf Kuznetsov Mikhail \frac{8}{15}+0 [F]TAIL+SBRCVTDEHS(F) 15,25 9
Ignatov Andrey Manifold training for predicting sets of quasi-periodic time series [120], pdf Ivkin Nikita 0+\frac{7}{16} [MF]TA+I+L+S+B+R+C+VTD>E+HS (J if ed) 18 10
Karasikov Mikhail Search for effective methods of dimensionality reduction in solving problems of multiclass classification by reducing it to solving binary problems [121], pdf Yu.V. Maksimov 0+0 [MF]TAI+L+SBRC+V+TDESH(J) 15 10
Kulunchakov Andrey Detecting Isomorphic Structures of Essentially Nonlinear Predictive Models [122], pdf Sologub Roman, Kuznetsov Mikhail \frac{10}{15}+\frac{14}{16} [F]T+AI+L+S+BR+CVT++D+EHS(J ed-ed) 17 10
Lipatova Anna Detecting Patterns in a Set of Time Series by Structural Learning Methods [123], pdf A. P. Motrenko \frac{8}{15}+\frac{6}{16} [MF]TA+I+LSBR-CVTDE (J when ed) 14,25 10
Makarova Anastasia Using non-linear forecasting when looking for dependencies between time series [124], pdf A. P. Motrenko 0+0 [F]TAI-LSB+R-CVTD>E>(F) 12,75 9
Plavin Alexander Optimizing the Number of Topics in Probabilistic Topic Models with a String Sparse Regularizer [125], pdf Potapenko Anna \frac{13}{15}+\frac{14}{16} [F]T+A+I+L+S+BR++CVTD+>>(?) 14 10
Maria Popova Choosing the optimal model for predicting human physical activity based on accelerometer measurements [126], pdf Tokmakova Alexandra \frac{11}{15}+\frac{6}{16} [MF]T+AI+L++SB++R+CV+TD+(JV ed) 15,25 10
Shvets Mikhail Interpretation of multimodels in the processing of sociological data [127], pdf Alexander Aduenko \frac{11}{15}+\frac{4}{16} [M+F]T+A+I+L+S+B+R+CVTD+E(F) 16,25 9
Shinkevich Mikhail Influence of sparse, smoothing and decorrelation regularizers on the stability of a probabilistic topic model [128], pdf Dudarenko Marina \frac{15}{15}+\frac{9}{16} [MF]T+AIL+S+BR+CV+T+D+E+H(J ed) 17 10

1. Optimizing the Number of Topics in Probabilistic Topic Models with a String Sparse Regularizer

consultant: Potapenko Task: The probabilistic topic model describes the probabilities of occurrence of words w\in W in documents d\in D through latent topics t\in T< /text>:
</p><p><tex> p(w|d) = \sum_{t\in T} p(w|t)p(t|d) = \sum_{t\in T} \phi_{wt}\theta_{td}.

We need to test the hypothesis that by imposing constraints on the \Theta matrix using the string sparse regularizer, it is possible to determine the optimal number of topics.

Data: The collection of documents is specified by word frequencies. Since to solve the problem it is necessary to know the <<true>> number of topics, experiments are performed on realistic model or semi-model data.

References:

cops // Reports of the Russian Academy of Sciences. 2014. - V. 455, No. 3 (in press).

  • Vorontsov K. V. Probabilistic thematic modeling. — 2014.

http://www.MachineLearning.ru/wiki/images/2/22/Voron-2013-ptm.pdf

  • Teh Y. W., Jordan M. I., Beal M. J., Blei D. M. Hierarchical Dirichlet processes // Journal of the

American Statistical Association. - 2006. - Vol. 101, no. 476.-Pp. 1566–1581

Basic algorithm: Regularized EM-algorithm [2014: Vorontsov] is used to solve the optimization problem. A rational, stochastic or online version of the EM algorithm can be used.

Novelty: Dirichlet's HDP [2006: Teh et Al] hierarchical process model is commonly used to optimize the number of topics. It determines the number of topics is unstable, and at the same time it is difficult both to understand and to implement. Additive Regularization of Topic Models (ARTM) is a new approach to topic modeling that combines versatility, flexibility and simplicity. The task of optimizing the number of topics has not yet been considered in the framework of ARTM.

2. Differential diagnosis of diseases by electrocardiogram

consultant: Vlada Tselykh

Task: It is proposed to solve a typical classification problem. Signs are 216 characteristics calculated from the electrocardiogram. It is necessary to evaluate the quality of the classification on a delayed control sample. To do this, the fractions of errors of the first and second kind are calculated. Under the error of the first kind is meant the assignment of healthy people to the class of patients, the second kind - the assignment of patients to the class of healthy people. Preference is given to minimizing Type II errors.

Data: For each of the 5 diseases, there are 2 types of samples. Reference - more reliable, specially selected cases. The rest are cases when the diagnoses were established by doctors less reliably; these samples are proposed to be used for control.

References:

  • Vorontsov K. V. Metric classification algorithms. Lectures on machine learning. — 2014. http://www.MachineLearning.ru/wiki/images/c/c3/Voron-ML-Metric-slides.pdf
  • Uspensky V. M. Information function of the heart // Clinical Medicine, 2008. - V. 86, No. 5. - P. 4–13.
  • Uspensky V. M. Information function of the heart. Theory and practice of diagnosing diseases of internal organs by the method of information analysis of electrocardiosignals. - M .: "Economy and information", 2008. - 116 p.

Basic algorithm: To solve the problem, it is proposed to use a metric algorithm with greedy feature selection.

Novelty: The data were prepared using a unique technology for information analysis of electrocardiosignals, developed by prof. MD V.M.Uspensky. A classification algorithm is proposed and its generalizing ability is investigated.

3. Influence of sparse, smoothing and decorrelation regularizers on the stability of a probabilistic topic model

consultant: M.A. Dudarenko

Task:Probabilistic topic model describes the probabilities of occurrence of words w\in W in documents d\in D through latent topics t\in T< /text>:
</p><p><tex> p(w|d) = \sum_{t\in T} p(w|t)p(t|d) = \sum_{t\in T} \phi_{wt}\theta_{td}.

Matrix representation \|p(w|d)\|_{W\times D} as a product of two smaller matrices {\Phi=\|\phi_{wt}\|_{W\times T}} and {\Theta=\|\theta_{dt} \|_{T\times D}} is not the only one: \Phi \Theta = (\Phi S)(S^{-1}\Theta) = \Phi'\Theta' for some non-degenerate S. It is required to test the hypothesis that, by imposing restrictions on the matrices \Phi, \Theta using regularizers, it is possible to increase the stability of their recovery.

Data: The collection of documents is specified by word frequencies. Because for to solve the problem, it is necessary to know the “true” matrices \Phi, \Theta, experiments are performed on realistic model or semi-model data that satisfy the hypotheses of sparseness, weak correlation of topics and the presence of background topics.

References:

Basic algorithm: Regularized EM-algorithm [2014: Vorontsov] is used to solve the optimization problem. A rational, stochastic or online version of the EM algorithm can be used.

Novelty: Additive Regularization of Topic Models (ARTM) was proposed in [2014: Vorontsov] as a universal way to improve the stability and interpretability of topic models. However, the question of which particular combination of regularizers increases stability remains open. This study is aimed at solving this problem.

4. Building University Rankings: Panel Analysis and Sustainability Assessment

consultant: Kuznetsov Mikhail

Task: University ranking changes from year to year. This change may be due to the poor quality of the ranking calculation methodology, random changes in the institution's performance, and purposeful changes in the state of the institution. It is required to propose such a rating method that is resistant to random changes, which would allow interpreting the change in the state of the university.

Data: Eight years of data for the world's top 100 universities.

References:

  • Strizhov V.V. Refinement of expert assessments using measured data. Zavodskaya lab. Diagnostics of materials, 2006, 72(7) - 59-64.
  • Strizhov V.V. Refinement of Expert assessments in rank scales using measured data. Zavodskaya lab. Diagnostics of materials, 2011, 77(7) - 72-78.
  • Kuznetsov M.P., Strijov V.V. Methods of expert estimations concordance for integral quality estimation // Expert Systems with Applications, 2014.
  • Draft POF article on request.

Basic algorithm: A method for constructing the RUR rating and one of the redundantly stable algorithms for ranking scales.

Novelty: Introduced the concept of interpretability of the change in the rating position. The Task of choosing and optimal locally monotonous correction of indicators was solved. A technique for constructing a rating is proposed that allows interpreting the change in the state of a university for the purpose of monitoring. Option: solved the reverse Task of management: how to change the indicators of the university in order to achieve a given goal.

5. Detecting Patterns in a Set of Time Series by Structural Learning Methods

consultant: A. P. Motrenko

Task: To improve the quality of the time series forecast, I would like to use expert statements about the presence of a causal relationship between events. To do this, it is necessary to be able to assess the reliability of expert statements. It is impossible to prove the existence of a causal relationship by statistical methods. The researcher can only check the presence of a certain structure of communication. The purpose of the task is, based on expert statements about the presence of a connection between events, to examine the time series for the presence of various structural connections and find the structure that is most consistent with the Expert's opinion.

References:

  • R. B. Kline, Principles and Practice of Structural Equation Modeling. New York: Guilford. 2005.
  • J. Pearl, Graphs, Causality and Structural Equation Models. Sociological Methods and Research, 27-2(1998), 226-284.
  • J. Pearl, E. Bareinboim, Transportability of Causal and Statistical Relations: A Formal Approach // Proceedings of the 25th AAAI Conference on Artificial Intelligence, August 7-11, 2011, San Francisco. 247-254
  • Valkov A.S., Kozhanov E.M., Motrenko A.P., Khusainov F.I. Construction of cross-correlation dependences in the forecast of load of the railway junction // Machine learning and data analysis. 2013. T. 1, No. 5. C. 505-518.
  • Valkov A.S., Kozhanov E.M., Medvednikova M.M., Khusainov F.I. Nonparametric forecasting of railway junction system load based on historical data // Machine Learning and Data Analysis. 2012. T. 1, No. 4. C. 448-465.

Basic algorithm: structural equation modeling, SEM

Novelty: A method for assessing the reliability of Expert statements about the impact of exchange prices on major instruments on the volume of rail freight traffic is proposed. Various structures of links between time series are proposed. The concept of structure complexity is introduced. The relationship between the complexity of the structure and the assessment of the reliability of the statement is investigated.

18. Using non-linear forecasting when looking for dependencies between time series

consultant: A. P. Motrenko

Task: (As part of a study devoted to the discovery of patterns in time series sets) It is proposed to abandon the standard assumptions about the stationarity of the time series when searching for dependencies between time series and to study time series from the point of view of dynamical systems theory, within which irregular time dependences determined by the structure of the phase space are considered. It is required to study a set of approaches to the analysis of dynamic data and the identification of relationships between them; describe the limits of applicability of the basic algorithm and propose new options for the revealed structural relationships. Data: Synthetic data, historical stock prices for major instruments and rail freight data.

References:

  • Tools for the Analysis of Chaotic Data. HENRY D. I. ABARBANEL
  • Nonlinear forecasting as a way of distinguishing chaos from measurement error in time series, G. Sugihara, R.M. May.
  • George Sugihara et al. Detecting Causality in Complex Ecosystems. Science 338, 496 (2012);
  • Valkov A.S., Kozhanov E.M., Motrenko A.P., Khusainov F.I. Construction of cross-correlation dependences in the forecast of load of the railway junction // Machine learning and data analysis. 2013. T. 1, No. 5. C. 505-518.
  • Valkov A.S., Kozhanov E.M., Medvednikova M.M., Khusainov F.I. Nonparametric forecasting of railway junction system load based on historical data // Machine Learning and Data Analysis. 2012. T. 1, No. 4. C. 448-465.

Basic algorithm: convergent cross mapping

Novelty: Proposed different structures of relationships between time series and a method for checking the existence of relationships

6. Sequential Generation of Essentially Nonlinear Models in tasks of Document Ranking

consultant: Kuznetsov Mikhail

Task: Propose and test on test and real data an algorithm for generating essentially non-linear models. The algorithm should generate 1) a complete set of models 2) choose the optimal step for a fixed model structure (adding a superposition element).

Data: Synthetic data, data for LIG text collections.

References:

  • Goswami P., Moura1 S., Gaussier E., Amini M.R. Exploring the Space of IR Functions //
  • Ore G.I., Strizhov V.V. Algorithms for the inductive generation of superpositions for the approximation of measured data // Informatics and its applications, 2013, 7(1) - 17-26.
  • Ore G.I., Strizhov V.V. Simplification of superpositions of elementary functions with the help of graph transformations according to the rules // Intellectualization of information processing. Reports of the 9th international conference, 2012 - 140-143.
  • Vladislavleva E., Smith G., Hertog D., Order of Nonlinearity as a Complexity Measure for Models Generated by Symbolic Regression via Pareto Genetic Programming // IEEE Transactions on Evolutionary Computation, 2009. Vol. 13(2). pp. 333-349.
  • Vladislavleva E. Model-based Problem Solving through Symbolic Regression via Pareto Genetic Programming: PhD thesis, Tilburg University, Tilburg, the Netherlands, 2008.

Basic algorithm: An exhaustive enumeration algorithm for admissible superpositions of generating functions.

Novelty: An algorithm for sequential addition of superposition elements is proposed. A function of the distance between superpositions is proposed and its properties are investigated. The notion of superposition complexity and the notion of adjacent superpositions that differ in complexity by one are introduced. An algorithm for generating adjacent superpositions is proposed.

7. Detecting Isomorphic Structures of Essentially Nonlinear Predictive Models

consultant: Sologub Roman, Kuznetsov Mikhail

Task: Develop an algorithm for finding isomorphic subgraphs for trees (a variant - for directed acyclic graphs). Compare the complexity of the algorithm for checking the isomorphism of two superpositions for the proposed algorithm and for the algorithm for element-by-element comparison of mappings.

Data: Data on exchange options: dependence of option volatility on the price and time of its execution.

References:

  • Ore G.I., Strizhov V.V. Algorithms for the inductive generation of superpositions for the approximation of measured data // Informatics and its applications, 2013, 7(1) - 17-26.
  • Ore G.I., Strizhov V.V. Simplification of superpositions of elementary functions with the help of graph transformations according to the rules // Intellectualization of information processing. Reports of the 9th international conference, 2012 - 140-143.
  • Ehrig H., Ehrig G., Prange U., Taentzer. G. Fundamentals of Algebraic Graph Transformation. Springer, 2006.
  • Ehrig H., Engels G. Handbook of Graph Grammars and Computing by Graph Transformation. World Scientific Publishing, 1997.
  • Strizhov V.V., Sologub R.A. Inductive generation of regression models of implied volatility for option trading // Computational technologies, 2009, 14(5) — 102-113.

Basic algorithm: Algorithm for element-by-element comparison of mappings.

Novelty: A fast algorithm for simplifying superpositions and searching for isomorphic models is proposed. The incidence matrix of the set of generating functions is used.

8. Building predictive models as superpositions of expert-specified functions

consultant: Ivkin Nikita

Task: Required to assign a set of time series to one of several classes. It is proposed to do this using the automated feature generation procedure. To do this, Expert creates a set of generating functions that 1) transform the time series (by example, smooth, decompose into principal components), 2) extract its aggregated descriptions from the time series (by example, mean, variance, number of extrema). It is possible to generate a significant number of features by constructing superpositions of generating functions. The resulting features are used to classify a set of time series (for example, by the nearest neighbor method).

Data: data from the mobile phone's accelerometer.

References:

  • Problem statement \MLAlgorithms\Group074\Kuznetsov2013SSAForecasting\doc
  • Khaikin S. Neural networks. Williams, 2006.

Basic algorithm: neural network (option: deep learning neural network).

Novelty: A method for extracting features using automatically constructed superpositions of Expert-specified functions is proposed.

Comparison of structural and topological complexity in Task classification.

9. Manifold training for predicting sets of quasi-periodic time series

consultant: Ivkin Nikita

Task: The task of classifying human activity based on data from the mobile phone's accelerometer is solved. Data from the accelerometer are represented by quasi-periodic time series. It is required to attribute the time series to one of the types of activity: running, walking, etc. To solve the problem of classifying series, a method based on nearest neighbors in the space of manifolds is proposed.

Data: data from the mobile phone's accelerometer.

References:

  • Mi Zhang; Sawchuk, A.A., "Manifold Learning and Recognition of Human Activity Using Body-Area Sensors," Machine Learning and Applications and Workshops (ICMLA), 2011 10th International Conference on , vol.2, no., pp.7,13, 18- 21 Dec. 2011

Basic algorithm: neural network

Novelty: proposed a method for classifying quasi-periodic time series based on manifolds

10. Interpretation of multimodels in the processing of sociological data

consultant: Alexander Aduenko

Task: The task of credit scoring is to determine the level of creditworthiness of the borrower who applied for a loan. To do this, a borrower's questionnaire is used, containing both numerical data (age, income, time of residence in the country) and categorical features (gender, profession). It is required, having historical information on loan repayments by other borrowers, to determine whether the client in question will return the loan. Thus, it is required to solve the problem of classification. Since the data can be heterogeneous (for example, if there are different income regions in the country), the data can be described not by one, but by several models. In this paper, we propose to compare two methods for constructing multimodels: mixtures of logistic models and gradient boosting.

Data: data on consumer loans (\mlalgorithms\BSThesis\Aduenko2013\data).

References:

  • model blends (\mlalgorithms\BSThesis\Aduenko2013\doc, Bishop)
  • boosting (lecture "Compositional methods of classification and regression" by Vorontsov)

Basic algorithm: boosting.

Novelty: Identification and explanation of similarities and differences between the solutions obtained by the two specified algorithms.

11. Selection of Optimal Structures of Predictive Models by Structural Learning Methods

consultant: Varfolomeeva Anna

Task: It is proposed to solve the problem of forecasting in two stages: first, the structure of the predictive model is restored using the stories of constructing successful forecasts. The model parameters are then optimized; using the model, a time series forecast is built.

Data: synthetic sample, biomedical time series, accelerometer measurements.

References:

Basic algorithm: the metaprediction algorithm described in the thesis.

Novelty: A method for restoring model structures using a priori assumptions about these structures is proposed.

12. Invariants in Predicting Quasi-Periodic Series

consultant: Arsenty Kuzmin

Task: The Task of hourly price/electricity consumption forecasting for the day ahead is being solved. When constructing the plan matrix, it is proposed to use not the original segment of the time series, but its invariant representation.

Data: hourly data on electricity prices and volumes (insert link).

References:

  • Sandulyanu L.N., Strizhov V.V. Feature Selection in Autoregressive Forecasting Tasks // Information Technologies, 2012, 7 — 11-15.
  • (taken from Fadeev's last article)

Basic algorithm: autoregressive prediction described in Sanduleanu's work.

Novelty: An algorithm for joint estimation of the parameters of the invariants and autoregressive model is proposed, which makes it possible to significantly improve the accuracy of forecasting.

13. Forecasting the volume of rail freight traffic by pairs of branches

consultant: Stenina Maria (Medvednikova)

Task: Predict traffic volumes from branch to branch, compare with the basic algorithm for predicting the departure of wagons from branch. Test the hypothesis that the traffic forecast from branch to branch is more accurate than the forecast using the basic algorithm. Examine series for trend/periodicity. If there is a trend/periodicity, then include it in the model. Prepare a prediction algorithm for use.

Data: daily data for a year and a half on the transportation of 38 types of cargo in the Omsk region.

References:

  • Valkov A.S., Kozhanov E.M., Medvednikova M.M., Khusainov F.I. Nonparametric forecasting of railway junction system load based on historical data // Machine Learning and Data Analysis. - 2012. - No. 4.

Basic algorithm: histogram prediction described in the article.

Novelty: it is proposed to improve the quality of the forecast by dividing the data into smaller parts and forecast traffic for specific branches instead of forecasting the departure of wagons.

14. Choosing the optimal model for predicting human physical activity based on accelerometer measurements

consultant: Tokmakova Alexandra

Task: Suggest an algorithm for sequential modification of the neural network. The goal is to find the most simple, stable and accurate network configuration that allows solving the problem of two-class (variant: multi-class) physical activity prediction.

Data: Set of time series of accelerometer measurements.

References:

  • Decimation of neural families on Machinelearning.ru.
  • Khaikin S. Neural networks. Williams, 2006.

Basic algorithm: Optimal Brain Damage/Optimal Brain Surgery.

Novelty: A method for sequential generation of neural networks of optimal complexity is proposed. The stability of generated models is studied.

15. Time Series Metaprediction

consultant: A.S. Inyakin, Ivkin Nikita

Task: A set of time series forecasting algorithms is specified. According to the presented time series, it is required to indicate the algorithm that delivers the most accurate forecast. In this case, the algorithm itself is not supposed to be executed. To solve this problem, it is proposed to build a set of features that describe the Expert time series, but a set of generating functions is created that 1) transform the time series (by example, smooth, decompose into principal components), 2) extract its aggregated descriptions from the time series (by example, mean, variance , the number of extrema). It is possible to generate a significant number of features by constructing superpositions of generating functions.

Data: Library of quasi-periodic and aperiodic time series

References:

  • Kuznetsov M.P., Mafusalov A.A., Zhivotovsky N.K., Zaitsev E., Sungurov D.S. Smoothing forecasting algorithms // Machine learning and data analysis. 2011. T. 1, No. 1. C. 104-112.
  • Fadeev I.V., Ivkin N.P., Savinov N.A., Kornienko A.I., Kononenko D.S., Dzhamtyrova R.B. Autoregressive forecasting algorithms // Machine learning and data analysis. 2011. T. 1, No. 1. C. 92-103.

Basic algorithm: Use the SAS/SPSS algorithm.

Novelty: A method for fast selection of the optimal predictive algorithm based on the description of the time series is proposed.

16. Identification of a person by the image of the iris

consultant: Matveev I. A.

Task: In the problem of identifying a person by the image of the iris (iris), the most important role is played by the selection of the region of the iris in the original image (segmentation of the iris). However, the iris image is usually partially obscured (shaded) by eyelids, eyelashes, highlights, that is, part of the iris cannot be used for recognition and moreover, the use of data from shaded areas can generate false signs and reduce accuracy. Therefore, one of the important steps in the segmentation of the iris image is the rejection of shaded areas.

Data: bitmap monochrome image, typical size 640*480 pixels (however, other sizes are possible) and coordinates of centers and radii of two circles approximating pupil and iris.

References:

  • Problem description and proposed solutions
  • Monro D. University of Bath Iris Image Database // http:// www.bath.ac.uk/ elec-eng/ research/ sipg/ irisweb/
  • Chinese academy of sciences institute of automation (CASIA) CASIA Iris image database // http://www.cb-sr.ia.ac.cn/IrisDatabase.htm, 2005.
  • MMU Iris Image Database: Multimedia University // http://pesonna.mmu.edu.my/ccteo/
  • Phillips P.J., Scruggs W.T., O'Toole A.J. et al. Frvt2006 and ice2006 large-scale experimental results // IEEE PAMI. 2010. V. 32. No. 5. P. 831–846.
  • G.Xu, Z.Zhang, Y.Ma Improving the performance of iris recognition system using eyelids and eyelashes detection and iris image enhancement // Proc. 5Th Int. Conf. Cognitive Informatics. 2006. P.871-876.

Basic algorithm: method using sliding window and texture features [2006: Xu, Zhang, Ma].

Novelty: the mask of the open area of the iris has been built.

17. Search for effective methods of dimensionality reduction in solving problems of multiclass classification by reducing it to solving binary problems

consultant: Yu.V. Maksimov

Task: Explore different approaches to solving multi-class classification problems and compare their performance.

Data: Data with a different number of classes. 0. Toy example: Shuttle dataset. http://archive.ics.uci.edu/ml/datasets/Statlog+(Shuttle). Small sample, 7 classes. No need to do data preparation. 1. Reuters collection text data http://www.daviddlewis.com/resources/testcollections/reuters21578/. 2. Data from our LIG Kaggle contest http://www.kaggle.com/c/lshtc

References:

Basic algorithms: SVM with different cores, Adaboost. Basic approaches: one vs all(combined), one vs one(uncombined)

Homework-2: Trial Programming

Task Who is doing Number
A selection is given "Wine of different regions". It is required to determine the clusters (regions of origin of wines) and draw the result: the cluster object is marked with a colored dot; the colored circle indicates the class of this object taken from the sample. Task option: determine the number of clusters. Task option: use two algorithms, for example k-means and EM, and show a comparison of clustering results on a graph. Plavin 1
Suggest ways to visualize sets of 4D vectors, see example for Fisher's iris data. Write down your last name here. 2
Given a time series series describing electricity consumption. Approximate a series by several curvilinear models and plot the predicted and original series on the same graph. Kulunchakov Andrey. 3
Smooth the time series Prices (volumes) for the main exchange instruments using the exponential smoothing. Draw color plots of the antialiased rows with different  \alpha and the original row. Avdyukhov 4
Closed Curve Sample Fit [129]: Check if points lie on a circle? Generate data yourself. Gazizullina Rimma 5
A time series with gaps is given, using the example [130]. Suggest ways to fill in the gaps in the data, fill in the gaps. For each method, construct a histogram. Option: take a sample without gaps, randomly remove part of the data, fill in the gaps, compare with the histogram of the original sample. Ignatov Andrey 6
A selection is given "Wine of different regions". Choose two features. Consider different distance functions when classifying with nearest neighbor method. For each, depict the classification result in the space of selected features. Maria Popova 7
For various types of dependence  y = f(x) + \epsilon (linear, quadratic, logarithmic) build linear regression and plot the SSE deviations (standard deviations-?). Generate data yourself or take data "Price for bread". Efimova Irina 8
Estimate the area of a unit circle using the Monte Carlo method. Plot the result against the sample size. Shinkevich Mikhail 9
Construct a convex hull of points on a plane. Draw a graph: points and their convex hull is a closed broken line. Makarova Anastasia 10
A selection is given: Iris. Implement the decision tree classification procedure. Illustrate the results of classification on a plane in the space of two features. Zhukov Andrey 11
The time series is set - volumes of hourly electricity consumption (select any two days). Approximate the series with polynomial models of various degrees (1-7). *Suggest a method for determining the optimal degree of a polynomial. Karasikov Mikhail 12
Two one-dimensional [[Time series (library examples)] | time series]] of various lengths. Calculate row spacing using dynamic alignment. Grinchuk Alexey 13
Generate a set of points on the plane. Select and visualize the main components. Lipatova 14
Approximate the sample bread prices with a polynomial model. Draw a graph. Mark objects that are outliers using the three sigma rule. Shvets Mikhail 15
Divide the sample Iris into clusters. Illustrate the results of clustering on a graph, highlight the clusters in different colors. Gushchin Alexander 16
And more tasks to choose from
A sample of several features is given, without a target vector Y. For example, this https://dmba.svn.sourceforge.net/svnroot/dmba/Data/Diabets_LARS.csv You need to specify the feature that is well described (in terms of linear regression) by the rest (such a feature is usually excluded from the sample). 17
Smooth time series (see library) with moving average. Take several windows of different lengths and superimpose the result on the graph on top of each other. Kostyuk 18
Given a time series (see library). Based on its variational series, construct a histogram of n percentiles and draw it. What is the most common time series value? Gizzatullin Anvar 19
Show the difference in the speed of performing matrix operations and operations in a loop. You can use Singular value decomposition and other linear algebra methods as an example. Show the efficiency of parallel computing (parfor). 20
Understand how function superposition works. Using the @ function, generate all possible polynomials in n variables of degree at most p. Option: use the obtained polynomials to approximate the time series of bread prices (data).

2013

My first cross-reviewed scientific paper

Tasks

Task name Author Link MAIPVTDCHSJ
Definition of the printed image Pushnyakov Alexey [131] MAIPVTDCHSJ
Comparison of Fast Clustering Algorithms Alexander Katrutsa [132] MAIPVTDCHS
Vector autoregression and management of macroeconomic indicators Kashcheeva Maria [133] MAIPVTDCHS
Marking up bibliographic records using logical algorithms Ryskina Maria [134] MAIPVTDCHS
Determination of the exact border of the pupil Chinaev Nikolai [135] MAIPV.DCHS
Vector autoregression and management of macroeconomic indicators Grinchuk Oleg [136] MAIPVTD.HS
Generating Neural Networks with Expert-Defined Activation Functions Perekrestenko Dmitry [137] MAIPVTDСHS
Comparative analysis of feature selection algorithms: accuracy, stability, complexity of regression models Yashkov Daniel [138] MAI.VTD.HS
Invariant transformations in tasks of local forecasting Kostin Alexander [139] MAI.VT.HS
Genetic Programming Algorithm for Solving the Prediction Problem Voronov Sergey [140] MAIPVTDC.S
Grouping of Nominal Variables in Bank Credit Scoring tasks Mityashov Andrey [141] MAIPVTDCHS
Modeling the process of learning and forgetting when assessing the quality of production Neklyudov Kirill [142] MAI..DC.S
Overview of Algorithms for Simplifying Algebraic Expressions Shubin Andrey [143] MAIPVTD.S
Search algorithms for the most informative objects and features in logistic regression Ibraimova Aizhan [144] MAIP.TD..
Interpretation of expert assessments of species of the Red Book of the Russian Federation by selecting reference (representative) objects Byrdin Alexander [145] MAI.TD.S
Visualization of Pair Distance Matrix in Topic Modeling Vdovina Evgenia [146] MAI.TDC.S
Algorithm for Estimating the Reliability of Expert Judgments on the Relationship of Time Series Antipova Natasha [147] MAIP.T..S

Task 2. Surname2013MassProduction (*eng)

  • Name Generation and optimization of logical descriptions when building production lines.
  • Task It is required to set the task of synthesizing admissible superpositions, develop an algorithm and test it on synthetic data.
  • Data Required to create.
  • References: Need a search (most likely German publications).
  • Proposed algorithm On discussion.
  • Basic algorithm None.

Task 3. Surname2013LearnForget (eng)

  • Name Modeling the process of learning and forgetting when assessing the quality of production.
  • Task Find an adequate regression model that describes the activities of a group of people.
  • Data Data on the speed and quality of the assembly of paper airplanes.
  • References:. Need to find.
  • Proposed algorithm The procedure for analyzing regression residuals.
  • Basic algorithm Regression model in the attached article.

Task 4. Surname2013GeneticProg

  • Name Genetic Programming Algorithm for Solving the Prediction Problem.
  • Task Create a genetic programming algorithm that solves the problems named by Ivan Zelinka. Suggest a way to test the resulting models, organize a sliding control. Compare its performance on a test set of tasks with the performance of other GPU algorithms and with neural networks.
  • Data Test set of tasks, take on the UCI or on the Polygon.
  • References:. Zelinka, Oplatkova, Vladislavleva; find works of recent years on this topic. Especially for testing these algorithms.
  • Proposed algorithm GPU.
  • Basic algorithm GPU, neural networks.

Task 5. Surname2013Simplify

  • Name Overview of Algorithms for Simplifying Algebraic Expressions.
  • Task It is required to find literature on algorithms that simplify expressions, compare algorithms, program the algorithm proposed in the work by Ruda/Strizhov V.V.
  • Data Collect a test collection of expressions.
  • References:. Graph rewriting.
  • Proposed algorithm R/S, comparison of algorithms.

Task 6. Surname2013RedListExplanation

  • Name Interpretation of expert assessments of species of the Red Book of the Russian Federation by selecting reference (representative) objects.
  • Task Selection of reference objects (STOLP algorithm). This algorithm can be interesting for Experts: it quickly finds noise objects, which in our terms are considered to be inconsistent with Expert data and "out of their class", and also selects reference objects that are also interpreted in a curious way. From a mathematical point of view, it is interesting, firstly, to observe different metrics (generalizations of the Hamming distance) and, most importantly, it is necessary to generalize the margin formula for the case of monotone classes, apparently by introducing the weight function of objects.
  • Data expert assessments of Red Data Book species.
  • References:. References: according to metric classification algorithms.
  • Proposed algorithm A method or algorithm that tells the Expert why (sic!) an object is not in the Expert's intended class.

Task 7. Surname2013RedListClassification

  • Name Algorithm for monotonic classification of objects described in rank scales.
  • Task Apply a decision tree to the Expert Estimates of Threatened Species in the Red Data Book. Compare with previously proposed algorithms. To substantiate operations with rank features, to introduce a generalization of the concept of informativeness for the case of monotone classes, apparently, to generalize the hypergeometric distribution.
  • Data expert assessments of Red Data Book species.
  • References:. You should try to avoid referring to trivial sources. Search for similar works in foreign magazines.

Task 11. Surname2013Invaraint4LocalForecast

  • Name Invariant transformations in tasks of local forecasting.
  • Task Combine algorithms for invariant transformation of time and amplitude of predicted time series.
  • Data Time series of pulse wave measurement.
  • References:. Find, avoid trivial references.

Task 8. Surname2013PlausibleExpert

  • Name Algorithm for Estimating the Reliability of Expert Judgments on the Relationship of Time Series.
  • Task Study of the relationship between exchange prices for the main instruments and rail freight.
  • Data Time series for 1.5 years. But it is better to choose a synthetic example.
  • References:. Publications on CCM.
  • Proposed algorithm CCM modifications.

Task 9. Surname2013DeepLearning

  • Name Generating Neural Networks with Expert-Defined Activation Functions.
  • Task It is required to raise the current state of the DeepLearning area, program the algorithm, test it on the problem of predicting consumption volumes and electricity prices.
  • Data Daily data for three years.
  • References:. Deep Learning.
  • Proposed algorithm Building a neural network and estimating its parameters.

Task 16. Surname2013ScoringSelection

  • Name Search algorithms for the most informative objects and features in logistic regression.
  • Task Using a genetic algorithm to find informative objects and features.
  • Data Consumer credit data.
  • References:. -

Task 10. Surname2013ScoringFeatureSelection

  • Name Grouping of Nominal Variables in Bank Credit Scoring tasks.
  • Task Create a genetic algorithm for reducing the dimension of a feature space.
  • Data Historical data on cash loans.
  • References:. SAS, find more.

Task 15. Surname2013InverseVAR

  • Name Vector autoregression and management of macroeconomic indicators.
  • Task Solve the inverse forecasting problem. According to the given state of the economy, set such a value of managed macroeconomic indicators that would bring the economy to the desired state.
  • Data Macroeconomic indicators of Russia over the past 16 years.
  • References:. S.A. Ayvazyan works.

Task 12. Surname2013DistanceVisualizing

  • Name Visualization of Pair Distance Matrix in Topic Modeling.
  • Task Display abstracts of the conference on the plane with the preservation of clusters.
  • Data EURO conference abstracts.
  • References: Zinoviev on ML, references on the topic.
  • Proposed algorithm PCA.
  • Basic algorithm Algorithm with minimization of the energy criterion.

Task 13. Surname2013RhoNets

  • Name Comparison of Fast Clustering Algorithms.
  • Task Compare clustering algorithm using $\rho$-networks and a fast $k$-means algorithm.
  • Data A selection of amino acid sequences. We need a test sample from the UCI or from comparison papers.
  • References: $k$-средних, $\varepsilon$-networks.
  • Proposed algorithm $\rho$-networks.
  • Basic algorithm $k$-means.

Task 17. Surname2013FeatureSelection

  • Name Comparative analysis of feature selection algorithms: accuracy, stability, complexity of regression models.
  • Task Build a series of test problems to compare algorithms. Propose a feature selection algorithm with the analysis of covariance matrices based on the Belsley method.
  • Data Synthetic.
  • References:. Leontieva/Strizhov V.V., search for modern reviews.

Task 1. Surname2013Txt2Bib

  • Name Marking up bibliographic records using logical algorithms.
  • Task It is required to create a text markup algorithm. Novelty in the formulation of the problem. The relevance is that a more complete library of logical expressions will be created and an adequate algorithm will be selected.
  • Data MLAlgorithms.
  • References:. The work of A. Ivanova and everything that is on the topic over the past two years.
  • Proposed algorithm Choose from logical classification algorithms; optional clustering.
  • Basic algorithm Dead-end coatings.

Task 14. Surname2013FindTheFormula (Risky)

  • Name Algorithm for searching text structures in a document.
  • Task Suggest an algorithm that would look for formulas in a TeX document that are equivalent to a given one.
  • Data Synthetic, MLAlgorithms collection.
  • References Have to search. Search by chemical compounds in WoK works well.

Task 18. Surname2013ScannedImage (Image)

  • Name Form type definition.
  • Task Determine the type of form from the scan.
  • Data A set of images in TIF.

Task 19. Surname2013SpectrumImage (Image)

  • Name Definition of the printed image.
  • Task Make a spectral transformation of the image, explore the spectrum.
  • Data A set of JPG images classified into two classes.
Task Who is doing
A set of three-element vectors is given. Draw the first two elements along the abscissa and ordinate axes. The third element is displayed as a circle with a proportional radius. Choose proportions based on a sense of beauty. Compare the resulting graph with plot3. What's better? Mityashov Andrey
Given a five-element vector. Draw %D0%B2%D0%B0 Chernov's face]. Which is better - Chernov's face or D0%BE%D0%B2%D0%B0%D1%8F+%D0%B4%D0%B8%D0%B0%D0%B3%D1%80%D0%B0%D0%BC%D0%BC%D0% B0%3F&aq=f&oq=%D0%9B%D0%B5%D0%BF%D0%B5%D1%81%D1%82%D0%BA%D0%BE%D0%B2%D0%B0%D1%8F+ %D0%B4%D0%B8%D0%B0%D0%B3%D1%80%D0%B0%D0%BC%D0%BC%D0%B0%3F&aqs=chrome.0.57j0l3.7857&sourceid=chrome&ie=UTF- 8 diagram? Neklyudov Kirill
Understand how regexp works in Matlab. Make code that highlights everything that is inside the brackets of some arithmetic expression. Ryskina Maria
Understand how function superposition works. Using the @ function, generate all possible polynomials in n variables of degree at most p. Shubin Andrey
Understand how a web connection and regexp works. Make a search query on a topic and make up a BibTeX entry from it.
Given a time series of m + 1 (random) points. Approximate its first m points by polynomials of degree from 1 to m. Calculate the mean error in points. Which degree gives the largest error? Voronov Sergey
Rotate and zoom in on a flat figure, make a zoom effect with frame-by-frame rotation. Antipova Natasha
Two matrices are given. Check if they have an intersection - a submatrix? Vdovina Evgenia
A sample of several features is given, without a target vector Y. For example, this https://dmba.svn.sourceforge.net/svnroot/dmba/Data/Diabets_LARS.csv You need to specify the feature that is well described (in terms of linear regression) the rest (such a feature is usually excluded from the sample). Grinchuk Oleg
Given a sample that has several outliers. It is known that it can be described by one-dimensional linear regression. It is required to find the outliers by enumeration. Show them on a chart. Pushnyakov Alexey
Given a sample of two classes on a plane. It is required to find all the objects that got into a foreign class. Show them on a chart. Kashcheeva Maria
The input is the incidence matrix of the tree. The function returns a list (vector) of vertices in the order they were visited. Ibraimova Aizhan
Classify iris flowers with an arbitrary algorithm, draw the “most visual” pair of features on the plane, indicate what was classified correctly and what was not. Yashkov Daniel
Given a time series. Based on its variational series, build a histogram of n percentiles, draw it. What is the most common time series value?
Create several groups of points on the plane and perform their clustering using any algorithm of your choice. Visualize the resulting clusters. Calculate the average intracluster distance for one cluster. Perekrestenko Dmitry
Upload a sound sequence, preferably a few piano notes. Select and play a specific note.
Download video. Delete every second frame. Process to taste. Write back. Byrdin Alexander
Show the difference in the speed of performing matrix operations and operations in a loop. Show the efficiency of parallel computing (parfor and others). Alexander Katrutsa
Suggest options for visualization of four-dimensional vectors and spaces. Compare them to a built-in function.
Smooth the time series with a moving average. Take several windows of different lengths and superimpose the result on the graph on top of each other. Chinaev Nikolai
Draw a surface. Replace each point of the surface with a median of n neighbors. Draw the result. Kostin Alexander

2012

Thematic Modeling: paper in the Higher Attestation Commission journal

For the status of the publication of works, see the bottom of the page, section "Publishing of works". All papers are expected to be published by the end of May 2013.


Task list

Task name Author Link to work Comments
Calculation of integral indicators in rank scales by co-clustering methods Medvednikova Maria [148] Published
Hierarchical thematic abstract clustering and visualization Arsenty Kuzmin [149] Published
Joint selection of objects and features in Tasks of multiclass classification. Alexander Aduenko [150] Published
Building hierarchical topic models Tsyganova Svetlana [151] Published
Feature Selection in tasks Structural Regression Varfolomeeva Anna [152] Accepted
Statistical tests for homogeneity and goodness of fit for highly sparse discrete distributions Vlada Tselykh

[153]

Published
Building logical rules when marking up texts Ivanova Alina [154] Accepted
Checking the adequacy of the topic model Stepan Lobastov [155] Redaction

Vorontsov Topics

  • 2012SparceDistribution Statistical tests for homogeneity and goodness of fit for highly sparse discrete distributions

2012LatentModels

  • Name: Checking the adequacy of the topic model.
  • Teaser: Methods for detecting incorrect thematic classification on conference materials. Methods for constructing a thematic model similar to the given one. Article clustering, hierarchical topic models with topic interpretability. Hierarchical thematic clustering of abstracts.
  • Data: Texts of Euro 2012 conference abstracts, 1862 abstracts.
  • References: for latent models.
  • Keywords: soft clustering, latent models.
  • Proposed algorithm:: hHDP.
  • Basic algorithm:HDP.
  • Problem: It is required to build a thematic model using the clustering method and check the correctness of the current text classification. To do this, (hierarchical) clustering of texts is performed, each cluster is assigned a topic name corresponding to the majority of articles from the cluster. After building the model, each article is checked and refers to its own or someone else's topic.

Links

https://mlalgorithms.svn.sourceforge.net/svnroot/mlalgorithms/utilities SugarSync/remarks contains a document with one of the possible text spacing functions.

References:

https://www.sugarsync.com File storage where project materials are located. Access to the corresponding folder is provided by e-mail address. The materials include publications on each topic.

Publication of works

Legend: Redaction >> Submit (design for the journal) >> Submitted >> Accepted (by reviewers) >> Layout (reviewers' and editor's comments taken into account) >> Published (the issue has been published).

Task name Author Link to the journal The original text of the work Date of application State
Feature selection and metric optimization when clustering a collection of documents Aduenko A.A., Kuzmin A.A., Strizhov V.V. Izvestiya TulGu [156] 12.10.2012 Published
Estimating the Probabilities of Strings in a Collection of Documents Budnikov E.A., Strizhov V.V. Information Technology [157] 24.09.2012 Published
Checking the adequacy of the topic models of a collection of documents Kuzmin A.A., Strizhov V.V. Software engineering [158] 17.12.2012 Published
Algorithm for the optimal location of the names of a collection of documents Aduenko A.A., Strizhov V.V. Software engineering [159] 13.11.2012 Published
Visualization of the matrix of paired distances between documents Aduenko A.A., Strizhov V.V. Scientific and technical statements of S.-Pb.PSU [160] 29.10.2012 Submitted
Construction of an integral indicator of the quality of scientific publications by co-clustering methods Medvednikova M.M., Strizhov V.V. Izvestiya TulGu [161] 15.11.2012 Published
Joint selection of objects and features in Tasks of multiclass classification of a collection of documents Aduenko A.A., Strizhov V.V. Infocommunication technologies [162] 18.12.2012 Published
Algorithm for constructing logical rules when marking up texts Иванова А.В., Aduenko A.A., Strizhov V.V. Software engineering [163] 24.01.2013 Accepted
Building hierarchical topic models of document collections Tsyganova S.V., Strizhov V.V. Прикладная информатика [164] 27.01.2013 Published
Choice of features when marking bibliographic lists by methods of structured learning Varfolomeeva A.A., Strizhov V.V. Scientific and technical statements of S.-Pb.PSU [165] 27.01.2013 Reviewed
Goodness-of-fit criteria for sparse discrete distributions and their application in topic modeling Tselykh V.R., Vorontsov K. V. Machine learning and data analysis [166] 17.12.2012 Published
Checking the adequacy of the topic model Stepan Lobastov [167] Redaction

List of works accepted for publication

  • 1. Aduenko A. A., Strizhov V. V. V.V. Visualization of the matrix of paired distances between documents // Scientific and technical bulletin of St. Petersburg. PGU. Computer science. Telecommunications. Management, 2013, 1 - ?.
  • 2. Aduenko A. A., Kuzmin A. A., Strizhov V. V. V. V. Feature selection and metric optimization when clustering a collection of documents // Proceedings of the Tula State University, Natural Sciences, 2012, No. 3. P. 119-132.
  • 3. Aduenko A. A., Strizhov V. V. V.V. Algorithm for the optimal location of the names of a collection of documents // Software engineering, 2013. No. 3. P.21-25.
  • 4. Budnikov E. A., Strizhov V. V. V. V. Estimating the Probabilities of Strings in a Collection of Documents // Information Technology, 2013. No. 4.
  • 5. Kuzmin A. A., Strizhov V. V. Checking the adequacy of the topic models of a collection of documents // Software engineering, 2013. No. 4.
  • 6. Medvednikova M. M., Strizhov V.V. Construction of an integral indicator of the quality of scientific publications by co-clustering methods // Proceedings of the Tula State University, Natural Sciences, 2013. No. 1.
  • 7. Aduenko A. A., Strizhov V. V. V. V. Joint selection of objects and features in Tasks of multiclass classification of a collection of documents // Infocommunication technologies, 2013. No. 2.
  • 8. Ivanova A.V., Aduenko A.A., Strizhov V.V. V.V. Algorithm for constructing logical rules when marking up texts // Software engineering, 2013. No. 4(5).
  • 9. Tsyganova S.V., Strizhov V.V. V. V. Building hierarchical topic models of document collections // Applied Informatics, 2013. No. 1.
  • 10. Varfolomeeva A.A., Strizhov V.V. V. V. Choice of features when marking bibliographic lists by methods of structured learning // Scientific and Technical Bulletin of St. Petersburg. PGU. Computer science. Telecommunications. Management, 2013.
  • 11. Tselykh V.R., Vorontsov K. V. Goodness-of-fit criteria for sparse discrete distributions and their application in topic modeling // JMLDA, 2012. No. 4. pp. 432-442.

My first cross-reviewed scientific paper

Task list

Task name Author Reviewer Link to work Comments
CMARS: spline approximation Vlada Tselykh Tatiana Shpakova Celyh2012CMARS [.]сaipvdstrj(10)
Algorithmic foundations for constructing bank scoring cards Alexander Aduenko Alina Ivanova Aduenko2012economics [.]сaipvdstrj(10)
Using the method of principal components in the construction of integral indicators Maria Medvednikova Svetlana Tsyganova Medvednikova2012PCA [r]сaipvdstrj(10)
Multi-level classification for price movement detection Arsenty Kuzmin Varfolomeeva A.A. Kuzmin2012TimeRows [r]сaipvdstjr(10)
Local forecasting methods with the choice of an invariant transformation Svetlana Tsyganova Maria Medvednikova Tsyganova2012 LocalForecast [r]сaipvdstjr(10)
Prediction of Quasi-Periodic Multivariate Time Series by Non-Parametric Methods (example) Egor Klochkov Alexander Shulga Klochkov2012Goods4Cast [r]сaipvdstj.(10)
Search algorithms for the most informative objects and features in logistic regression (example) Stepan Lobastov Egor Klochkov Lobastov2012FOSelection [r]сaipvdstrj(10)
Local forecasting methods with the choice of metric Varfolomeeva A.A. Arsenty Kuzmin Varfolomeeva2012 LocForecastMetrics [r]сaipvdstjr(10)
Chebyshev polynomials and time series forecasting Valeria Bochkareva Stepan Lobastov Bochkareva2012TimeSeriesPrediction [.]сaipvdst-r(9)
Clustering and compiling a dictionary of amino acid sequences Tatiana Shpakova Vlada Tselykh Shpakova2012Clustering [.]сaipvdst.(9)
Vector autoregression and management of macroeconomic indicators Alexander Shulga Shulga2012VAR [.]сaipvds..(9)
Approximation of empirical distribution functions Alina Ivanova Alexander Aduenko Ivanova2012 ApproximateFunc [r]сaipvd..(9)

Annotations

Search algorithms for the most informative objects and features in logistic regression

Logistic regression is a statistical model that is used to predict the probability of an event occurring based on the values of a set of features. It has applications, for example, in medicine [168] and credit scrolling. In real conditions, the number of features is usually large, and the most important task is to select only essential features, as well as to search for objects that are atypical for one reason or another. Keywords: logit model, feature selection, boosting.

Using the method of principal components in the construction of integral indicators

This paper considers Using the method of principal components in the construction of integral indicators. The results obtained are compared with the results given by the Pareto stratification method. An integral indicator is being built for Russian universities. For this, biographies of the 30 richest businessmen in Russia according to the Forbes magazine for 2011 are used. Keywords: integral indicator, expert estimates, parameter weights, principal component method, Pareto stratification method.

Approximation of empirical distribution functions

The work is devoted to methods of approximation of functions for efficient calculation of integrals. Practical Tasks usually have data at certain points in time or space. When making assumptions about the remaining points, it becomes necessary to approximate the distribution function of the quantity under study, as well as to estimate the corresponding error. For its calculation, it is possible to use methods of different accuracy.

Keywords: Monte Carlo method, calculation of distribution functions, empirical distribution functions.

Local prediction methods with choice of transformation

Time series forecasting problems have many applications in various fields such as economics, physics, and medicine. Their solution is a forecast for the near future based on the already known values of the predicted series at previous points in time. In the work, a local forecasting algorithm will be built taking into account transformations, which allows, without human intervention, to identify visually similar sections of the time series.

Keywords: local prediction, transformation

Draft task list

  1. Clustering and compiling a dictionary of amino acid sequences
  2. Oblivious decision trees: алгоритм Яндекс для системы Полигон
  3. Comparative analysis of regression residuals in SVN regression
  4. Algorithms for finding Gaussian mixtures
  5. Prediction of quasi-periodic multivariate time series by non-parametric methods
  6. Multi-level classification for price movement detection
  7. CMARS: spline approximation
  8. Chebyshev polynomials and sweep method in time series forecasting
  9. Comparison of ARMA and FLS methods for hindsight
  10. Local forecasting methods with the choice of metric
  11. Local forecasting methods with the choice of an invariant transformation
  12. Search algorithms for the most informative objects and features in logistic regression
  13. Vector autoregression and management of macroeconomic indicators
  14. Building a ranking of Russian universities based on open data on the career success of their graduates

More tasks

  1. Text analysis by structural learning methods
  2. Approximation of empirical distribution functions
  3. Algorithmic foundations for constructing bank scoring cards
  4. Singular value decomposition and search engine
  5. Comparison of multiobjective optimization algorithms
  6. Refinement of Expert estimates on data in rank scales (interval, cones, weights of Experts, copulas)
  7. Refinement of Expert assessments in the analysis of the mechanism of sustainable development of the energy sector
  8. Visualization of the parameter space of regression models
  9. Regression recovery by principal component method
  10. Estimation of hyperparameters by sampling
  11. Decimation of essentially non-linear models using hyperparameters
  12. Occam's factor for parametric models with a known parameter domain
  13. Creation of algorithms for sequential modification of models
  14. Generation and selection of classification models

More tasks

  • Distance function between formulas and search.
  • Search for objects (technical work).

+

  • Autoregression
  • Vector autoregression
  • Exponential smoothing
  • Local methods, search metrics
  • Local methods with invariants, fixed metric
  • ARIMA
  • Multidimensional track, track length selection
  • Multidimensional track, row selection
  • Prediction using DTW
  • Moving average, selection of cores
  • Moving average with history forgetting
  • Moving average of time series with a periodic component
  • Prediction by neural networks
  • Forecast quality analysis
  • Time series meta description
  • Logic prediction
  • SVN - Regression
  • Discrete prediction, music.

Compose

  • List of common typographical errors
  • List of BibTeX errors

2011

Publication in the JMLDA journal

Recommended reading before completing assignments

Tasks

Name Author Reviewer Link Comments
Estimation of hyperparameters of linear regression models in the selection of noise and correlated features Tokmakova Alexandra A. P. Motrenko Tokmakova2011HyperPar
Choice of forecasting models for electricity prices Leontieva Lyubov Grebennikov Evgeny Leonteva2011ElectricityConsumption
Multiclass prediction of the probability of myocardial infarction and estimation of the required sample size of patients (example) A. P. Motrenko Tokmakova Alexandra Motrenko2011HAPrediction
Algorithms for generating essentially non-linear models Georgy Rudoy Nikolai Baldin Rudoy2012Generation
Event Modeling and Financial Time Series Forecast Alexander Romanenko Budnikov E. A. Romanenko2011Event
Overview of some statistical models of natural language Budnikov E. A. Alexander Romanenko Budnikov2011Statistical

My first publication in JMLDA magazine

Recommended reading before completing assignments

See also

Tasks

Name Author Работу рецензирует Link Comments
Using the Granger Test in Time Series Forecasting Anastasia Motrenko Leontieva Lyubov Motrenko2011GrangerForc Published at JMLDA
Choosing an Activation Function for Predicting Neural Networks Georgy Rudoy Nikolai Baldin Rudoy2011NNForecasting Published at JMLDA
Multidimensional caterpillar, choice of length and number of caterpillar components (example) Leontieva Lyubov Mikhail Burmistrov Leonteva2011GaterpillarLearning Published at JMLDA
Prediction by Discrete Argument Functions (example) Budnikov E. A. Alexander Romanenko Budnikov2011DiscreteForecasting Published at JMLDA
Investigation of Convergence in Prediction by Neural Networks with Feedback Nikolai Baldin Georgy Rudoy Baldin2011FNNForecasting Published at JMLDA
Time series alignment: Forecasting with DTW Alexander Romanenko Budnikov E. A. Romanenko2011DTWForecasting Published at JMLDA
Isolation of the periodic component of the time series (example) Tokmakova Alexandra Budnikov E. A. Tokmakova2011Periodic Published at JMLDA

Brief description of tasks

Task 1: Non-parametric forecasting: kernel selection, parameter tuning

The paper describes the method of nuclear smoothing of the time series, as one of the types of nonparametric regression. The essence of the method consists in restoring the function of time as a weighted linear combination of points from some neighborhood. A continuous bounded symmetric real weight function is called a kernel. The resulting kernel estimate is used to predict the next point in the series. The dependence of the quality of prediction on the parameters of the kernel and the superimposed noise is investigated.

Task 2: Exponential Smoothing and Prediction

The paper investigates the application of the exponential smoothing algorithm to time series forecasting. The algorithm is based on taking into account the previous values of the series with weights decreasing as you move away from the studied section of the time series. The behavior of the algorithm on model data in various models of weights is studied. An analysis of the operation of the algorithm on real data - stock indices was carried out.

Task 3: Isolation of the periodic component of the time series (example)

The project examines the time series for the presence of a periodic component, builds a trigonometric interpolation of the proposed time series using the least squares method. The parameters of the function of the least squares method are estimated depending on the quality of forecasting. In a computational experiment, the results of the work of the correlation function and the least squares method on a noisy model sine and a real time series of an electrocardiogram are presented.

Task 4: Multivariate caterpillar, choice of length and number of caterpillar components (comparison of smoothed and unsmoothed time series) (example)

The paper describes the caterpillar method and its application for time series forecasting. The algorithm is based on the selection of its informative components from the studied time series and the subsequent construction of a forecast. The dependence of the accuracy of forecasts on the choice of the caterpillar length and the number of its components is investigated. In a computational experiment, the results of the algorithm's operation on periodic series with different patterns within a period, on series with violation of periodicity, as well as on real time series of hourly temperature, are presented.

Task 5: Prediction by Discrete Argument Functions (example)

The paper investigates short time series on the example of monophonic musical melodies. There is a prediction of one note by exponential smoothing, a local method, as well as a method of searching for constant patterns. The computational experiment is carried out on two melodies, one of which has exactly repeating fragments.

Task 7: Local forecasting methods, search for metrics

The time series is divided into separate sections, each of which is associated with a point in the n-dimensional feature space. The local model is calculated in three successive stages. The first one finds the k-nearest neighbors of the observed point. The second one builds a simple model using only these k neighbors. The third - using this model, predicts the next one based on the observed point. Many researchers use the Euclidean metric to measure distances between points. This work is intended to compare the accuracy of forecasting when using different metrics. In particular, it is required to investigate the optimal set of weights in the weighted metric to maximize the prediction accuracy.

Task 8: Local prediction methods, search for invariant transformation

The project uses local forecasting methods time series. There is no temporary representation in these methods series in the class of given functions of time. Instead, the prediction is made on the basis of data about some part of the time series (local information is used). In this paper, we study in detail the following method (a generalization of the classical "nearest neighbour").

Let there be a time series and Task should continue it. It is assumed that such a continuation is determined prehistory, i.e. in a series you need to find the part that after some transformation of A becomes similar to the part we are trying to predict. Finding such a transformation A and is the goal of this project. To determine the degree of similarity, the function B is used - the function of the proximity of two segments time series (for more details see 8C%D0%BD%D1%8B%D0%B5_%D0%BC%D0%B5%D1%82%D0%BE%D0%B4%D1%8B_%D0%BF%D1%80%D0%BE% D0%B3%D0%BD%D0%BE%D0%B7%D0%B8%D1%80%D0%BE%D0%B2%D0%B0%D0%BD%D0%B8%D1%8F%2C% D0%BF%D0%BE%D0%B8%D1%81%D0%BA_%D0%BC%D0%B5%D1%82%D1%80%D0%B8%D0%BA%D0%B8_%28% D0%BF%D1%80%D0%B8%D0%BC%D0%B5%D1%80%29&action=edit here). This is how we find the closest neighbor to our backstory. In general, we are looking for several nearest neighbors. The continuation will be written as their linear combination.

Task 9: Time Series Flattening: Forecasting with DTW (example)

Time series is a sequence of time-ordered values of some real variable $\mathbf{x}=\{x_{t}\}_{t=1}^T\in\mathbb{R }^T$. The task that accompanies the appearance of time series is the comparison of one data sequence with another. Comparison of sequences is greatly simplified after the deformation of the time series along one of the axes and its alignment. Dynamic time warping (DTW) is a technique for effectively leveling time series. DTW methods are used in speech recognition, information analysis in robotics, industry, medicine and other areas.

The purpose of the work is to give an example of alignment, to introduce a comparison functional for two time series, which has the natural properties of commutativity, reflexivity and transitivity. The functional should take two time series as input, and at the output give a number characterizing the degree of their "similarity".

Task 10: Choosing an Activation Function for Predicting Neural Networks

The aim of the project is to study the dependence of the quality of prediction by neural networks without feedback (single- and multilayer perceptrons) on the chosen activation function of neurons in the network, as well as on the parameters of this function.

The result of the project is to evaluate the quality of forecasting by neural networks depending on the type and parameters of the activation function.

Task 12: Investigation of Convergence in Prediction by Neural Networks with Feedback

The dependence of the convergence rate in time series forecasting on the parameters of a neural network with feedback is investigated. The concept of feedback is typical for dynamic systems in which the output signal of some element of the system affects the input signal of this element. The output signal can be represented as an infinite weighted the sum of the current and previous input signals. The Jordan network is used as a neural network model. It is proposed to investigate the rate of convergence depending on the choice of the activation function (sigmoid, hyperbolic tangent), on the number of neurons in the intermediate layer and on the width of the sliding window. We also explore a way to increase the rate of convergence using the generalized delta rule.

Task 13: Multidimensional caterpillar, choice of length and number of caterpillar components (example)

The work is devoted to the study of one of the methods for analyzing multivariate time series - the "caterpillar" method, also known as Singular Spectrum Analysis or SSA. The method can be divided into four stages - the representation of the time series in the form of a matrix using a shift procedure, the calculation of the covariance matrix of the sample and its singular value decomposition, the selection of principal components related to various components of the series (from slowly changing and periodic to noise), and, finally, line restoration.

The scope of the algorithm is Tasks of both meteorology and geophysics, and economics and medicine. The purpose of this work is to find out the dependence of the efficiency of the algorithm on the choice of time series used in its work.

Task 14: Using the Granger Test in Time Series Forecasting

When predicting a series, it can be useful to determine whether a given series is "dependent" on some other series. The Granger test, based on statistical tests, helps to identify such a relationship (in this case, the method does not guarantee an accurate result - when comparing two rows that depend on another row, an error is possible). The method is used in forecasting economic and natural phenomena (for example, earthquakes).

The purpose of the work is to propose an algorithm that makes the best use of this method; investigate the effectiveness of the method depending on the predicted series.

Task 15: Prediction and spline approximation

Description.

Task 16: ARIMA and GARCH for predicting high volatility series

Description.

Task 17: Prediction and SVN Regression

Description.

Reports and exam (clarifications are possible)

  • Report-1 April 6
  • Checkpoint 12 May
  • Exam May 19
Личные инструменты