# Участник:Strijov/Drafts

### Материал из MachineLearning.

< Участник:Strijov(Различия между версиями)

(→BCI, Matrix and tensor approximation) |
Текущая версия (22:28, 26 февраля 2023) (править) (отменить) |
||

(278 промежуточных версий не показаны.) | |||

Строка 1: | Строка 1: | ||

- | |||

{{TOCright}} | {{TOCright}} | ||

+ | ==2023== | ||

+ | ===Problem 112=== | ||

+ | * '''Title:''' Modeling an FMRI reading from a video of a shown person | ||

+ | * '''Problem description:''' It is required to build a dependence model of the readings of FMRI sensors and the video sequence that a person is viewing at this moment. | ||

+ | * '''Data:''' The sample for approximation is presented in the work of J. Berezutskay, in which there are various types of parallel signals. | ||

+ | * '''Literature:''' Berezutskaya J., et al Open multimodal iEEG-fMRI dataset from naturalistic stimulation with a short audiovisual film // Sci Data 9, 91, 2022. | ||

+ | * '''Predecessor code:''' | ||

+ | * '''Base algorithm:''' Running code based on transformer models. | ||

+ | * '''Novelty:''' Analysis of the relationship between sensor readings and human perceptions of the external world. It is required to test the hypothesis of the relationship between the data, as well as to propose a method for approximating FMRI readings based on the video sequence being viewed. | ||

+ | * '''Authors:''' Expert Grabovoi Andrey. | ||

- | = | + | ===Problem 113=== |

- | + | * '''Title:''' Modeling of the FMRI indication on the sound range that a person hears | |

- | + | * '''Problem description:''' It is required to build a model of the dependence of the readings of the FMRI sensors and the sound accompaniment that a person is listening to at this moment. | |

+ | * '''Data:''' The sample for approximation is presented in the work of J. Berezutskay, in which there are various types of parallel signals. | ||

+ | * '''Literature:''' Berezutskaya J., et al Open multimodal iEEG-fMRI dataset from naturalistic stimulation with a short audiovisual film // Sci Data 9, 91, 2022. | ||

+ | * '''Predecessor code:''' | ||

+ | * '''Base algorithm:''' Running code based on transformer models. | ||

+ | * '''Novelty:''' Analysis of the relationship between sensor readings and human perceptions of the external world. It is required to test the hypothesis of the relationship between the data, as well as to propose a method for approximating the FMRI readings from the listening sound series. | ||

+ | * '''Authors:''' Expert Grabovoi Andrey. | ||

- | == | + | ===Problem 114=== |

- | The | + | * '''Title:''' Simulating the Dynamics of Physical Systems with Physics-Informed Neural Networks |

+ | * '''Problem description:''' The problem of choosing the optimal model for predicting the dynamics of a physical system is being solved. Under the dynamics of the system is understood the change in time of the parameters of the system. Neural networks do not have a priori knowledge about the system being modeled, which does not allow obtaining optimal parameters that take into account physical laws. The Lagrangian neural network takes into account the law of conservation of energy when modeling dynamics. In this paper, a Noetherian Agrangian neural network is proposed that takes into account the laws of conservation of momentum and angular momentum in addition to the law of conservation of energy. It is shown that for this problem the Noetherian Lagrangian neural network is optimal among the fully connected neural network model, the neural network with long-term short-term memory and the Lagrangian neural network. The simulation comparison was carried out on artificially generated data for the double pendulum system, which is the simplest chaotic system. The results of the experiments confirm the hypothesis that the introduction of a priori knowledge about the physics of the system improves the quality of the model. | ||

+ | * '''Problem description:'''Generate a set of convolutions from the available data and choose the best one using order and dimensionality reduction techniques. | ||

+ | * '''Data:''' Biomedical accelerometer and gyroscope data, ocean currents, dune movement, air currents. | ||

+ | * '''Literature:''' The base work contains references. | ||

+ | * '''Base algorithm:''' Neural network, Lagrangian neural networks. | ||

+ | * '''Solution:''' Nesterov neural network. | ||

+ | * '''Novelty:''' The proposed network takes into account the symmetry. | ||

+ | * '''Authors:''' Experts Severilov, Strijov V.V., consultant - Panchenko. | ||

- | The | + | ===Problem 115=== |

+ | * '''Title:''' Knowledge distillation in deep networks and alignment of model structures | ||

+ | * '''Problem description:''' It is required to build a network of the simplest structure, a student model, using a high quality teacher model. Show how the student's accuracy and stability change. The result of the experiment is a graph complexity-accuracy-stability, where each model is accurate. | ||

+ | * '''Data:''' CIFAR-10. It is assumed that the teacher has a structure open for analysis with a large number of layers. | ||

+ | * '''Literature:''' Hinton's original work on distillation, work by Andrei Grabovoi, work by Maria Gorpinich | ||

+ | * '''Base algorithm:''' Training (models with a given structure of controlled complexity) without distillation. Teaching (ditto) with Hinton distillation. Layered learning. Neuronal transfer learning. | ||

+ | * '''Solution:''' As in paragraph 2, only in layers. Building the path of least cost over neurons. We consider the covariance matrices of each neuron of each layer for the teacher and for the student. We propose an error function that includes the cost of the least cost path. We propose a way to construct the path of the least cost. The main idea: the transfer goes through pairs of neurons and the most similar distributions (expectation and covariance matrix) from teacher to student. | ||

+ | * '''Novelty:''' The proposed transfer significantly reduces complexity without loss of accuracy and solves the problem of interchangeability of neurons by identifying them. | ||

+ | * '''Authors:''' Experts Bakhteev Oleg, Strijov V.V., Consultant Gorpinich Maria. | ||

- | == | + | ===Problem 116=== |

- | + | * '''Title:''' Neural differential equations for modeling physical activity - selection and generation of mathematical models | |

- | + | * '''Problem description:''' The problem of choosing the optimal mat. models as the problem of genetic optimization. The optimality criterion is defined in terms of the accuracy, complexity, and stability of the model. The sampling procedure itself consists of two steps: generating a new structure and rejecting this structure if it does not satisfy the optimality criterion. Required on 'pendulum' type data - accelerometer, myogram, pulse wave - to choose the optimal model. | |

- | + | * '''Data:''' WISDM, own collection of biomedical data | |

- | + | * '''Literature:''' Neural CDE | |

- | + | * '''Base algorithm:''' Neuro ODE/CDE on a two-layer neural network. | |

- | + | * '''Solution:''' A number of experiments have already been performed, where sampling is performed by a genetic algorithm. Acceptable results have been obtained. It is proposed to analyze and improve them. | |

- | + | * '''Solution:''' Algorithm for generating mathematical models in the form of ordinary differential equations. Comparison of models and solvers on biomedical data. | |

+ | * '''Authors:''' Expert Strijov V.V., consultant Eduard Vladimirov | ||

- | == | + | ===Problem 117=== |

- | + | * '''Title:''' Search for dependencies of biomechanical systems (do people dance in pairs or independently?) and (Method of Convergence Cross-Mpping, Takens theorem) | |

- | + | * '''Problem description:''' When forecasting complex time series that depend on exogenous factors and have multiple periodicity, it is required to solve the problem of identifying related pairs of series. It is assumed that the addition of these series to the model improves the quality of the forecast. In this paper, to detect relationships between time series, it is proposed to use the convergent cross-mapping method. With this approach, two time series are connected if their trajectory subspaces exist, the projections onto which are connected. In turn, the projections of series onto trajectory subspaces are related if the neighborhood of the phase trajectory of one series is mapped to the neighborhood of the phase trajectory of another series. The problem of finding trajectory subspaces that reveal the connection of series is set. | |

+ | * '''Literature:''' Everything Sugihara wrote in Science and Nature (ask the collection). Usmanova K.R., Strijov V.V. Detection of dependencies in time series in the problems of building predictive models // Systems and means of informatics, 2019, 29(2). Neural CDE | ||

+ | * '''Data:''' Accelerometer, gyroscope, and other data describing dynamic systems | ||

+ | * '''Solution:''' Basic in Karina's work. Ours is to build the Neural ODE for both signals and decide if both models belong to the same dynamic system. | ||

+ | * '''Authors:''' Expert Strijov V.V., consultants Vladimirov, Samokhina | ||

- | == | + | ===Problem 118=== |

- | + | * '''Title:''' Continuous time when building a BCI neural interface | |

+ | * '''Problem description:''' In signal decoding The problems, data is represented as multidimensional time series. When solving problems, a discrete representation of time is used. However, recent work on neural ordinary differential equations illustrates the ability to work with the hidden state of recurrent neural networks, as with solutions to differential equations. This allows us to consider time series as continuous in time. | ||

+ | * '''Data:''' For classification: dataset P300, which was used to write an article with Alina, DEAP dataset dataset similar to it in the format of records, find a modern dataset, ask U.Grenoble-Alpes | ||

+ | * '''Literature:''' Neural CDE | ||

+ | * '''Base algorithm:''' Alina Samokhina's algorithm | ||

+ | * '''Solution:''' Using NeurODE variations to approximate the original signal. Comparative analysis of existing approaches to the application of differential equations for EEG classification. (Encoder-tensor decomposition, NeuroCDE decoder) | ||

+ | * '''Novelty:''' suggests a way to construct a continuous signal representation. Working with the functional space of the signal, not its discrete representation. Using the parameters of the resulting function as a feature space of the resulting model. | ||

+ | * '''Authors:''' Expert Strijov V.V. (was Problem 109), consultant Tikhonov | ||

- | == | + | ===Problem 119=== |

- | + | * '''Title:''' Analysis of the dynamics of multiple learning | |

- | + | * '''Problem description:''' Consider a supervised multiple learning problems in which the training set is not fixed but is updated depending on the predictions of the trained model on the test set. For the process of multiple training, prediction and updating of the sample, we build a mathematical model and study the properties of this process based on the constructed model. Let f(x) be a feature distribution density function, G be an algorithm for training the model, generating predictions on the test set and mixing predictions into the training set, as a result of which the feature distribution changes. Let the space of non-negative smooth functions F(x) be given, whose integral on R^n is equal to one. f_{t+1}(x) = G(f_{t})(x), where G(f) is the evolution operator on the space of these functions F and the initial function f_0(x) is known. In general, G can be an arbitrary operator, not necessarily smooth and/or continuous. Question 0. Find conditions on the operator G under which the image of G lies in the same class of distribution density functions F. In particular, should G be bounded, the operator norm ||G|| <= 1, so that the image of G(f) \in F is also a distribution density function for any f from F? Does there exist a unit in the space F with respect to the operator G, and what will be the identity function f in such F? Question 1. Under what conditions will there be a t_0 on G such that for all t > t_0 the tail of the sequence {f} will be bounded? Question 2. Under what conditions will the operator G have a fixed point? Data In a computational experiment, it is proposed to check the significance of the restriction / the significance of the conditions under which the answer to questions 0-2 is obtained. For example, for a problem of linear regression and/or regression with a multilevel fully connected neural network with different proportions of predictions mixed into the training set on synthetic data sets. | |

- | + | * '''Literature:''' | |

- | + | *# Khritankov A., Hidden Feedback Loops in Machine Learning Systems: A Simulation Model and Preliminary Results, https://doi.org/10.1007/978-3-030-65854-0_5 | |

- | + | *# Khritankov A.. Pilkevich A. Existence Conditions for Hidden Feedback Loops in Online Recommender Systems, https://doi.org/10.1007/978-3-030-91560-5_19 | |

- | + | *# Katok A.B., Hasselblat B. Introduction to the modern theory of dynamical systems.1999. 768 p. ISBN 5-88688-042-9. | |

- | + | *# Nemytsky V. V., Stepanov V. V. Qualitative theory of differential equations, published in 1974. | |

- | + | * '''Authors:''' Expert Khritankov A.S., Expert Afanasiev A.P. | |

- | + | ||

- | + | ||

- | + | ||

- | + | ||

- | + | ||

- | + | ||

- | # | + | |

- | # | + | |

- | + | ||

- | # | + | |

- | + | ||

- | + | ||

- | + | ||

- | == | + | ===Problem 120=== |

- | + | * '''Title:''' Differentiated algorithm for searching ensembles of deep learning models with diversity control | |

- | * | + | * '''Problem description:''' The problem of selecting an ensemble of models is considered. It is required to propose a method for controlling the diversity of basic models at the stage of application. |

- | * | + | * '''Data:''' Fashion-MNIST, CIFAR-10, CIFAR-100 datasets |

- | * | + | * '''Literature:''' |

- | * | + | *# Neural Architecture Search with Structure Complexity Control |

+ | *# Neural Ensemble Search via Bayesian Sampling | ||

+ | *# DARTS: Differentiable Architecture Search | ||

+ | * '''Base algorithm:''' It is proposed to use DARTS [3] as the basic algorithm. | ||

+ | * '''Solution:''' To control the diversity of basic models, it is proposed to use a hypernet [1], which shifts the structural parameters in terms of the Jensen-Shannon divergence. At the application stage, base architectures are sampled with a given offset to build an ensemble. | ||

+ | * '''Novelty:''' The proposed method allows building ensembles with any number of base models without additional computational costs relative to the base algorithm. | ||

+ | * '''Authors:''' K.D. Yakovlev, Bakhteev Oleg | ||

+ | ===Problem 121=== | ||

+ | * '''Problem description:''' building predictive analytics for air pollution sensors. | ||

+ | * '''Problem description:''' Data available for air quality monitoring stations in Moscow and the Moscow region (time series). The problem is to check the achievable predictive ability to predict the time series of station readings by their history + when connecting additional features (take into account the stations in aggregate, taking into account their location, time of day and weekend / working day, history and weather forecast (wind)) | ||

+ | * '''Data:''' Real data and simulations for Moscow and Moscow Region | ||

+ | * '''Authors:''' Artem Mikhailov, Vladimir Vanovsky | ||

+ | ===Problem 122=== | ||

+ | * '''Problem description:''' Reducing the dimension of space in a generative modeling problem using reversible models. | ||

+ | * '''Problem description:''' An example of a generative modeling problem is image generation. Some kinds of new models, such as normalization flows or diffusion models, define reversible transformations. But at the same time they work in a space of very high dimensions. It is proposed to combine 2 approaches: dimensionality reduction and generative modeling. | ||

+ | * '''Data:''' Any image dataset (MNIST/CIFAR10). | ||

+ | * '''Novelty:''' By reducing the dimension, you can achieve a significant acceleration of generative models, which will reduce the complexity of such models. | ||

+ | * '''Author:''' Roman Isachenko | ||

+ | |||

+ | ===Problem 123=== | ||

+ | * '''Problem description:''' Analysis of distribution bias in contrast distribution problem. | ||

+ | * '''Problem description:''' There is the same problem as Representation learning. One of the most popular approaches to solving this problem is contrastive learning. At the same time, in the data we learn from, there are often markup errors: false positive/false negative. It is proposed to analyze various ways to eliminate these biases caused by errors. And also to explore the properties of the proposed models. | ||

+ | * '''Data:''' Any image dataset (MNIST/CIFAR10). | ||

+ | * '''Novelty:''' Current models are very error sensitive. If you manage to take into account the bias in the distributions, many methods of ranking products will greatly increase in quality. | ||

+ | * '''Author:''' Roman Isachenko | ||

+ | |||

+ | ===Problem 124=== | ||

+ | * '''Title:''' Speed up sampling from diffusion models using adversarial networks | ||

+ | * '''Problem description:''' The most popular generative model today is the diffusion model. Its main disadvantage is the speed of sampling. To sample 1 picture, you need to run 1 neural network 100-1000 times. There are ways to speed up this process. One such way is to use adversarial networks. It is proposed to develop this method and explore various ways to set the functional for sampling | ||

+ | * '''Data:''' Any image dataset (MNIST/CIFAR10). | ||

+ | * '''Novelty:''' By speeding up diffusion models, they will become even more popular and easier to use. | ||

+ | * '''Author:''' Roman Isachenko | ||

+ | |||

+ | ===Problem 125=== | ||

+ | * '''Title:''' Influence of the lockdown on the dynamics of the spread of the epidemic | ||

+ | * '''Problem description:''' The introduction of a lockdown is considered an effective measure to combat the epidemic. However, contrary to intuition, it turned out that under certain conditions, a lockdown can lead to an increase in the epidemic. This effect is absent for the classical models of the spread of the epidemic “on average”, but was revealed when modeling the epidemic on the contact graph. The problem is to find formulaic and quantitative relationships between the parameters under which the lockdown can lead to an increase in the epidemic. It is necessary both to identify such relationships in the SEIRS/SEIR/SIS/etc models based on the SEIRS+ epidemiological distribution framework (and its modifications), and to theoretically substantiate the relationships obtained from specific implementations of the epidemia. | ||

+ | * '''Data:''' The problem involves working with model and synthetic data: there are ready-made data, and it is also possible to generate new ones in the process of solving the problem. This The problem belongs to unsupervised learning, since the implementation of the epidemic on the contact graph has a high proportion of random events, and therefore requires analysis on average over many synthetically generated implementations of the epidemic | ||

+ | * '''Literature:''' T. Harko, Francisco S. N. Lobo, and M. Mak. "Exact analytical solutions of the Susceptible-Infected-Recovered (SIR) epidemic model and of the SIR model with equal death and birth rates" | ||

+ | * '''Authors:''' A.Yu. Bishuk, A.V. Zuhba | ||

+ | |||

+ | ===Problem 126=== | ||

+ | * '''Title:''' Machine generation style change detection | ||

+ | * '''Problem description:'''It is required to propose a detection method | ||

+ | * '''Data:''' The sample for approximation is presented in the work of J. Berezutskay, in which there are various types of parallel signals. | ||

+ | * '''Literature:''' | ||

+ | *# G. Gritsay, A. Grabovoy, Y. Chekhovich. Automatic Detection of Machine Generated Texts: Need More Tokens // Ivannikov Memorial Workshop (IVMEM), 2022. | ||

+ | *# M. Kuznetsov, A. Motrenko, R. Kuznetsova, V. Strijov. Methods for intrinsic plagiarism detection and author diarization // Working Notes of CLEF, 2016, 1609 : 912-919. | ||

+ | *# RuATD competition. | ||

+ | * '''Base algorithm:''' Using the results of the RuATD competition as base models for classifying proposals. Use the method from Kuznetsov et all. | ||

+ | * '''Novelty:''' Suggest a method for detecting machine-generated fragments in the text using methods for changing the writing style. | ||

+ | * '''Authors:''' Expert Grabovoi Andrey | ||

+ | |||

+ | ===Problem 128=== | ||

+ | * '''Title:''' Build a deep learning model based on The problem data | ||

+ | * '''Problem description:''' is considered The problem optimization of the deep learning model for the new dataset. It is required to propose a model optimization method that allows generating new models for a new dataset with low computational costs. | ||

+ | * '''Data:''' CIFAR10, CIFAR100 | ||

+ | * '''Literature:''' variational inference for neural networks, hypernets, similar work tailored to change the model depending on a predetermined complexity | ||

+ | * '''Base algorithm:''' Retrain the model directly. | ||

+ | * '''Solution:''' The proposed method is to represent a deep learning model as a hypernet (a network that generates the parameters of another network) using a Bayesian approach. Probabilistic assumptions about the parameters of deep learning models are introduced, and a variational lower estimate of the Bayesian validity of the model is maximized. The variation estimate is considered as a conditional value, depending on the information about the problem data. | ||

+ | * '''Novelty:''' The proposed method allows you to generate models in one-shot mode (practically without retraining) for the required The problem, which significantly reduces the cost of optimization and retraining. | ||

+ | * '''Authors:''' Olga Grebenkova and Bakhteev Oleg | ||

+ | |||

+ | ===Problem 129=== | ||

+ | * '''Title:''' Spatiotemporal Prediction with Convolutional Networks and Tensor Decompositions | ||

+ | * '''Problem description:'''Generate a set of convolutions from the available data and choose the best one using order and dimensionality reduction techniques. | ||

+ | * '''Data:''' Consumption and price of electricity, ocean currents, dune movement, air currents | ||

+ | * '''Literature:''' | ||

+ | *# [http://irep.ntu.ac.uk/id/eprint/32719/1/PubSub10184_Sanei.pdf](Tensor-based Singular Spectrum Analysis for Automatic Scoring of Sleep EEG | ||

+ | *# [https://ieeexplore.ieee.org/document/6661921](Tensor based singular spectrum analysis for nonstationary source separation) | ||

+ | * '''Base algorithm:''' Caterpillar, tensor caterpillar. | ||

+ | * '''Solution:''' Find a multi-periodic time series, build its tensor representation, decompose into a spectrum, collect, show the forecast. | ||

+ | * '''Novelty:''' Show that a multilinear model is a convenient way to construct convolutions for dimensions in space and time. | ||

+ | * '''Authors:''' Expert Strijov V.V., consultant Nadezhda Alsakhanova | ||

+ | |||

+ | ===Problem 130=== | ||

+ | * '''Title:''' Automatic highlighting of terms for topic modeling | ||

+ | * '''Problem description:''' Build an ATE (Automatic Term Extraction) model for automatic extraction of phrases that are terms of the subject area in the texts of scientific articles. It is supposed to use effective collocation detection methods (TopMine or more modern) and thematic models to determine the "thematic" of the phrase. The model must be trained without a teacher (unsupervised). | ||

+ | * '''Data:''' Collection of scientific articles in the field of machine learning. Marked up articles with highlighted terms for evaluating models. | ||

+ | * '''Literature:''' | ||

+ | *# El-Kishky A., Song Y., Wang C., Voss C. R., Han J. Scalable topical phrase mining from text corpora // Proc. VLDB Endowment. _ 2014._ Vol. 8, no. 3._Pp. 305_316. | ||

+ | *# Vorontsov K. V. "Probabilistic thematic modeling: theory, models, algorithms and the BigARTM project" (http://www.machinelearning.ru/wiki/images/d/d5/Voron17survey-artm.pdf) | ||

+ | *# Nikolay Shatalov. Unsupervised learning methods for automatically highlighting compound terms in text collections. 2019. VMK MSU. | ||

+ | *# Vladimir Polushin. Topic models for ranking text content recommendations. 2017. VMK MSU. | ||

+ | *# Hanh Thi Hong Tran, Matej Martinc, Jaya Caporusso, Antoine Doucet, Senja Pollak. The Recent Advances in Automatic Term Extraction: A survey. 2023. https://arxiv.org/abs/2301.06767 | ||

+ | * '''Base algorithm:''' TopMine collocation search method • BigARTM thematic modeling library. • Modern methods based on neural network language models | ||

+ | * '''Solution:''' Application of the TopMine collocation search algorithm followed by filtering by topic. Selection of thematic model hyperparameters and thematicity criterion. Comparison of this approach with modern methods based on neural network models of the language. | ||

+ | * '''Novelty:''' Previous studies of the proposed approach have shown good results both in terms of completeness and computational efficiency. However, they have not yet been compared with neural network models. | ||

+ | * '''Authors:''' Polina Potapova, Vorontsov K.V. | ||

+ | |||

+ | ===Problem 131=== | ||

+ | * '''Title:''' Iterative improvement of the topic model with user feedback | ||

+ | * '''Problem description:''' Topic modeling is widely used in socio-humanitarian research to understand the thematic structure of large text collections. A typical use case would involve the user rating topics as relevant, irrelevant, and junk. If the number of garbage topics is too large, then the user tries to build another model. The problem is to use custom markup for each such rebuild in such a way that relevant topics are preserved, new relevant ones stand out from irrelevant and garbage topics if possible, and there are as few garbage topics as possible. | ||

+ | * '''Data:''' Any collection of natural language texts about which the thematic structure is known (about how many topics, how many documents on different topics) is suitable as data. For example, you can take a collection of Lenta news, a Wikipedia dump, posts from Habrahabr, 20 Newsgroups, Reuters, articles from PostNauka. The subject of the collection should be of interest to the researcher himself, so that there is motivation to evaluate topics manually. | ||

+ | * '''Literature:''' | ||

+ | *# Vorontsov K. V. "Probabilistic thematic modeling: theory, models, algorithms and the BigARTM project" (http://www.machinelearning.ru/wiki/images/d/d5/Voron17survey-artm.pdf ). | ||

+ | *# Alekseev V. et al. "TopicBank: Collection of coherent topics using multiple model training with their further use for topic model validation" (https://www.sciencedirect.com/science/article/pii/S0169023X21000483). | ||

+ | * '''Solution:''' Using the BigARTM theme modeling library. Use of smoothing and decorrelation regularizers. Development of methods of initialization when rebuilding thematic models. Finding a ready-made tool or developing a simple, fast, convenient way to view and markup topics. | ||

+ | * '''Novelty:''' The problem of non-uniqueness and instability of models still does not have a final solution in probabilistic thematic modeling. The proposed study is an important step towards building models with the maximum number of interpretable topics that are meaningfully useful from the point of view of humanitarian research. | ||

+ | * '''Authors:''' Vasily Alekseev, Vorontsov K. V. | ||

+ | |||

+ | ===Problem 132=== | ||

+ | * '''Title:''' Ranking of scientific articles for semi-automatic summarization | ||

+ | * '''Problem description:''' Build a ranking model that takes a selection of texts of scientific articles as input and outputs the sequence of their mention in the abstract. | ||

+ | * '''Data:''' - Overview sections (for example, Introduction and Related Work) of articles from the S2ORC collection (81.1M English-language articles) are used as a training sample. The object of the training set is a sequence of references to articles from the bibliography mentioned in the review sections. For each document there is a set of metadata - year of publication, journal, number of citations, number of citations of the author, etc. Also, there is an abstract and, possibly, the full text of the article. - Kendall's rank correlation coefficient is used as a metric. | ||

+ | * '''Literature:''' | ||

+ | *# Kryzhanovskaya S. Yu. "Technology of semi-automatic summation of thematic collections of scientific articles". | ||

+ | *# Vlasov A. V. "Methods of semi-automatic summation of collections of scientific articles". | ||

+ | *# Kryzhanovskaya S. Yu., Vorontsov K. V "Technology for semi-automatic summarization of thematic collections of scientific articles" (http://www.machinelearning.ru/wiki/images/f/ff/Idp22.pdf, p. 371), S2ORC: The Semantic Scholar Open Research Corpus. | ||

+ | * '''Base algorithm:''' Pair-wise ranking methods. Gradient boosting. | ||

+ | * '''Solution:''' The simplest solution is to rank the articles in chronological order, according to the year they were published. To solve the problem, it is proposed to build a ranking model based on gradient boosting. As signs, you can use the year of publication, the citation of the article, the citation of its authors, the semantic proximity of the publication to the review, to its local context, etc. | ||

+ | * '''Novelty:''' The problem is the first step for semi-automatic summarization of thematic collections of scientific publications (machine aided human summarization, MAHS). After the abstract script is built, the system generates prompt phrases for each article, from which the user selects phrases to continue his abstract. | ||

+ | * '''Author:''' Kryzhanovskaya Svetlana, Vorontsov K. V. | ||

+ | |||

+ | ===Problem 133=== | ||

+ | * '''Title:''' Diffusion models in the problem of generating the structure of a molecule with optimal energy | ||

+ | * '''Problem description:''' For an organic small molecule (the number of atoms is less than 100), knowing only the topology of the molecular graph is not enough to obtain the spatial structure. A molecule can have many possible configurations (conformers), each of which corresponds to a local minimum of the potential. In practice, of greatest interest are the most stable conformers, which have the lowest energy. Recent studies show the success of the application of diffusion models for the generation of molecular structures. This approach shows advanced results in the problem of generating molecules and their conformers for a small number of heavy atoms (QM9 dataset up to 9 heavy atoms in a molecule), as well as in assessing the binding of a molecule and a protein. It is proposed to build a model for the generation of conformers with minimum energy for larger molecules. | ||

+ | * '''Data:''' Base dataset QM9 | ||

+ | * '''Literature:''' | ||

+ | *# Different theoretical approaches to the diffusion model: https://arxiv.org/abs/2011.13456 | ||

+ | *# Diffusion in molecular generation: https://arxiv.org/abs/2203.17003 | ||

+ | *# Diffusion in the problem of binding a protein and a molecule: https://arxiv.org/abs/2210.01776 | ||

+ | *# Diffusion in the problem of conformer generation: https://arxiv.org/abs/2203.02923 | ||

+ | *# Tutorial on equivariant neural networks: https://arxiv.org/abs/2207.09453 | ||

+ | * '''Base algorithm:''' GeoDiff[4]. | ||

+ | * '''Solution:''' Implement conformer generation similar to DiffDock[3] for QM9 dataset. Check the performance of the model for larger molecules. | ||

+ | * '''Novelty:''' The novelty of the work lies in the design of a model for generating large conformers, which is of great practical importance. | ||

+ | * '''Author:''' Philip Nikitin | ||

+ | |||

+ | ===Problem 134=== | ||

+ | * '''Title:''' Combining distillation of models and data | ||

+ | * '''Problem description:''' Knowledge distillation is the transfer of knowledge from a more meaningful representation to a compact, concise representation. There are two kinds of knowledge distillation. The first is the distillation of models. In this case, the large model transfers knowledge (distilled) to the small model. The second is data distillation. In this case, a minimum data set is created, on which, after training the model, it achieves a quality comparable to training on a full sample. At the moment, there is no solution that can implement simultaneous distillation of model and knowledge. Therefore, the goal of The problem is to propose a basic solution for model distillation and compare with approaches to model distillation and data distillation. | ||

+ | * '''Data:''' MNIST handwritten digit sampling, CIFAR-10 image sampling | ||

+ | * '''Literature:''' | ||

+ | *# A collection of various papers on the distillation of data. | ||

+ | *# Review on methods of distillation models. | ||

+ | *# Basic knowledge distillation solution. | ||

+ | *# Basic solution for model distillation. | ||

+ | * '''Base algorithm:''' Basic Model Distillation Solution, Hinton Distillation Basic Dataset Distillation Solution, Dataset Distillation by Matching Training Trajectories | ||

+ | * '''Solution:''' It is proposed to implement data distillation as a basic algorithm. Then train a larger model on the data and distill it into a smaller model. Next, compare with the original model and the model trained on distilled data. | ||

+ | * '''Novelty:''' The novelty of the work lies in the combination of two distillation approaches, which has not been implemented before | ||

+ | * '''Authors:''' Andrey Filatov | ||

+ | |||

+ | ===Problem 135=== | ||

+ | * '''Title:''' Proximity measures in self-supervised learning The problems | ||

+ | * '''Problem description:''' The idea of self-supervised learning is to solve an artificially selected The problem to get useful representations of data without markup. One of the most popular approaches is the use of contrastive learning, during which the model is trained to minimize the distance between representations of augmented copies of the same object. The purpose of The problem is to investigate the quality of the resulting representations depending on the choice of the proximity measure (similarity measure) used in training, and to offer our own version of distance measurement | ||

+ | * '''Data:''' CIFAR-100 | ||

+ | * '''Literature:''' | ||

+ | *# Solution using squared Euclidean distance. | ||

+ | *# Solution using cosine similarity. | ||

+ | *# Decision based on the information principle. | ||

+ | * '''Base algorithm:''' VicReg, Barlow Twins, SimSiam | ||

+ | * '''Solution:''' One of the distance options that can be proposed is an analogue of the Vaserstein metric, which would allow taking into account the dependencies between features. | ||

+ | * '''Novelty:''' Propose a new way to determine the measure of proximity, which would be theoretically justified / contributed to obtaining representations with given properties | ||

+ | * '''Authors:''' Polina Barabanshchikova | ||

+ | |||

+ | ===Problem 136=== | ||

+ | * '''Title:''' Stochastic Newton with Arbitrary Sampling | ||

+ | * '''Problem description:''' We analyze second order methods solving Empirical Risk Minimization problem of the form min f(x) in R^d. Here x is a parameter vector of some Machine Learning model, f_i(x) is a loss function on i-th training point (a_i,b_i). Our desire to solve it using Newton-type method that requires access to only one data point per iteration. We investigate different sampling strategies of index i_k on iteration k. See description in PDF. | ||

+ | * '''Data:''' It is proposed to use open SVM library as a data for experimental part of the work. | ||

+ | * '''References:''' | ||

+ | *# Stochastic Newton and Cubic Newton Methods with Simple Local Linear-Quadratic Rates | ||

+ | *# Parallel coordinate descent methods for big data optimization | ||

+ | * '''Base algorithm:''' As a base method it is proposed to use Algorithm 1 from the paper Stochastic Newton and Cubic Newton Methods with Simple Local Linear-Quadratic Rates. | ||

+ | * '''Solution:''' Is is proposed to adjust existing sampling strategies from Parallel coordinate descent methods for big data optimization in this work. | ||

+ | * '''Novelty:''' In the literature of Second Order methods there are a few works on incremental methods. The idea is to analyze the existing method by applying different sampling strategies. It is known that the proper sampling strategies may improve the performance of a method. | ||

+ | * '''Authors:''' Islamov Rustem, Vadim Strijov | ||

+ | |||

+ | ===Problem 139=== | ||

+ | * '''Title:''' Distillation of models on multidomain selections. | ||

+ | * '''Problem description:''' The problem of reducing the complexity of the approximating model when transferred to new data of lower power is investigated. | ||

+ | * '''Data:''' Samples MNIST, CIFAR-10, CIFAR-100, Amazon products. | ||

+ | * '''Literature:''' Diploma Kamil Bayazitov | ||

+ | * '''Base algorithm:''' The basic solution and experiments are presented in the thesis. | ||

+ | * '''Authors:''' Grabovoi Andrey | ||

+ | |||

+ | ===Problem 140=== | ||

+ | * '''Title:''' Tailoring the architecture of a performance-controlled deep learning model | ||

+ | * '''Problem description:''' considers The problem adapting the structure of a trained deep learning model for limited computing resources. It is assumed that the resulting architecture (or several architectures) should work efficiently on several types of computing servers (for example, on different GPU models or different mobile devices). It is required to propose a model search method that allows controlling its complexity taking into account the target performance characteristics. | ||

+ | * '''Data:''' MNIST, CIFAR | ||

+ | * '''Literature:''' | ||

+ | *# Grebenkova O.S., Bakhteev Oleg O., Strijov V.V. V.V. Variational optimization of a deep learning model with complexity control // Informatics and its applications, 2021, 15(2). PDF | ||

+ | *# Yakovlev K. D. et al. Neural Architecture Search with Structure Complexity Control //Recent Trends in Analysis of Images, Social Networks and Texts: 10th International Conference, AIST 2021, Tbilisi, Georgia, December 16–18, 2021, Revised Selected Papers. Cham: Springer International Publishing, 2022. - pp. 207-219. | ||

+ | *# FBNet: choosing a model architecture based on target characteristics | ||

+ | * '''Base algorithm:''' FBNet and random search of model substructure | ||

+ | * '''Solution:''' The proposed method is to use a differentiable neural network architecture search algorithm (FBNet) with parameter complexity control using a hypernet. A hypernetwork is a model that generates the structure of the model depending on the input parameters. It is proposed to use the normalized running time of basic operations on target computing resources as hypernet parameters. Thus, the resulting model will allow adapting the architecture of the model for an arbitrary device. * '''Novelty:''' The proposed method allows you to control the complexity of the model, in the process of searching for an architecture without additional heuristics. | ||

+ | * '''Authors:''' Konstantin Yakovlev, Bakhteev Oleg | ||

+ | |||

+ | ==2022== | ||

+ | ===Results=== | ||

{|class="wikitable" | {|class="wikitable" | ||

|- | |- | ||

- | ! | + | ! Author |

- | ! | + | ! Topic |

- | + | ||

! Links | ! Links | ||

+ | ! Consultant | ||

+ | ! Letters | ||

|- | |- | ||

- | | | + | |[https://github.com/anton39reg Pilkevich Anton] |

- | | | + | | Existence conditions for hidden feedback loops in recommender systems |

- | | | + | |[https://github.com/Intelligent-Systems-Phystech/2021-Project-74 GitHub], [https://docs.google.com/document/d/1OLCqkmArjqFn8M9pB5C_kLoYOv0l1w9RjHy0y0upPew/edit?usp=sharing LinkReview], |

- | + | [https://github.com/Intelligent-Systems-Phystech/2021-Project-74/raw/main/docs/Pilkevich2021HiddenFeedbackLoops.pdf Paper], [https://github.com/Intelligent-Systems-Phystech/2021-Project-74/raw/main/docs/Pilkevich2021Presentation/Pilkevich2021Presentation.pdf Slides], | |

+ | [https://www.youtube.com/watch?v=xW_lXGn1WHs&t=24s Video], [https://youtu.be/9ELhIqjFSE8 Video] | ||

+ | |[https://intelligent-systems-phystech.github.io/ru/people/khritankov_as/index.html Khritankov] | ||

+ | | AILB.P-X+R-B-H1CVO.T-EM.H1WJSF | ||

|- | |- | ||

- | | | + | |[https://github.com/Edyarich Vladimirov Eduard] |

- | | | + | |Restoration of the trajectory of hand movement from video |

+ | |[https://github.com/Intelligent-Systems-Phystech/2022-Project-90 GitHub], [https://docs.google.com/document/d/1RpWz1sqpgwnf-ewTe4OHI_WODGklx5FBjLfzvHkIUYQ/edit?usp=sharing LinkReview], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2022-Project-90/raw/master/paper/Vladimirov2022RestoringHandMovement.pdf Paper], [https://github.com/Intelligent-Systems-Phystech/2022-Project-90/blob/master/slides/Vladimirov2022Presentation.pdf Slides] | ||

+ | |[https://github.com/r-isachenko Isachenko] | ||

+ | |(B.O.H1M)ALI+PXRBС+V+TED? | ||

+ | |- | ||

+ | |[https://github.com/pkseniya Petrushina Ksenia] | ||

+ | | Anti-Distillation: Knowledge Transfer from Simple Model to a Complex One | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2022-Project-97 GitHub], [https://docs.google.com/document/d/1ekpNeQnvnpXP_Jwp07llyZArH85IZO7Bz1UAlTme7Xs/edit?usp=sharing LinkReview], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2022-Project-97/blob/master/paper/Petrushina2022AntiDistillation.pdf Paper], [https://github.com/Intelligent-Systems-Phystech/2022-Project-97/blob/master/slides/Petrushina2022Presentation.pdf Slides] | ||

+ | |[https://github.com/andriygav Grabovoi] | ||

+ | | (B.O.H1M)ALIPXRBСVTED | ||

+ | |- | ||

+ | |[https://github.com/Jhomanik Kornilov Nikita] | ||

+ | | Winterstorm risk prediction via machine learning methods | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2022-Project-93-1 GitHub], [https://docs.google.com/document/d/1XAld9YsJ-R7Jv-i5SkIGNxX5Hy8vShPv8BA_jig9XcQ/edit?usp=sharing LinkReview], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2022-Project-93-1/raw/master/paper/Kornilov2022Winterstorm.pdf Paper], [https://github.com/Intelligent-Systems-Phystech/2022-Project-93-1/raw/master/slides/Winterstorm_presentation.pdf Slides] | ||

+ | | Yuri Maksimov | ||

+ | | (B.O.H1M?)ALIPXRBСV+TE0D | ||

+ | |- | ||

+ | |[https://github.com/AlievAE Aliyev Alen] | ||

+ | | Geometric Deep Learning for Protein-Protein Binding Affinity Prediction | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2022-Project-103 GitHub], [https://docs.google.com/document/d/1J6nfi3nclsB6TOgcoqokSlli0u0YOqPpKzhZ7h0Xltw LinkReview], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2022-Project-103/blob/master/docs/Aliev2022PpbAffinityPrediction.pdf Paper], [https://github.com/Intelligent-Systems-Phystech/2022-Project-103/blob/master/slides/Aliev2022Presentation.pdf Slides] | ||

+ | | Ilya Igashov | ||

+ | | (B.O.H1M?)ALIPXRBСVTED? | ||

+ | |- | ||

+ | |[https://github.com/IvanLukianenko Lukyanenko Ivan] | ||

+ | | Hail Prediction Using Graph Neural Networks | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2022-Project94 GitHub], [https://docs.google.com/document/d/1ntAjEcvUhdgxM4CZCwmWDq8fBXiOrqBKh92rto4C92Q/edit?usp=sharingLinkReview], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2022-Project-94/blob/master/paper/Hail%20risk%20prediction%20with%20HailNet.pdf Paper], [https://github.com/Intelligent-Systems-Phystech/2022-Project-94/blob/master/slides/Hail%20risk%20prediction%20via%20Graph%20Neural%20Networks%20Slides.pdf Slides] | ||

+ | | Yuri Maksimov | ||

+ | | (B.O.H1M?)ALIPXRBСV+TED? | ||

+ | |- | ||

+ | |[https://github.com/Maxgaponov Gaponov Maxim] | ||

+ | | Choosing Interpretable Recurrent Deep Learning Models | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2022-Project-99 GitHub], [https://docs.google.com/document/d/1R-IAGa-w5Edc23jfB_68OZ34EiBlRq6Yaoc1XR_mQ9g/edit?usp=sharing LinkReview], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2022-Project-99/blob/master/paper/Gaponov2022InterpretableRNN.pdf Paper], [https://github.com/Intelligent-Systems-Phystech/2022-Project-99/blob/master/slides/Gaponov2022InterpretableRNNSlides.pdf Slides] | ||

+ | |[https://github.com/bahleg Bakhteev Oleg] | ||

+ | | (B.O.H1M)AL+IPXRBСVT???ED | ||

+ | |- | ||

+ | |[https://github.com/MelnikovIgor1 Melnikov Igor] | ||

+ | | Stochastic Newton with Arbitrary Sampling | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2022-Project-101 GitHub], [https://docs.google.com/document/d/1wwLvqBrUV3atwJfnlqVRAhSk-KlUzbUpW6K_aaJ8arQ/edit?usp=sharing LinkReview], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2022-Project-101/raw/master/paper/Melnikov2022StochasticNewtonWithArbitrarySampling.pdf Paper], [https://github.com/Intelligent-Systems-Phystech/2022-Project-101/raw/master/slides/one-slide.pdf Slides] | ||

+ | |[https://github.com/Rustem-Islamov Rustem Islamov] | ||

+ | | (B.O.H1M)ALIPXСRBVTED | ||

+ | |- | ||

+ | |[https://github.com/fzmushko Zmushko Philip] | ||

+ | | Continuous time when building a BCI neural interface | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2022-Project-109 GitHub], [https://docs.google.com/document/d/1tpH34r2x4vRWgaBeBkf8yp__-qGyDQNST-w7X29qgPg/edit?usp=sharing LinkReview], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2022-Project-109/blob/master/paper/Zmushko2022ContinuousTime.pdf Paper], [https://github.com/Intelligent-Systems-Phystech/2022-Project-109/blob/master/slides/Zmushko2022Presentation.pdf Slides] | ||

+ | |[https://github.com/Alina-Samokhina Samokhina] | ||

+ | | (B.O.H1M)ALI0P0XR?BСVTE?D? | ||

+ | |- | ||

+ | |[https://github.com/hadingus Tishchenko Evgeny] | ||

+ | | Cross-language duplicate search | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2022-Project-104 GitHub], [https://docs.google.com/document/d/13bZ_Cs5Q-tAfuSEPXVMw-uqTtZkkvoUxF35pRSfx7bI/edit?usp=sharing LinkReview], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2022-Project-104/blob/master/paper/Tishchenko2022PlagiatDetecting.pdf Paper], [https://github.com/Intelligent-Systems-Phystech/2022-Project-104/blob/master/slides/Tishchenko2022AntiplagiatDetectionSlides.pdf Slides] | ||

+ | | Konstantin Vorontsov | ||

+ | | (B.O.H1M)ALIPXRB0СV0T?E?D? | ||

+ | |- | ||

+ | |[https://github.com/JustAnotherArchetype Antyshev Tikhon] | ||

+ | | Compression for Federated Random Reshuffling | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2022-Project107 GitHub], [https://docs.google.com/document/d/1T0bsAXp2P8kWmhCtI2lV0KVi4neEdu6FabkWxrAd3aI/edit?usp=sharing LinkReview], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2022-Project107/blob/master/paper/Antyshev2022CompressionforFedRR.pdf Paper], [https://github.com/Intelligent-Systems-Phystech/2022-Project107/blob/master/slides/Antyshev2022Presentation.pdf Slides] | ||

+ | |[https://grigory-malinovsky.github.io/ Malinovsky] | ||

+ | | (B.O.H1_M?)ALI-PXRBСVT? | ||

+ | |- | ||

+ | |[https://github.com/vladpyzh Pyzh Vladislav] | ||

+ | | Flood risk prediction via machine learning methods | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2022-Project-93-2 GitHub], [https://docs.google.com/document/d/1eKr7KS_ONyhj9B5ZupALz_ejm9SgO1rmoTvOAmW10G8/edit?usp=sharing LinkReview], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2022-Project-93-2/raw/master/docs/Pyzh2022Title.pdf Paper], [https://www.overleaf.com/read/tbrgqmyttnnb Online Draft], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2022-Project-93-2/raw/master/docs/presentation.pdf Slides] | ||

+ | | Yuri Maksimov | ||

+ | | (B.O.H10M?)ALI0P0XRBСVT0ED? | ||

+ | |- | ||

+ | |[https://github.com/Egor-s-gor Zharov Georgy] | ||

+ | | Forest fire risk assessments using machine learning methods | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2022-Project-93 GitHub], [https://docs.google.com/document/d/17LqpAAdnIwbVIq9dLdZA7z9eBnaf_nd-Dp0_kBKXxYA/edit?usp=sharing LinkReview], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2022-Project-93/blob/master/paper/First_paper_Zharov_Wildfires.pdf Paper], [https://github.com/Intelligent-Systems-Phystech/2022-Project-93/blob/master/slides/talk.pdf Slides] | ||

+ | | Yuri Maksimov | ||

+ | | (B.O.H1)ALIPX0R0B0С0V0T?E0D? | ||

+ | |- | ||

+ | |[https://github.com/TimkaMLG Muradov Timur] | ||

+ | | Choosing Interpretable Convolutional Deep Learning Models | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2022-Project99 GitHub], [https://docs.google.com/document/d/177wuzjmAuY4BpG7325QSH9SkS4SBCgWCKBFXzc68YA0/edit LinkReview], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2022-Project99/raw/master/paper/Muradov2022InterpretableCNN.pdf Paper], [https://github.com/Intelligent-Systems-Phystech/2022-Project99/raw/master/slides/Muradov2022Presentation.pdf Slides] | ||

+ | |[https://github.com/bahleg Bakhteev] | ||

+ | | (B.O.H1)ALI0P0XRBСV0T0E?D? | ||

+ | |- | ||

+ | |[https://github.com/YHx07 Pavlov Dmitry] | ||

+ | | Machine learning approach to startup success prediction | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2022-Project-vc GitHub], [https://www.overleaf.com/read/zswjpqgmrcmw Online Draft], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2022-Project-vc/blob/master/paper/2022_Project_vc.pdf Paper], [https://github.com/Intelligent-Systems-Phystech/2022-Project-vc/blob/master/slides/2022_Project_vc.pdf Slides] | ||

+ | | Anton Moiseev, Yuri Ammosov | ||

+ | | (B.O.H10M?)ALI?P?XRBСV?T0E0D0 | ||

+ | |- | ||

+ | |} | ||

+ | |||

+ | ===Problem 100.2022 (group)=== | ||

+ | * '''Title:''' Multi-model representation of dynamical systems | ||

+ | * '''Problem description:''' The system described by attractors in several phase spaces is considered. Particular models are constructed that approximate measurements of the state of the system in each space. A matching multimodel is built. The parameters of private models are specified. | ||

+ | * '''Data:''' Human motion video, accelerometer, gyroscope, electroencephalogram signals | ||

+ | * '''Literature:''' Our work on accelerometers and BCI, dissertations by Motrenko, Isachenko, Grabovoi | ||

+ | * '''Base algorithm:''' Particular models are neural networks, multimodel is canonical correlation analysis and multimodel is distilled. | ||

+ | * '''Solution:''' Generalize canonical correlation analysis and distillation to the case of an arbitrary number of models. | ||

+ | * '''Novelty:''' Alignment space built for a set of heterogeneous models | ||

+ | * '''Authors:''' A.V. Grabovoi, Strijov V.V. | ||

+ | |||

+ | ===Problem 90.2022=== | ||

+ | * '''Title:''' Hand movement recovery from video | ||

+ | * '''Problem description:''' A skeletal representation of a person's pose is restored from the video sequence. The trajectory of the movement of human limbs sets the initial phase space. The accelerometer signal from the limbs sets the target phase space. Build a model that connects the attractors of the trajectories of the source and target spaces. | ||

+ | * '''Data:''' The initial sample is collected by the authors of the project. Parts of the selection are in the library examples. | ||

+ | * '''Solution:''' Theoretical part executed by the extended command. Perform a theoretical study: show that the canonical correlation analysis method (and in particular the PLS, NNPLS, seq2seq, Neur ODE methods) are special cases of the Sugihara convergent cross mapping method. | ||

+ | * '''Novelty:''' A reversible model has been introduced that maps the coordinates recovered from the video sequence into the accelerations of the mobile phone's accelerometer. | ||

+ | * '''Authors:''' A.D. Kurdyukova, R.I. Isachenko, Strijov V.V. | ||

+ | |||

+ | ===Problem 91.2022=== | ||

+ | * '''Title:''' Clustering human movement trajectories | ||

+ | * '''Problem description:''' This paper analyzes the periodic signals in the time series to recognize human activity by using a mobile accelerometer. Each point in the timeline corresponds to a segment of historical time series. This segments form a phase trajectory in phase space of human activity. The principal components of segments of the phase trajectory are treated as feature descriptions at the point in the timeline. The paper introduces a new distance function between the points in new feature space. To reval changes of types of the human activity the paper proposes an algorithm. This algorithm clusters points of the timeline by using a pairwise distances matrix. The algorithm was tested on synthetic and real data. This real data were obtained from a mobile accelerometer | ||

+ | * '''Data:''' USC-HAD, new accelerometer samples | ||

+ | * '''Literature:''' Grabovoy A.V., Strijov V.V. Quasi-periodic time series clustering for human activity recognition // Lobachevskii Journal of Mathematics, 2020, 41 : 333-339. | ||

+ | * '''Base algorithm:''' Caterpillar | ||

+ | * '''Solution:''' Bring Grabovoi's article from the Lobachevsky Journal of Mathematics to perfection | ||

+ | * '''Novelty:''' Use Neuro ODE to plot the phase trajectory and classify it | ||

+ | * '''Authors:''' A.V. Grabovoi (ask!!), Strijov V.V. | ||

+ | |||

+ | ===Problem 97.2022=== | ||

+ | * '''Title:''' Anti-distillation or teacher training: knowledge transfer from a simple model to a complex one | ||

+ | * '''Problem description:''' The problem of adapting the model to a new sample with a large amount of information is considered. For adaptation, it is proposed to build a new model of greater complexity with further transfer of information from a simple model to it. When transferring information, it is necessary to take into account not only the quality of the forecast on the original sample, but also the adaptability of the new model to the new sample and the robustness of the solution obtained. | ||

+ | * '''Data:''' MNIST handwritten digit sampling, CIFAR-10 image sampling | ||

+ | * '''Literature:''' Original distillation problem statement: Hinton G. et al. Distilling the knowledge in a neural network //arXiv preprint arXiv:1503.02531 | ||

+ | * '''Base algorithm:''' It is proposed to increase the complexity of the model by including constant values close to zero in the model. This approach is basic, because can lead to a decrease in the robustness of the model and worse adaptability to a new sample. | ||

+ | * '''Solution:''' It is proposed to consider several approaches to increase the complexity of the model, including both probabilistic (adding noise to new parameters, taking into account operational requirements) and algebraic (expanding the parametric space of the model, taking into account the requirements for robustness and constant Lipschitz of the original model) | ||

+ | * '''Novelty:''' obtaining a method that allows you to adapt the existing model to complicate the training sample without losing information | ||

+ | * '''Authors:''' Bakhteev, Grabovoi, Strijov V.V. | ||

+ | |||

+ | ===Problem 98.2022=== | ||

+ | * '''Title:''' Deep learning model selection with expert model matching control | ||

+ | * '''Problem description:''' is considered The problem classification. An expert model of low complexity is specified. It is required to build a deep learning model that gives a high quality of the forecast and is similar in behavior to the expert model. | ||

+ | * '''Data:''' Sociological samples, CIFAR image sample | ||

+ | * '''Literature:''' Yakovlev Konstantin, Grebenkova Olga, Bakhteev Oleg, Strijov Vadim. Neural architecture search with structure complexity control // Communications in Computer and Information Science (Proceedings of the 10th International Conference on Analysis of Images, Social Networks and Texts), 2021 | ||

+ | * '''Base algorithm:''' building an expert model. | ||

+ | * '''Solution:''' The proposed method consists in hypernetworks with control of the consistency of the found model with the expert model. A hypernetwork is a deep learning model that generates the parameters of the target model. | ||

+ | * '''Novelty:''' the proposed method allows to take into account expert judgment in the process of model selection and architecture search. | ||

+ | * '''Authors:''' Grebenkova, Bakhteev, Strijov V.V. | ||

+ | |||

+ | ===Problem 99.2022=== | ||

+ | * '''Title:''' Selection of interpretable convolutional deep learning models | ||

+ | * '''Problem description:''' Considers The problem of choosing an interpretable deep learning classification model. Interpretability is understood as the ability of the model to: a) return the most significant features of an object for classification, b) determine clusters of objects that are similar from the point of view of the classifier | ||

+ | * '''Data:''' MNIST handwritten digit sampling, CIFAR-10 image sampling | ||

+ | * '''Literature:''' | ||

+ | *# [https://arxiv.org/pdf/1802.06259.pdf Exact and Consistent Interpretation for Piecewise Linear Neural Networks: A Closed Form Solution] | ||

+ | *# [https://arxiv.org/abs/1602.04938 "Why Should I Trust You?": Explaining the Predictions of Any Classifier] | ||

+ | * '''Base algorithm:''' The LIME(1) algorithm interprets the model by local approximation | ||

+ | * '''Solution:''' A solution based on the method described in (2) is proposed. In this paper, a generalization of the multilayer perzpetron model with a piecewise linear activation function was proposed. Such an activation function allows us to consider the classifier for each sample object as a locally linear one, without using approximation. It is proposed to generalize the proposed approach to the main nonlinear functions used in convolutional neural networks: convolution, pooling and normalization functions. | ||

+ | * '''Novelty:''' is to obtain a new class of neural models that lend themselves to good interpretation. | ||

+ | * '''Authors:''' Yakovlev, Bakhteev, Strijov V.V. | ||

+ | |||

+ | ===Problem 01.2022=== | ||

+ | * '''Title:''' Stochastic Newton with Arbitrary Sampling | ||

+ | * '''Problem:''' We analyze second order methods solving Empirical Risk Minimization problem of the form min f(x) in R^d. Here x is a parameter vector of some Machine Learning model, f_i(x) is a loss function on i-th training point (a_i,b_i). Our desire to solve it using Newton-type method that requires access to only one data point per iteration. We investigate different sampling strategies of index i_k on iteration k. See description in [http://www.machinelearning.ru/wiki/images/5/5c/Stochastic_Newton_with_Arbitrary_Sampling.pdf PDF]. | ||

+ | * '''Dataset:''' It is proposed to use open SVM library as a data for experimental part of the work. | ||

+ | * '''References:''' | ||

+ | *# Stochastic Newton and Cubic Newton Methods with Simple Local Linear-Quadratic Rates | ||

+ | *# Parallel coordinate descent methods for big data optimization | ||

+ | * '''Base algorithm:''' As a base method it is proposed to use Algorithm 1 from the paper Stochastic Newton and Cubic Newton Methods with Simple Local Linear-Quadratic Rates. | ||

+ | * '''Solution:''' Is is proposed to adjust existing sampling strategies from Parallel coordinate descent methods for big data optimization in this work. | ||

+ | * '''Novelty:''' In the literature of Second Order methods there are a few works on incremental methods. The idea is to analyze the existing method by applying different sampling strategies. It is known that the proper sampling strategies may improve the performance of a method. | ||

+ | * '''Authors:''' Islamov Rustem, Vadim Strijov | ||

+ | |||

+ | ===Problem 107.2022=== | ||

+ | * '''Title:''' Compression for Federated Random Reshuffling | ||

+ | * '''Problem:''' We analyze first order methods solving Empirical Risk Minimization problem of the form min f(x) in R^d. Here x is a parameter vector of some Machine Learning model, f_i(x) is a loss function on i-th training point (a_i,b_i). We focus on distributed setting of this problem. We are going to apply compression techniques to reduce number of communicated bits to overcome communication bottleneck. Also we want to combine it with server-side updates. We desire to generalize and get improvement in theory and practice. | ||

+ | * '''Dataset:''' It is proposed to use open SVM library as a data for experimental part of the work. | ||

+ | * '''References:''' | ||

+ | *# [https://fl-icml.github.io/2021/papers/FL-ICML21_paper_34.pdf Federated Random Reshuffling with Compression and Variance Reduction] | ||

+ | *# [https://arxiv.org/pdf/2102.06704.pdf Proximal and Federated Random Reshuffling] | ||

+ | *# [https://arxiv.org/pdf/2201.11066.pdf Server-Side Stepsizes and Sampling Without Replacement Provably Help in Federated Optimization] | ||

+ | * '''Base algorithm:''' As a base method we use Algorithm 3 from [https://arxiv.org/pdf/2102.06704.pdf Proximal and Federated Random Reshuffling]. | ||

+ | * '''Solution:''' Is is proposed to combine the method with two stepsizes with compression operators. | ||

+ | * '''Novelty:''' This would be the first method combining 4 popular federated learning techniques: local steps, compression, reshuffling of data and two stepsizes. | ||

+ | * '''Authors:''' Grigory Malinovsky | ||

+ | |||

+ | ===Problem 108.2022=== | ||

+ | * '''Title:''' Distillation of knowledge using sample representation in the common latent space of models | ||

+ | * '''Problem description:''' Considers The problem of distillation - the transfer of information from one or more teacher models to the student. A special case is considered when teachers have incomplete information about the sample, and each model has useful information only about some subset. | ||

+ | * '''Data:''' Sample CIFAR-10 images; sampling of handwritten MNIST digits | ||

+ | * '''Literature:''' | ||

+ | *# Hinton G. et al. Distilling the knowledge in a neural network //arXiv preprint arXiv:1503.02531. - 2015. - Vol. 2. - No. 7. | ||

+ | *# Oki H. et al. Triplet Loss for Knowledge Distillation //2020 International Joint Conference on Neural Networks (IJCNN). - IEEE, 2020. - P. 1-7. | ||

+ | * '''Base algorithm:''' Hinton distillation [1]. | ||

+ | * '''Solution:''' It is proposed to consider hidden representations of teachers and students obtained using dimensionality reduction algorithms. To align the model spaces, it is proposed to use the autoencoder model with triplet constraints (see, for example, [2]). | ||

+ | * '''Novelty:''' The proposed method will allow the distillation of heterogeneous models, using information from several teachers. | ||

+ | * '''Authors:''' Gorpinich, Bakhteev, Strijov V.V. | ||

+ | |||

+ | ===Problem 93.2022=== | ||

+ | * '''Title:''' Estimating the risk of forest fires using machine learning methods. | ||

+ | * '''Problem description:''' Wildfire risk prediction based on climate variables (water/air temperature, atmospheric pressure) since 1991. Forecasting is carried out (a) in the short-term range (2-5 years; stationary time series) and (b) in the long-term range (up to 50 years; non-stationary time series). A feature of forecasting in the long range is the (probable) significant change in the behavior of climate variables (CMIP5 scenarios). The key features of problem (1) are the need for a sufficiently accurate prediction of extreme risk values (maximum values of the time series), while the algorithm can make a significant number of errors in the region of small values of the series. (2) the spatial data structure of the series. | ||

+ | * '''Data:''' | ||

+ | *# [https://developers.google.com/earth-engine/datasets/catalog/IDAHO_EPSCOR_TERRACLIMATE Google Earth Data] - data on climate variables and landscape available via API (there is a jupyter notebook through which you can download data locally) | ||

+ | *# [https://www.worldclim.org/data/cmip6/cmip6_clim2.5m.html CMIP5] climate scenarios (there is a jupyter notebook through which you can download data locally) | ||

+ | *# [https://daac.ornl.gov/cgi-bin/theme_dataset_lister.pl?theme_id=8 Wildfire Risk Database] | ||

+ | *# [https://www.visualcrossing.com/weather/weather-data-services Severe Weather Dataset] | ||

+ | * '''Literature:''' | ||

+ | *# [http://staff.ustc.edu.cn/~hexn/papers/kdd19-timeseries.pdf Daizong Ding, Mi Zhang, Xudong Pan, Min Yang, Xiangnan He. Modeling Extreme Events in Time Series Prediction. KDD-2019]. | ||

+ | *# [https://arxiv.org/abs/2004.09140 Roman Kail, Alexey Zaytsev, Evgeny Burnaev. Recurrent Convolutional Neural Networks help to predict the location of Earthquakes]. | ||

+ | *# [http://roseyu.com/time-series-workshop/submissions/TSW2017_paper_3.pdf Nikolay Laptev, Jason Yosinski, Li Erran Li, Slawek Smyl. Time-series Extreme Event Forecasting with Neural Networks at Uber]. | ||

+ | * '''Base algorithm:''' (1) method from article 1, (2). ST-LSTM | ||

+ | * '''Solution:''' is proposed to solve the problem in two steps. At the first step, Algorithm 1 (with the addition of a spatial component) restores (averaged over a certain range) the behavior of the time series. Next, the discrepancy between the values of the series and the model is analyzed. Based on this, the noise distribution is restored and a probabilistic model is built to achieve a certain level of risk in a given territory in the required time range. | ||

+ | * '''Novelty:''' (geo)-spatial time series prediction is an open area with great potential for theoretical and practical work. In particular, fire risk assessment is necessary for (1) predicting the probability of accidents (electric power industry, gas transport complex); (2) prioritization of fire prevention measures by region; (3) assessing the financial risks of companies operating in the region. | ||

+ | * '''Authors:''' Yuri Maksimov, Alexey Zaitsev | ||

+ | * '''Consultants:''' Yuri Maksimov, Alexey Zaitsev, Alexander Lukashevich. | ||

+ | |||

+ | ===Problem 94.2022=== | ||

+ | * '''Title:''' Hail forecast using graph neural networks | ||

+ | * '''Problem description:''' Hail risk prediction based on climate variables (water/air temperature, atmospheric pressure) since 1991. Forecasting is carried out (a) in the short-term range (2-5 years; stationary time series) and (b) in the long-term range (up to 50 years; non-stationary time series). A feature of forecasting in the long range is the (probable) significant change in the behavior of climate variables (CMIP5 scenarios). Key features of The problem (1) rare events, the case of hail in Russia over the past 30 years was less than 700 throughout the country (2) the spatial structure of the data series. | ||

+ | * '''Data:''' | ||

+ | *# [https://developers.google.com/earth-engine/datasets/catalog/IDAHO_EPSCOR_TERRACLIMATE Google Earth Data] - data on climate variables and landscape available via API (there is a jupyter notebook through which you can download data locally) | ||

+ | *# [https://www.worldclim.org/data/cmip6/cmip6_clim2.5m.html CMIP5] climate scenarios (there is a jupyter notebook through which you can download data locally) | ||

+ | *# [https://www.ncdc.noaa.gov/stormevents/ftp.jsp NOAA Storm Events Database] | ||

+ | *# [https://eswd.eu/cgi-bin/eswd.cgi European Severe Weather Database] | ||

+ | *# [https://www.visualcrossing.com/weather/weather-data-services Severe Weather Dataset] | ||

+ | * '''Literature:''' | ||

+ | *# Ayush, Kumar, et al. "Geography-aware self-supervised learning." [https://openaccess.thecvf.com/content/ICCV2021/papers/Ayush_Geography-Aware_Self-Supervised_Learning_ICCV_2021_paper.pdf Proceedings of the IEEE/CVF International Conference on Computer Vision. 2021]. | ||

+ | *# Cachay, Salva Rühling, et al. "Graph Neural Networks for Improved El Ni\~ no Forecasting." arXiv preprint arXiv:2012.01598 (2020). [https://arxiv.org/pdf/2012.01598.pdf NeurIPS Clima Workshop]. | ||

+ | *# Cai, Lei, et al. "Structural temporal graph neural networks for anomaly detection in dynamic graphs." [https://dl.acm.org/doi/pdf/10.1145/3459637.3481955 Proceedings] of the 30th ACM International Conference on Information & Knowledge Management. 2021. | ||

+ | * '''Base algorithm:''' classification with extremely rare events, the most basic variant of log-regression + SMOTE. The paper proposes to take a combination of algorithms from articles 2 and 3 as a basis. | ||

+ | * '''Solution:''' suggests that a combination of the algorithms from articles 2 and 3 can improve classification in such The problems with exceptionally rare events. In addition, it is supposed to use physical information to regularize the classifier (combination of temperature/humidity factors at which hail is most likely) | ||

+ | * '''Novelty:''' (geo)-spatial time series prediction is an open area with great potential for theoretical and practical work. In particular, fire risk assessment is necessary for (1) predicting the probability of damage (agriculture, animal husbandry); (2) assessment of insurance and financial risks. | ||

+ | * '''Authors:''' Yuri Maksimov (point of contact), Alexey Zaitsev | ||

+ | * '''Consultants:''' Yuri Maksimov (point of contact), Alexey Zaitsev, Alexander Bulkin. | ||

+ | |||

+ | ===Problem 95.2022=== | ||

+ | * '''Title:''' Identification the transmission rate and time-dependent noise for the stochastic SIER disease model with vital rates (Time-dependent parameter identification for a stochastic epidemic model) | ||

+ | * '''Problem description:''' The problem is set to find the optimal time-dependent parameters for the known stochastic SIER disease propagation model. The optimal parameters are the parameters of the stochastic equation, under which the sample of the rate of spread of the virus in a limited population, when using comparison with the optimal sample. It is proposed to use the adaptive generalized method of moments with local delay (LLGMM) based on the generalized method of moments (GMM). | ||

+ | * '''Data:''' Hopkins Institution's Coronavirus Increasing Data is available from various sources. You can also download the data yourself from the link. | ||

+ | * '''Literature:''' | ||

+ | *# Anna Mummert, Olusegun M. Otunuga Parameter identification for a stochastic SEIRS epidemic model: case study influenza PDF | ||

+ | *# David M. Drukker Understanding the generalized method of moments (GMM): A simple example LINK | ||

+ | * '''Keywords:''' Compartment disease model, Stochastic disease model, Local lagged adapted generalized method of moments, Time-dependent transmission rate | ||

+ | * '''Base algorithm:''' there are several different options on the Internet, for example, the article B.Tseytlin Actually forecasting COVID-19 LINK, the current program does not give good convergence, because it always uses a fixed number of points for prediction | ||

+ | * '''Novelty:''' a new LLGMM method of moments that increases the accuracy of prediction& The basic idea of the method of moments is to use in moment conditions (moment functions or simply moments) instead of mathematical expectations, sample means, which, according to the law of large numbers under sufficiently weak conditions, should converges asymptotically to the mathematical expectations. Since the number of conditions for moments in the general case is greater than the number of estimated parameters, this system of conditions does not have a unique solution. The generalized method of moments suggests a situation where it is possible to obtain more conditions for moments than estimated parameters. The method constructs moment conditions (moment functions), also called orthogonality conditions, in a more general form as some function of model parameters and data. The parameters are estimated by minimizing a certain positive quadratic form from the sample means for the moments (moment functions). The quadratic form is in an iterative process with the required accuracy. If the model contains more than one parameter (this is our case) to be identified, then the second and higher moments are used to construct moment conditions. LLGMM defines time-dependent parameters by using a limited number of "points" in a data time series to form moment conditions, rather than the entire series. So the method is late. In addition, the number of time series elements used varies for each estimate over time. Thus, the method is local and adaptive. | ||

+ | * '''Author:''' expert Vera Markasheva (Laboratory of Computational Bioinformatics of the Center for Systems Biology) | ||

+ | |||

+ | ===Problem 96.2022=== | ||

+ | * '''Title:''' Impact of the lockdown on the dynamics of the epidemic | ||

+ | * '''Problem description:''' The introduction of a lockdown is considered an effective measure to combat the epidemic. However, contrary to intuition, it turned out that under certain conditions, a lockdown can lead to an increase in the epidemic. This effect is absent for classical models “on average”, but was revealed when modeling the spread of the epidemic, taking into account the contact graph. The problem is to find formulaic and quantitative relationships between the parameters under which the lockdown can lead to an increase in the epidemic. | ||

+ | * '''Data:''' Real data on the spread of the epidemic on contact graphs, especially considering the need for scenario analysis, is not available. The problem involves working with model and synthetic data: there are ready-made data, and it is also assumed that new ones can be generated in the process of solving the problem. | ||

+ | * '''Authors:''' Anton Bishuk, A.V. Zuhba | ||

+ | |||

+ | ===Problem 102.2022=== | ||

+ | * '''Title:''' Graph neural networks in the problem of regression of pairs of graphs | ||

+ | * '''Problem description:''' Considered The problem regression on a pair of graphs. In a pair, each vertex of one graph corresponds to a vertex of the second graph. It is required to establish the optimal architecture of the graph neural network, taking into account the given order specified on the vertices. | ||

+ | * '''Data:''' It is suggested to use chemical reaction datasets [https://github.com/hesther/reactiondatabase github]. For a given dataset, a pair of graphs is specified in a natural way. These are graphs of molecules of initial substances and products of a chemical reaction. | ||

+ | * '''Literature:''' | ||

+ | *# [https://chemrxiv.org/engage/chemrxiv/article-details/60c74e0f9abda2cf1af8d58a DRACON: disconnected graph neural network for atom mapping in chemical reactions.] | ||

+ | *# [https://chemrxiv.org/engage/api-gateway/chemrxiv/assets/orp/resource/item/6112ac487117507542e68bef/original/machine-learning-of-reaction-properties-via-learned-representations-of-the- condensed-graph-of-reaction.pdf Machine learning of reaction properties via learned representations of the condensed graph of reaction.] | ||

+ | *# [https://ieeexplore.ieee.org/abstract/document/9046288 A comprehensive survey on graph neural networks.] | ||

+ | * '''Base algorithm:''' The graph relationship is set at the level of graph embeddings. That is, a separate embedding vector is built for each graph, and then the vector data is concatenated. In this case, information about the correspondence of vertices in graphs is not explicitly used. | ||

+ | * '''Novelty:''' On the example of the architecture of a graph neural network with fixed hyperparameters, from a theoretical and practical point of view, to study ways to add information about the relationship of graphs to a graph neural network. | ||

+ | * '''Authors:''' Filipp Nikitin, Vadim Strijov V.V., Alexander Isaev. | ||

+ | |||

+ | ===Problem 103.2022=== | ||

+ | * '''Requirement:''' Fluent English to collaborate, Python and PyTorch (medium level and higher), Git, Bash, Background in computational biology is a plus | ||

+ | * '''Introduction:''' [http://www.machinelearning.ru/wiki/images/f/fa/M1p_ppis.pdf See full description here]. Proteins are involved in several biological reactions by means of interactions with other proteins or with other molecules such as nucleic acids, carbohydrates, and ligands. Among these interaction types, protein–protein interactions (PPIs) are considered to be one of the key factors as they are involved in most of the cellular processes [1]. The binding of two proteins can be viewed as a reversible and rapid process in an equilibrium that is governed by the law of mass action. Binding affinity is the strength of the interaction between two (or more than two) molecules that bind reversibly (interact). It is translated into physico-chemical terms in the dissociation constant Kd, the latter being the concentration of free protein at which half of all binding sites of the second protein type are occupied [2]. | ||

+ | * '''Objectives:''' Three main objectives of this work can be formulated as follows: 1. Refine PDBbind [12] data and a standard binding affinity dataset [3], and compile a novel benchmark of PPIs with known binding affinity values. 2. Employ graph-learning toolset to predict binding affinities of PPIs from the new dataset. 3. Benchmark the resulting method against existing state-of-the-art approaches | ||

+ | * '''Data & Metrics:''' In this work, we will operate on experimentally-observed three-dimensional structures of protein-protein complexes annotated with the binding affinity values. Two main sources of data are the following: | ||

+ | * PDBbind dataset [12] that includes around 2k PPIs | ||

+ | * Standard dataset introduced in [3] that includes 144 PPIs As main regression metrics, we suggest to consider Mean Squared Error (MSE), Mean Absolute Error (MAE) and Pearson correlation. | ||

+ | * '''Novelty:''' To the best of our knowledge, geometric deep learning methods have never been applied to the protein-protein binding affinity prediction problem so far. | ||

+ | * '''Authors:''' Arne Schneuing, Ilia Igashov | ||

+ | |||

+ | ===Problem 109.2022=== | ||

+ | * '''Title:''' Continuous time when building a BCI neural interface | ||

+ | * '''Problem description:''' In Signal Decoding The problems, data is represented as multivariate time series. When solving problems, a discrete representation is used time. However, recent work on neural ordinary differential equations illustrates the ability to work with the hidden state of recurrent neural networks, as with solutions to differential equations. This allows us to consider time series as continuous in time. | ||

+ | * '''Data:''' For classification: | ||

+ | *# dataset P300, according to which the article was written | ||

+ | *# dataset DEAPdataset similar to it in the format of records. | ||

+ | *# Definition of emotions. | ||

+ | *# Same SEED emotion classification | ||

+ | *# Not EEG, but accelerometer data with activity/position classification | ||

+ | *# For regression, you can take the same neurotycho, if you want to complicate life somewhat with respect to classification problems. | ||

+ | * '''Literature:''' | ||

+ | *# Neural Ordinary Differential Equations | ||

+ | *# Neural controlled differential equations for irregular time series | ||

+ | *# Latent ODEs for Irregularly-Sampled Time Series (?) | ||

+ | *# GRU-ODE-Bayes: Continuous modeling of sporadically-observed time series (?) | ||

+ | *# Neural Rough Differential Equations for Long Time Series (?) | ||

+ | *# ODE2VAE: Deep generative second order ODEs with Bayesian neural networks (?) | ||

+ | *# Go with the Flow: Adaptive Control for Neural ODEs | ||

+ | *# Legendre Memory Units: Continuous-Time Representation in Recurrent Neural Networks | ||

+ | *# My master's | ||

+ | * '''Base algorithm:''' Alina Samokhina's algorithm | ||

+ | * '''Solution:''' Using NeurODE variations to approximate the original signal. (Bayes, partial derivatives, etc.). Comparative analysis of existing approaches to the application of differential equations for EEG classification | ||

+ | * '''Novelty:''' suggests a way to construct a continuous signal representation. Working with the functional space of the signal, not its discrete representation. Using the parameters of the resulting function as a feature space of the resulting model. | ||

+ | * '''Authors:''' Alina Samokhina, Strijov V.V. | ||

+ | |||

+ | ===Problem 104.2022=== | ||

+ | * '''Title:''' (Clarification awaited) Cross-language duplicate search | ||

+ | * '''Problem description:''' The problem of cross-language search for text plagiarism is set. The search for duplicates of the original text is carried out among texts in 100 different languages. | ||

+ | * '''Data:''' | ||

+ | *# A selection of scientific articles from the scientific electronic library eLIBRARY.ru, as well as articles from the Wikipedia online encyclopedia, is used as a training sample. | ||

+ | *# The State Rubricator of Scientific and Technical Information (SRSTI), the Universal Decimal Classifier (UDC) are considered as scientific rubricators. | ||

+ | *# The following are used as search quality metrics: | ||

+ | *# average frequency - the frequency, averaged over the control languages, with which the query document falls into the top 10% of documents among which the search is carried out | ||

+ | *# average percentage - the percentage of documents, averaged over the control languages, that are in the top 10% of translation documents that have the same scientific heading as the query document | ||

+ | * '''Literature:''' Vorontsov K. V. Probabilistic thematic modeling: review of models and additive regularization [http://www.machinelearning.ru/wiki/images/d/d5/Voron17survey-artm.pdf PDF] | ||

+ | * '''Base algorithm:''' | ||

+ | *# Hierarchical topic models | ||

+ | *# Topic models with one-pass document vectorization | ||

+ | * '''Solution:''' To solve the search problem, a multimodal thematic model was built. 100 languages were used as modalities, as well as scientific headings, which included articles from the training data. A series of experiments was carried out to improve search quality metrics, including: selection of the optimal tokenization method, addition of regularizers, selection of thematic vector comparison functions, ranking functions, etc. | ||

+ | * '''Novelty:''' Most systems for finding documents in large collections are based on vectorization of the documents in the collection and the search document in one way or another. The latest ways to vectorize documents are usually limited to one language. In this case, the problem arises of creating a uniform system for obtaining vector embeddings of a multilingual collection of documents. The proposed approach makes it possible to train a topic model that encodes information about the distribution of words in a text, regardless of their language affiliation. Also, the solution is subject to restrictions on the size of the model and training time, due to the possibility of practical use of the described model. | ||

+ | * '''Author:''' Polina Potapova, Konstantin Vorontsov | ||

+ | |||

+ | ===Problem 52.2022=== | ||

+ | * '''Title:''' (pending clarification) Predicting the quality of protein models using spherical convolutions on 3D graphs. | ||

+ | * '''Problem description:''' The purpose of this work is to create and study a new convolution operation on three-dimensional graphs within the framework of solving the problem of assessing the quality of three-dimensional protein models (The problem regression on graph nodes). | ||

+ | * '''Data:''' [http://predictioncenter.org Models generated by CASP contestants] are used. | ||

+ | * '''Literature:''' | ||

+ | *# [https://drive.google.com/file/d/1pXCED8XBcxbjwtg_1wZG0oAjvUCxFlua/view?usp=sharing The problem details]. | ||

+ | *# [https://arxiv.org/abs/1806.01261 Relational inductive biases, deep learning, and graph networks]. | ||

+ | *# [https://arxiv.org/abs/1611.08097 Geometric deep learning: going beyond euclidean data]. | ||

+ | * '''Base algorithm:''' As a base algorithm, we will use a neural network based on the graph convolution method, which is generally described in [https://arxiv.org/abs/1806.01261]. | ||

+ | * '''Solution:''' The presence of a peptide chain in proteins allows you to uniquely enter local coordinate systems for all graph nodes, which makes it possible to create and apply spherical filters regardless of the graph topology. | ||

+ | * '''Novelty:''' In general, graphs are irregular structures, and in many Graph Learning The problems, sample objects do not have a single topology. Therefore, the existing operations of convolutions on graphs are greatly simplified or do not generalize to different topologies. In this paper, we propose to consider a new method for constructing a convolution operation on three-dimensional graphs, for which it is possible to uniquely choose local coordinate systems associated with each node. | ||

+ | * '''Author:''' Sergey Grudinin | ||

+ | |||

+ | ===Problem 110. 2022 (technical)=== | ||

+ | * '''Title:''' Detection of defects on the car body | ||

+ | * '''SubThe problems:''' Classification of cars by type and brand, Classification of car parts (door, hood, roof, etc.), Segmentation of defective areas on different parts of the car, Classification of defects by type (dent, scratch, glass damage), Assessment of the degree of damage, | ||

+ | * '''Data:''' | ||

+ | *# Coco Car Damage Detection Dataset - 70 photos of damaged cars with frames, semantic mask and damage type (headlight, front bumper, hood, door, rear bumper) | ||

+ | *# Сar_damage - 920 photos of damaged cars with labeled masks | ||

+ | *# CarDent-Detection-Assessment - 100 photos of damaged cars with labeled masks | ||

+ | *# CarAccidentDataset - 52 photos of damaged cars with labeled masks | ||

+ | *# Car damage detection - 950 photos of damaged and 1150 photos of whole cars | ||

+ | *# Car Damage - 1512 photos of damaged cars. Labeled to classify the type of damage | ||

+ | *# Cars Dataset - 16185 photos of whole cars, 196 models. Images with different angles, labels and frames of machine elements for matching angles. | ||

+ | * '''Author:''' Andrey Inyakin | ||

+ | |||

+ | ===Problem 111.2022 (technical)=== | ||

+ | * '''Title:''' Recognition of named entities in informational Russian-language news | ||

+ | * '''SubThe problems:''' Estimating the accuracy of available NER models (up to 2 weeks for data collection and markup) | ||

+ | * '''Base algorithm:''' Development of an algorithm for saturation (augmentation) of the training sample with rare named entities | ||

+ | * '''Data:''' To solve the problem, datasets of news from Interfax with the markup of named entities will be prepared. | ||

+ | |||

+ | ==2021== | ||

+ | {|class="wikitable" | ||

+ | |- | ||

+ | ! Author | ||

+ | ! Topic | ||

+ | ! Links | ||

+ | ! Consultant | ||

+ | ! Letters | ||

+ | ! Reviewer | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Magistrkoljan Grebenkova Olga] | ||

+ | |Variational optimization of deep learning models with model complexity control | ||

+ | |[https://docs.google.com/document/d/1gHyVeYgzFgco1vUTZRjxT2FbO03GsB27EVEstLWTzdM/edit?usp=sharing LinkReview] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020-Project60 GitHub] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020-Project60/raw/master/docs/Grebenkova2020Optimization.pdf Paper] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020-Project60/raw/master/slides/Grebenkova2020OptimizationSlides.pdf Slides] | ||

+ | [https://youtu.be/9ELhIqjFSE8 Video] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Oleg_Bakhteev Oleg Bakhteev] | ||

+ | |AILP+UXBR+HCV+TEDWSS | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Vshokorov Shokorov Vyacheslav] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020_Project_9/raw/master/review%20Grebenkova.pdf Review] | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Anton39reg Pilkevich Anton] | ||

+ | |Existence conditions for hidden feedback loops in recommender systems | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2021-Project-74 GitHub] | ||

+ | [https://docs.google.com/document/d/1OLCqkmArjqFn8M9pB5C_kLoYOv0l1w9RjHy0y0upPew/edit?usp=sharing LinkReview] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2021-Project-74/raw/main/docs/Pilkevich2021HiddenFeedbackLoops.pdf Paper] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2021-Project-74/raw/main/docs/Pilkevich2021Presentation/Pilkevich2021Presentation.pdf Slides] | ||

+ | [https://www.youtube.com/watch?v=xW_lXGn1WHs&t=24s Video] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Khritankov Khritankov Anton] | ||

+ | |AILB*P-X+R-B-H1CVO*T-EM*H1WJSF | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Gorpinich Gorpinich Maria] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2021-Project-84/raw/main/docs/Pilkevich2021HiddenFeedbackLoops_review.pdf Review] | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Antonina_Kurdyukova Antonina Kurdyukova] | ||

+ | |Determining the phase and disorder of human movement based on the signals of wearable devices | ||

+ | |[https://docs.google.com/document/d/1ts2i6Cq6CCFf3YWGPhtDxDlfj3OCoQGC3RcXou9bo1I/edit?usp=sharing LinkReview] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2021-Project77 GitHub] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2021-Project77/raw/main/docs/Kurdyukova2021WearableDevices.pdf Paper] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2021-Project77/raw/main/slides/Kurdyukova2021Presentation_ru.pdf Slides] | ||

+ | [https://www.youtube.com/watch?v=xW_lXGn1WHs&t=684s Video] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:KormakovG Georgy Kormakov] | ||

+ | |AILB*PXBRH1CVO*TEM*WJSF | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Anton39reg Pilkevich Anton] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2021-Project-74/raw/main/docs/review_Kurdyukova.pdf Review] | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Yakovlev_kd Yakovlev Konstantin] | ||

+ | |A differentiable search algorithm for model architecture with control over its complexity | ||

+ | |[https://docs.google.com/document/d/1cxWRiZ1a4JR83kYvxtXwpiOR-g8ar4_NaQ6E2ealEF0/edit LinkReview] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2021-Project85 GitHub] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2021-Project85/raw/main/docs/DARTS2021Yakovlev.pdf Paper] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2021-Project85/raw/main/slides/Yakovlev2021Presentation_ru.pdf Slides] | ||

+ | [https://www.youtube.com/watch?v=xW_lXGn1WHs&t=1157s Video] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Magistrkoljan Grebenkova Olga] | ||

+ | |AILB*PXBRH1CVO*TEM*WJSF | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Vitalii_kondratiuk Pyrau Vitaly] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2021-Project-Planning/raw/main/docs/Yakovlev2021DARTS_Review.pdf Review] | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Gorpinich Gorpinich Maria] | ||

+ | |Trajectory Regularization of Deep Learning Model Parameters Optimization Based on Knowledge Distillation | ||

+ | |[https://docs.google.com/document/d/1kQj66GEPv4Dx21A1_zJJKLRR1OujsLgkJrgKO5DCz70/edit?usp=sharing LinkReview] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2021-Project-84 GitHub] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2021-Project-84/raw/main/docs/Gorpinich2021DistillingKnowledge.pdf Paper] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2021-Project-84/raw/main/docs/slides/Gorpinich2021DistillingKnowledgeSlides.pdf Slides] | ||

+ | [https://www.youtube.com/watch?v=xW_lXGn1WHs&t=1625s Video] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Oleg_Bakhteev Oleg Bakhteev] | ||

+ | |AILB*P+XBRC+VH1O*TEM*WJSF | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Kulackov Kulakov Yaroslav] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2021-Project-17/raw/main/docs/GorpinichMaria2020PaperReview.pdf Review] | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Alexandr_Tolmachev Alexandr Tolmachev] | ||

+ | |Analysis of the QPFS Feature Selection Method for Generalized Linear Models | ||

+ | |[https://docs.google.com/document/d/1mtJc1ZqMSmPh9nRjdCZCV-zOSfDNp3Sejo3sx8mHw9Q/edit?usp=sharing LinkReview] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2021-Project-87 GitHub] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2021-Project-87/raw/main/docs/Tolmachev2021BayesApproach.pdf Paper] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2021-Project-87/raw/main/Slides/Tolmachev2021Presentation.pdf Slides] | ||

+ | [https://www.youtube.com/watch?v=xW_lXGn1WHs&t=2201s Video] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Aduenko Aduenko Alexander] | ||

+ | |AILB*PXB-R-H1CVO*TEM*WJSF | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Antonina_Kurdyukova Antonina Kurdyukova] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2021-Project77/raw/main/docs/Tolmachev2021BayesApproach_Review.pdf Review] | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Kulackov Kulakov Yaroslav] | ||

+ | |BCI: Selection of consistent models for building a neural interface | ||

+ | |[https://docs.google.com/document/d/1w28UOFRZgXhvt2MZqgdj682vGS9fjP6EUijrQqYoUPs/edit?usp=sharing LinkReview] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2021-Project-17 GitHub] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2021-Project-17/raw/main/docs/Kulakov2021BCI.pdf Paper] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2021-Project-17/raw/main/presentation/Kulakov2021Presentation.pdf Slides] | ||

+ | [https://www.youtube.com/watch?v=xW_lXGn1WHs&t=2850s Video] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Isachenkoroma Isachenko Roman] | ||

+ | |AILB*PXBRH1CVO*TEM*WJ0SF | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Zverev.eo Zverev Egor] | ||

+ | [https://raw.githubusercontent.com/Intelligent-Systems-Phystech/2021-Project-86/main/docs/PeerReviewForKulakov(RUS).pdf Review] | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Vitalii_kondratiuk Pyrau Vitaly] | ||

+ | |Experimental comparison of several problems of operational planning of biochemical production. | ||

+ | |[https://docs.google.com/document/d/115kv-KWPdX5R_UkEA8UlZV9opw-OmnevRM87R3xrn6k/edit?usp=sharing LinkReview] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2021-Project-Planning GitHub] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2021-Project-Planning/raw/main/docs/Pirau2021SchedulingInProcessIndustry.pdf Paper] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2021-Project-Planning/raw/main/slides/Pirau2021Presentation_ru.pdf Slides] | ||

+ | [https://www.youtube.com/watch?v=xW_lXGn1WHs&t=3491s Video] | ||

+ | |[https://mipt.ru/education/chairs/dm/staff/trenin.php Trenin Sergey Alekseevich] | ||

+ | |AILB*PXBRH1CVO*TEM*WJSF | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Yakovlev_kd Yakovlev Konstantin] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2021-Project85/raw/main/docs/Pirau2021_Scheduling_Review.pdf Review] | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Bazhenov.aa Bazhenov Andrey] | ||

+ | |Search for the boundaries of the iris by the method of circular projections | ||

+ | |[https://docs.google.com/document/d/1rmd1MQemJhgHG7W3p3qH2Di7KAxtClImVB_Gx_lOWCY/edit?usp=sharing LinkReview] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2021-Project88 GitHub] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2021-Project88/raw/master/docs/Bazhenov2021.pdf Paper] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2021-Project88/raw/master/slides/Bazhenov2021Presentation.pdf Slides] | ||

+ | [https://www.youtube.com/watch?v=xW_lXGn1WHs&t=4712s Video] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:IvanMatveev Matveev Ivan Alekseevich] | ||

+ | |AILB*PXB0RH1CVO*TEM*WJ0SF | ||

| | | | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Zverev.eo Zverev Egor] | ||

+ | |Learning co-evolution information with natural language processing for protein folding problem | ||

+ | |[https://docs.google.com/document/d/1x4TGjGlGjtr2m4hhzY3qGSFU7bZ6wv03eHwpTOVC3-8/edit?usp=sharing LinkReview] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2021-Project-86 GitHub] | ||

+ | [https://raw.githubusercontent.com/Intelligent-Systems-Phystech/2021-Project-86/main/docs/Zverev2021CoevolutionFromLMs.pdf Paper] [https://raw.githubusercontent.com/Intelligent-Systems-Phystech/2021-Project-86/main/docs/Zverev2021Presentation.pdf Slides] | ||

+ | [https://www.youtube.com/watch?v=xW_lXGn1WHs&t=4184s Video] | ||

+ | |[https://team.inria.fr/nano-d/team-members/sergei-grudinin/ Sergei Grudinin], |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Igashov Ilya Igashov] | ||

+ | |AILB*PXBRH1CVO*TEM*WJSF | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Alexandr_Tolmachev Alexandr Tolmachev] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2021-Project-87/raw/main/docs/Zverev2021Review.pdf Review] | ||

+ | |- | ||

+ | |[https://www.youtube.com/channel/UC1uHZnZPsxDpSYlerkdTvXA/videos Gorchakov Vyacheslav] | ||

+ | |Importance Sampling for Chance Constrained Optimization | ||

+ | |[https://docs.google.com/document/d/199SSt922JQRTBhj8USTVNfDi0bn_t3NnuCCQtDf3NwY/edit?usp=sharing LinkReview] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2021-Project-78 Github] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2021-Project-78/raw/main/docs/Gorchakov_Importance_Sampling_for_Chance_Constrained_Optimization.pdf Paper] | ||

+ | [https://www.youtube.com/watch?v=xW_lXGn1WHs&t=5441s Video] | ||

+ | |[https://faculty.skoltech.ru/people/yurymaximov Yuri Maksimov] | ||

+ | |AILB*PX0B0R0H1C0V0O*0T0E0M*0W0JS0F | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Bazhenov.aa Bazhenov Andrey] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2021-Project88/raw/master/docs/Gorchakov2021_Importance_Sampling_for_Chance_Constrained_Optimization_Review.pdf Review] | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:NikLin Lindemann Nikita] | ||

+ | |Training with an expert for a sample with many domains | ||

+ | |[https://docs.google.com/document/d/1wL99D7UyY2uJqHwvxTfTKX3REoauyub5L8bFnRnwpJU/edit?usp=sharing LinkReview] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2021-Project-82 Github] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2021-Project-82/raw/main/docs/Lindemann2021DomainAdaptation.pdf Paper] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2021-Project-82/raw/main/Slides/Lindemann2021PresentationDomainAdaptation.pdf Slides] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Andriygav Andrey Grabovoi] | ||

+ | |AILPXBRH1C0V0O*TE0M*0W0J0SF0 | ||

| | | | ||

|- | |- | ||

- | |9 | + | |} |

+ | |||

+ | ===Problem 74.2021=== | ||

+ | * '''Title:''' Existence conditions for hidden feedback loops in recommender systems | ||

+ | * '''Problem description:''' In recommender systems, the effect of artificially inadvertently limiting the user's choice due to the adaptation of the model to his preferences (echo chamber / filter bubble) is known. The effect is a special case of hidden feedback loops. (see - Analysis H.F.L.). It is expressed in the fact that by recommending the same objects of interest to the user, the algorithm maximizes the quality of its work. The problem is a) lack of variety b) saturation / volatility of the user's interests. | ||

+ | * '''Problem description:'''It is clear that the algorithm does not know the interests of the user and the user is not always honest in his choice. Under what conditions, what properties of the learning algorithm and dishonesty (deviation of the user's choice from his interests) will the indicated effect be observed? Clarification. The recommendation algorithm gives the user a_t objects to choose from. The user selects one of them c_t from Bernoulli from the model of interest mu(a_t) . Based on the user's choice, the algorithm changes its internal state w_t and gives the next set of objects to the user. On an infinite horizon, you need to maximize the total reward sum c_t. Find the conditions for the existence of an unlimited growth of user interest in the proposed objects in a recommender system with the Thomson Sampling (TS) MAB algorithm under conditions of noisy user choice c_t. Without noise, it is known that there is always unlimited growth (in the model) [1]. | ||

+ | * '''Data:''' are created as part of the experiment (simulation model) by analogy with the article [1], external data is not required. | ||

+ | * '''References:''' | ||

+ | *# Jiang, R., Chiappa, S., Lattimore, T., György, A. and Kohli, P., 2019, January. Degenerate feedback loops in recommender systems. In Proceedings of the 2019 AAAI/ACM Conference on AI, Ethics, and Society (pp. 383-390). | ||

+ | *# Khritankov, A. (2021). Hidden Feedback Loops in Machine Learning Systems: A Simulation Model and Preliminary Results. In International Conference on Software Quality (pp. 54-65). Springer, Cham. | ||

+ | *# Khritankov A. (2021). Hidden feedback loop experiment demo. https://github.com/prog-autom/hidden-demo | ||

+ | * '''Base algorithm:''' The initial mathematical model of the phenomenon under study is described in the article [1]. The method of experimental research is in the article [2]. The base source code is available at [3] | ||

+ | * '''Solution:''' It is necessary to derive conditions for the existence of positive feedback for the Thomson Sampling Multi-armed Bandit algorithm based on the known theoretical properties of this algorithm. Then check their performance in the simulation model. For verification, a series of experiments is performed with the study of parameter ranges and the estimation of the error (variance) of the simulation. The results are compared with the previously constructed mathematical model of the effect. There is an implementation of the experiment system that can be improved for this The problem. | ||

+ | * '''Novelty:''' The studied positive feedback effect is observed in real and model systems and is described in many publications as an undesirable phenomenon. There is his model for the limited case of the absence of noise in the user's actions, which is not implemented in practice. Under the proposed conditions, The problem has not previously been posed and not solved for recommender systems. For the regression problem, the solution is known. | ||

+ | * '''Authors:''' Expert, consultant Anton Khritankov | ||

+ | |||

+ | ===Problem 77.2021=== | ||

+ | * '''Title:''' Determining the phase and disorder of human movement by signals from wearable devices | ||

+ | * '''Problem description:''' A wide class of periodic movements of a person or an animal is investigated. It is required to find the beginning and end of the movement. It is required to understand when one type of movement ends and another begins. For this, The problem of segmentation of time series is solved. The phase trajectory of one movement is constructed and its actual dimension is found. The purpose of the work is to describe a method for finding the minimum dimension of the phase space. By repetition of the phase, segment the periodic actions of a person. It is also necessary to propose a method for extracting the zero phase in a given space for a specific action. Bonus: find the discord in the phase trajectory and indicate the change in the type of movement. Bonus 2: do this for different phone positions by proposing invariant transformation models. | ||

+ | * '''Data:''' The data consists of time series read from a three-axis accelerometer with an explicit periodic class (walking, running, walking up and down stairs, etc.). It is possible to get your own data from a mobile device, or get model data from the dataset [https://archive.ics.uci.edu/ml/datasets/Human+Activity+Recognition+Using+Smartphones UCI HAR] | ||

+ | * '''References:''' | ||

+ | *# A. P. Motrenko, V. V. Strijov. Extracting fundamental periods to segment biomedical signals // Journal of Biomedical and Health Informatics, 2015, 20(6).P. 1466–1476. Time series segmentation with periodic actions: The segmentation problem was solved using a fixed-dimensional phase space. [http://strijov.com/papers/MotrenkoStrijov2014RV2.pdf PDF][http://sourceforge.net/p/mlalgorithms/code/HEAD/tree/Group874/Motrenko2014TSsegmentation/JBHI/MotrenkoStrijov2014RV2.pdf?format=raw URL] | ||

+ | *# A.D. Ignatov, V. V. Strijov. Human activity recognition using quasi-periodic time series collected from a single triaxial accelerometer. // Multimedia Tools and Applications, 2015, P. 1–14. Classification of human activity using time series segmentation: classifiers were studied on the resulting segments. [https://rdcu.be/6oBD PDF][http://strijov.com/papers/Ignatov2015HumanActivity.pdf URL] | ||

+ | *# Grabovoy, A.V., Strijov, V.V. Quasi-Periodic Time Series Clustering for Human Activity Recognition. Lobachevskii J Math 41, 333–339 (2020). Segmentation of time series into quasi-periodic segments: Segmentation methods were explored using principal component analysis and transition to phase space. [http://www.machinelearning.ru/wiki/images/c/cd/Grabovoy2019BSThesis.pdf Text] [http://www.machinelearning.ru/wiki/images/1/19/Grabovoy2019TimeSeriesClusteringSlides.pdf Slides] [https://doi.org/10.1134/S19950802200300751 DOI] | ||

+ | * '''Base algorithm:''' The basic algorithm is described in 1 and 3 works, [https://sourceforge.net/p/mlalgorithms/code/HEAD/tree/Group874/Motrenko2014TSsegmentation/ code here], work code 3 author. | ||

+ | * '''Solution:''' It is proposed to consider various dimensionality reduction algorithms and compare different spaces in which the phase trajectory is constructed. Develop an algorithm for finding the minimum dimension of the phase space in which the phase trajectory has no self-intersections up to the standard deviation of the reconstructed trajectory. | ||

+ | * '''Novelty:''' In Motrenko's article, the space dimension is equal to two. This shortcoming must be corrected. The phase trajectory must not intersect itself. And if we can distinguish one type of movement from another within one period (switched from running to a step and realized this within one and a half steps), it will be great. | ||

+ | * '''Authors:''' | ||

+ | consultants: Kormakov G.V., Tikhonov D.M., Expert Strijov V.V. | ||

+ | |||

+ | ===Problem 78. 2021=== | ||

+ | * '''Title:''' Importance Sampling for Scenario Approximation of Chance Constrained Optimization | ||

+ | * '''Problem description:''' Optimization problems with probabilistic constraints are often encountered in engineering practice. For example, The problem of minimizing energy generation in energy networks, with (randomly fluctuating) renewable energy sources. In this case, it is necessary to comply with safety restrictions: voltages at generators and consumers, as well as currents on the lines, must be less than certain thresholds. However, even in the simplest situations, The problem cannot be resolved exactly. The best-known approach is the chance constrained optimization methods, which often give a good approximation. An alternative approach is sampling the network operation modes and solving the problem on the data set of the classification * '''Problem description:''' separating bad modes from good ones with a given error of the second kind. At the same time, for a sufficiently accurate solution, a very large amount of data is required, which often makes the problem numerically inefficient. We suggest using “importance sampling” to reduce the number of scenarios. Importance sampling consists of substituting a sample from a nominal solution, which often carries no information since all bad events are very rare, with a synthetic distribution that samples the sample in a neighborhood of bad events. | ||

+ | * '''Problem statement:''' find the minimum of a convex function (price) under probabilistic constraints (the probability of exceeding a certain threshold for a system of linear/quadratic functions is small) and numerically show the effectiveness of sampling in this problem. | ||

+ | * '''Data:''' Data is available in the pypower and matpower packages as csv files. | ||

+ | * '''References:''' The proposed algorithms are based on 3 articles: | ||

+ | *# Owen, Maximov, Chertkov. Importance Sampling for the Union of Rare Events with Applications to Power Systems [https://statistics.sites.stanford.edu/sites/g/files/sbiybj6031/f/2017-10.pdf LINK] | ||

+ | *# A. Nemirovski. On safe tractable approximations of chance constraints [https://www2.isye.gatech.edu/~nemirovs/EUROXXIV.pdf LINK] | ||

+ | *# S. Tong, A. Subramanyam, and Vi. Rao. Optimization under rare chance constraints. [https://arxiv.org/pdf/2011.06052.pdf LINK] | ||

+ | *# In addition, the authors of the problem have a draft of the article, in which you need to add a numerical part. | ||

+ | * '''Base algorithm:''' A list of basic algorithms is provided in this lecture [http://niaohe.ise.illinois.edu/IE598_2020/IE598NH-lecture-10-11-CCP.pdf LINK] | ||

+ | * '''Solution:''' in numerical experiments, you need to compare the sample size requirements for standard methods (scenario approximation) and using importance sampling to obtain a solution of comparable quality (and inverse The problem, having equal sample lengths, compare the quality of the solution) | ||

+ | * '''Novelty:''' The problem has long been known in the community and scenario approximation is one of the main methods. At the same time, importance sampling helps to significantly reduce the number of scenarios. We have recently received a number of interesting results on how to calculate optimal samplers, with their use the complexity of the problem will be significantly reduced | ||

+ | * '''Authors:''' Expert Yuri Maksimov, consultant Yuri Maksimov and Alexander Lukashevich. | ||

+ | |||

+ | ===Problem 79.2021=== | ||

+ | * '''Title:''' Improving Bayesian Inference in Physics Informed Machine Learning | ||

+ | * '''Problem description:''' Machine learning methods are currently widely used in physics, in particular, in solving turbulence problems or analyzing the stability of physical networks. At the same time, the key issue is which modes to choose for training models. A frequent choice is a sequence of points that uniformly covers the admissible set. However, often such sequences are not very informative, especially if analytical methods give a region where the system is guaranteed to be stable. The problem proposes several methods of sampling: allowing to take into account this information. Our goal is to compare them and find the one that requires the smallest sample size (empirical comparison). | ||

+ | * '''Data:''' The experiment is proposed to be carried out on model and real data. The simulation experiment consists in analyzing the stability of (slightly non-linear) differential equations (synthetic data is self-generated). The second experiment is to analyze the stability of energy systems (data from matpower, pypower, GridDyn). | ||

+ | * '''References:''' | ||

+ | *# Art Owen. Quasi Monte Carlo Sampling. [https://statweb.stanford.edu/~owen/courses/362-1011/readings/siggraph03.pdf LINK ] | ||

+ | *# Jian Cheng & Marek J. Druzdzel. Computational Investigation of Low-Discrepancy Sequences in Simulation Algorithms for Bayesian Networks [https://arxiv.org/pdf/1301.3841.pdf LINK] | ||

+ | *# A. Owen, Y Maximov, M. Chertkov. Importance Sampling for the Union of Rare Events with Applications to Power Systems [https://statistics.sites.stanford.edu/sites/g/files/sbiybj6031/f/2017-10.pdf LINK] | ||

+ | *# Polson and Solokov. Deep Learning: A Bayesian Perspective [https://arxiv.org/pdf/1706.00473.pdf LINK] | ||

+ | *# In addition: the authors of the problem have a draft work on this topic | ||

+ | * '''Base algorithm:''' The basic algorithm we are improving is Quasi Monte Carlo (QMC, [https://statweb.stanford.edu/~owen/courses/362-1011/readings/siggraph03.pdf LINK ]). The problem to construct low discrepancy sequences not covering the polyhedral region and the region given by the intersection of the quadratic constraints. Another algorithm with which we need a comparison: E. Gryazina, B. Polyak. Random Sampling: a Billiard Walk Algorithm [https://www.sciencedirect.com/science/article/pii/S1474667016425711 LINK] and algorithms Hit and Run [https://statweb.stanford.edu/~cgates/PERSI/papers/hitandrun062207.pdf LINK] | ||

+ | * '''Solution:''' sampling methods by importance, in particular the extension of the approach (Boy, Ryi, 2014) and (Owen, Maximov, Chertkov, 2017) and their applications to ML/DL for physical problems | ||

+ | * '''Novelty:''' in a significant reduction in sample complexity and the explicit use of existing and analytical results and learning to solve physical problems, before that ML approaches and analytical solutions were mostly parallel courses | ||

+ | * '''Authors:''' Expert Yuri Maksimov, consultant Yuri Maksimov and Alexander Lukashevich, student. | ||

+ | |||

+ | ===Problem 81.2021=== | ||

+ | * '''Title:''' NAS — Generation and selection of neural network architectures | ||

+ | * '''Problem description:''' The problem of choosing the optimal neural network architecture is set as The problem of sampling the vector of structural parameters. The optimality criterion is defined in terms of the accuracy, complexity and stability of the model. The sampling procedure itself consists of two steps: generating a new structure and rejecting this structure if it does not satisfy the optimality criterion. It is proposed to explore various methods of sampling. The formulation of the problem of choosing the optimal structure is described in [https://drive.google.com/file/d/1Wn-CEhDKvjyZMvZdBHWUobxpizVF1G8l/view?usp=sharing Potanin-1] | ||

+ | * '''Data:''' : Two separate sets are offered as data. The first one consists of one element, this is the popular MNIST dataset. Pros - is a strong and generally accepted baseline, was used as a benchmark for the WANN article, quite large (multi-class classification). The second set is a set of datasets for the regression The problem. Size varies from very small to quite large. Here is a link to the dataset and laptop to download the data [https://drive.google.com/file/d/19Cxtf3dg7gHFHyDXYAI0cEoT7PaNl4IR/view?usp=sharing data]. | ||

+ | * '''References:''' | ||

+ | *# [https://drive.google.com/file/d/1Wn-CEhDKvjyZMvZdBHWUobxpizVF1G8l/view?usp=sharing Potanin - 1] | ||

+ | *# Potanin - 2. One more work, the text is given to the interested student, but without publication. | ||

+ | *# Strijov Factory laboratory [http://strijov.com/papers/Strijov2012ErrorFn.pdf Error function] | ||

+ | *# [http://strijov.com/papers/HyperOptimizationEng.pdf Informtica] | ||

+ | *# [https://weightagnostic.github.io/ WANN] | ||

+ | *# [https://arxiv.org/pdf/1806.09055.pdf DARTS] | ||

+ | *# [https://arxiv.org/pdf/1912.01412.pdf Symbols] | ||

+ | *# [http://nn.cs.utexas.edu/downloads/papers/stanley.cec02.pdf NEAT] | ||

+ | * '''Base algorithm:''' Closest [https://weightagnostic.github.io/ project], and its [https://github.com/google/brain-tokyo-workshop/tree/master/WANNRelease/WANN code]. Actual [https://drive.google.com/file/d/19Cxtf3dg7gHFHyDXYAI0cEoT7PaNl4IR/view?usp=sharing code] from consultant. | ||

+ | * '''Solution:''' A number of experiments have already been performed, where sampling is performed by a genetic algorithm. Acceptable results have been obtained. It is proposed to analyze and improve them. Namely, to distinguish two modules: generation and deviation and compare several types of sampling. Basic - Importance sampling, desirable - Metropolis-Hastings (or even Metropolis-Langevin) sampling. Since the genetic algorithm is considered by us as a process with jumps, it is proposed to take this into account when designing the sampling procedure. The bonus of MH is that it has a Bayesian interpretation. The first level of Bayesian inference as applied to MH is described in [Informatica]. It is required either to rewrite it in terms of the distribution of structural parameters, or to describe both levels in general, moving the structural parameters to the second level (by the way, approximately the same will be in the Aduenko problem). | ||

+ | * '''Novelty:''' Neural networks excel at The problems of computer vision, reinforcement learning, and natural language processing. One of the main goals of neural networks is to perform well The problems that are currently solved exclusively by humans, that is, natural human neural networks. Artificial neural networks still work very differently from natural neural networks. One of the main differences is that natural neural networks evolve over time, changing the strength of connections and their architecture. Artificial neural networks can adjust the strength of connections using weights, but cannot change their architecture. Therefore, The problem of choosing the optimal structures of neural networks for specific The problems seems to be an important step in the development of the capabilities of neural network models. | ||

+ | * '''Authors:''' consultant Mark Potanin, Expert Strijov V.V. | ||

+ | |||

+ | ===Problem 82.2021=== | ||

+ | * '''Title:''' Training with an Expert for a sample with many domains. | ||

+ | * '''Problem description:''' The problem of approximating a multi-domain sample by a single multi-model - a mixture of Experts is considered. As data, it is supposed to use a sample that contains several domains. There is no domain label for each object. Each domain is approximated by a local model. The paper considers a two-stage The problem optimization based on the EM algorithm. | ||

+ | * '''Data:''' Samples of reviews from the Amazon site for different types of goods are used as data. It is supposed to use a linear model as a local model, and use tf-idf vectors within each domain as an indicative description of reviews. | ||

+ | * '''References:''' | ||

+ | *# [https://arxiv.org/pdf/1806.00258.pdf https://arxiv.org/pdf/1806.00258.pdf] | ||

+ | *# [http://www.mysmu.edu/faculty/jingjiang/papers/da_survey.pdf http://www.mysmu.edu/faculty/jingjiang/papers/da_survey.pdf] | ||

+ | *# [https://dl.acm.org/doi/pdf/10.1145/3400066 https://dl.acm.org/doi/pdf/10.1145/3400066] | ||

+ | * '''Basic algorithm and Solution:''' The basic solution is presented [https://www.aclweb.org/anthology/D18-1498.pdf here]. The work uses the expert mixture method for the Multi-Soruce domain adaptation problem. The code for the article is available [https://github.com/jiangfeng1124/transfer link]. | ||

+ | * '''Novelty:''' At the moment, in machine learning there are more and more The problems related to data that are taken from different sources. In this case, there are samples that consist of a large number of domains. At the moment, there is no complete theoretical justification for constructing mixtures of local models for approximating such types of samples. | ||

+ | * '''Authors:''' Grabovoi A.V., Strijov V.V. | ||

+ | |||

+ | ===Problem 17.2021=== | ||

+ | * '''Title:''' BCI: Selection of consistent models for building a neural interface | ||

+ | * '''Problem:''' When building brain-computer interface systems, simple, stable models are used. An important step in building an interface is such a model is an adequate choice of model. A wide range of models is considered: linear, simple neural networks, recurrent networks, transformers. The peculiarity of the problem is that when making a prediction, it is required to model not only the initial signal taken from the cerebral cortex, but also the target signal taken from the limbs. Thus, two models are required. In order for them to work together, a space of agreements is being built. It is proposed to explore the properties of this space and the properties of the resulting forecast (neural interface) on various pairs of models. | ||

+ | * '''Data:''' ECoG/EEG brain signal data sets. | ||

+ | *# Need ECoG (dataset 25 contains EEG, EOG and hand movements) [http://bnci-horizon-2020.eu/database/data-sets http://bnci-horizon-2020.eu/database/data-sets] | ||

+ | *# neyrotycho — our old data. | ||

+ | * '''References:''' | ||

+ | *# Yaushev F.Yu., Isachenko R.V., Strijov V.V. Latent space matching models in the forecasting problem // Systems and Means of Informatics, 2021, 31(1). [http://strijov.com/papers/Isachenko2020CanonicCorrelation.pdf PDF] | ||

+ | *# Isachenko R.V. Choice of a signal decoding model in high-dimensional spaces. Manuscript, 2021. [https://github.com/r-isachenko/PhDThesis/raw/master/doc/Isachenko2021PhDThesis.pdf PDF] | ||

+ | *# Isachenko R.V. Choice of a signal decoding model in high-dimensional spaces. Slides, 2020. [https://github.com/r-isachenko/PhDThesis/raw/master/pres/Isachenko2020PhDThesisPres.pdf] | ||

+ | *# Isachenko R.V., Vladimirova M.R., Strijov V.V. Dimensionality reduction for time series decoding and forecasting problems // DEStech Transactions on Computer Science and Engineering, 2018, 27349 : 286-296. [http://strijov.com/papers/IsachenkoVladimirova2018PLS.pdf PDF] | ||

+ | *# Isachenko R.V., Strijov V.V. Quadratic Programming Optimization with Feature Selection for Non-linear Models // Lobachevskii Journal of Mathematics, 2018, 39(9) : 1179-1187. [https://rdcu.be/bfR32 PDF] | ||

+ | *# Motrenko A.P., Strijov V.V. Multi-way feature selection for ECoG-based brain-computer interface // Expert Systems with Applications, 2018, 114(30) : 402-413. [http://strijov.com/papers/MotrenkoStrijov2017ECoG_HL_2.pdf PDF] | ||

+ | *# Eliseyev A., Aksenova T. Stable and artifact-resistant decoding of 3D hand trajectories from ECoG signals using the generalized additive model //Journal of neural engineering. – 2014. | ||

+ | * '''Basic algorithm''': Described in the first work. The code is available. In that work, the data is two parts of an image. In our work, the signal of the brain and the movement of the hands. Super* '''Problem description:''' to finish the first job. Also the code and works [http://www.machinelearning.ru/wiki/index.php?title=BCI here]. | ||

+ | * '''Solution:''' The case is considered when the initial data are heterogeneous: the spaces of the independent and target variables are of different nature. It is required to build a predictive model that would take into account the dependence in the source space of the independent variable, as well as in the space of the target variable. It is proposed to investigate the accuracy, complexity and stability of pairs of various models. Since the inverse The problem is solved when building a forecast, it is required to build inverse transformations for each model. To do this, you can use both basic techniques (PLS) and streams. | ||

+ | * '''Novelty:''' Analysis of the prediction and latent space obtained by a pair of heterogeneous models. | ||

+ | * '''Authors:''' Consultant Roman Isachenko, Expert Strijov V.V. | ||

+ | |||

+ | ===Problem 69.2021=== | ||

+ | * '''Title:''' Graph Neural Network in Reaction Yield prediction | ||

+ | * '''Problem description:''' There are disconnected graphs of source molecules and products in a chemical reaction. The yield of the main product in the reaction is known. It is required to design an algorithm that predicts yield by solving the regression The problem on given disconnected graphs. | ||

+ | * '''Data:''' Database of reaction from US patents [https://www.repository.cam.ac.uk/handle/1810/244727] | ||

+ | * '''References:''' | ||

+ | *# [https://www.ncbi.nlm.nih.gov/pubmed/30046072] A general overview. | ||

+ | *# [https://pure.uva.nl/ws/files/33146507/1703.06103.pdf] Relational Graph Convolution Neural Network | ||

+ | *# [https://papers.nips.cc/paper/7181-attention-is-all-you-need] Transformer architecture | ||

+ | *# [http://www.machinelearning.ru/wiki/images/6/6c/NikitinMMPR201927.pdf] Graph neural network learning for chemical compounds synthesis | ||

+ | * '''Base algorithm:''' Transformer model. The input sequence is a SMILES representation of the source and product molecules. | ||

+ | * '''Solution:''' A pipeline for working with disconnected graphs is proposed. The pipeline includes the construction of extended graph with molecule and reaction representation, Relational Graph Convolution Neural Network, Encoder of Transformer. The method is applied to solve yield predictions. | ||

+ | * '''Novelty:''' A solution for regression problem on the given disconnected graph is constructed; the approach demonstrates better performance compared with other solutions | ||

+ | * '''Authors:''' Nikitin Filipp, Isayev Olexandr, Strijov V.V. | ||

+ | |||

+ | ===Problem 84.2021=== | ||

+ | * '''Title:''' Trajectory Regularization of Deep Learning Model Parameters Optimization Based on Knowledge Distillation | ||

+ | * '''Problem description:''' The problem of optimizing the parameters of a deep learning model is considered. The case is considered when the responses of a more complex model (teacher model) are available during optimization. The classical approach to solving such a problem is learning based on the responses of a complex model (knowledge distillation). Assignment of hyperparameters is made empirically based on the results of the model on delayed sampling. In this paper, we propose to consider a modification of the approach to knowledge distillation, in which the coefficient of significance of the distilling term, as well as its gradients, act as hyperparameters. Both of these groups of parameters allow you to adjust the optimization of the model parameters. To optimize hyperparameters, it is proposed to consider the optimization problem as a two-level optimization problem, where at the first level of optimization The problem of optimizing the model parameters is solved, and at the second level The problem of optimizing hyperparameters is approximately solved by the value of the loss function on the delayed sample. | ||

+ | * '''Data:''' Sampling of CIFAR-10 images | ||

+ | * '''References:''' | ||

+ | *# [https://arxiv.org/abs/1503.02531 Distillation of knowledge] | ||

+ | *# [https://arxiv.org/abs/1511.06727 Hyperparameter Optimization in a Bilevel * '''Problem description:''' Greedy Method] | ||

+ | *# [http://strijov.com/papers/Bakhteev2017Hypergrad.pdf Hyperparameter Optimization in a Bilevel * '''Problem description:''' Comparison of Approaches] | ||

+ | *# [https://arxiv.org/abs/1606.04474 Meta Optimization: neural network instead of optimization operator] | ||

+ | * '''Basic algorithm: Model optimization without distillation and with standard distillation approach | ||

+ | * '''Solution:''' Using a two-level problem for model optimization. The combination of gradients for both terms is processed by a separate model (LSTM) | ||

+ | * '''Novelty:''' A new approach to model distillation will be proposed to significantly improve the performance of models trained in privileged information mode. It is also planned to study the dynamics of changes in hyperparameters in the optimization process. | ||

+ | * '''Authors:''' Oleg Bakhteev, Strijov V.V. | ||

+ | |||

+ | ===Problem 85.2021=== | ||

+ | * '''Title:''' A differentiable search algorithm for model architecture with control over its complexity | ||

+ | * '''Problem description:''' The problem of choosing the structure of a deep learning model with a predetermined complexity is considered. It is required to propose a method for searching for a model that allows controlling its complexity with low computational costs. | ||

+ | * '''Data:''' MNIST, CIFAR | ||

+ | * '''References:''' | ||

+ | *# Grebenkova O.S., Oleg Bakhteev, Strijov V.V.Variational optimization of a deep learning model with complexity control // Informatics and its applications, 2021, 15(2). [http://strijov.com/papers/Grebenkova2020HyperNet.pdf PDF] | ||

+ | *# [https://arxiv.org/abs/1806.09055 DARTS] | ||

+ | *# [https://arxiv.org/abs/1609.09106 hypernets] | ||

+ | * '''Basic algorithm: DARTS | ||

+ | * '''Solution:''' The proposed method is to use a differentiable neural network architecture search algorithm (DARTS) with parameter complexity control using a hypernet. | ||

+ | * '''Novelty:''' The proposed method allows you to control the complexity of the model, in the process of searching for an architecture without additional heuristics. | ||

+ | * '''Authors:''' Oleg Bakhteev, Grebenkova O. S. | ||

+ | |||

+ | ===Problem 86. 2021=== | ||

+ | * '''Title:''' Learning co-evolution information with natural language processing for protein folding problem | ||

+ | * '''Problem:''' One of the most essential problems in structural bioinformatics is protein fold recognition since the relationship between the protein amino acid sequence and its tertiary structure is revealed by protein folding. A specific protein fold describes the distinctive arrangement of secondary structure elements in the nearly-infinite conformation space, which denotes the structural characteristics of a protein molecule. | ||

+ | * '''Problem description:''': request | ||

+ | * '''Authors:''' Sergei Grudinin, Maria Kadukova. | ||

+ | |||

+ | ===Problem 87.2021=== | ||

+ | * '''Title:''' Bayesian choice of structures of generalized linear models | ||

+ | * '''Problem description:''' The work is devoted to testing methods for feature selection. It is assumed that the sample under study contains a significant number of multicollinear features. Multicollinearity is a strong correlation between the features selected for analysis that jointly affect the target vector, which makes it difficult to estimate regression parameters and identify the relationship between features and the target vector. There is a set of time series containing the readings of various sensors that reflect the state of the device. The readings of the sensors correlate with each other. It is necessary to choose the optimal set of features for solving the forecasting problem. | ||

+ | * '''Novelty:''' One of the most preferred feature selection algorithms has been published. It uses structural parameters. But there is no theoretical justification. It is proposed to build a theory by describing and analyzing various functions of a priori distribution of structural parameters. In works on the search for structures of neural networks, there is also no clear theory and a list of a priori assumptions. | ||

+ | * '''Data:''' Multivariate time series with readings from various sensors from paper 4, for starters, all samples from paper 1. | ||

+ | * '''References:''' Keywords: bootstrap aggregation, Belsley method, vector autoregression. | ||

+ | *# Katrutsa A.M., Strijov V.V. Comprehensive study of feature selection methods to solve multicollinearity problem according to evaluation criteria // Expert Systems with Applications, 2017, 76 : 1-11. [http://strijov.com/papers/Katrutsa2016QPFeatureSelection.pdf PDF] | ||

+ | *# Katrutsa A.M., Strijov V.V. Stresstest procedure for feature selection algorithms // Chemometrics and Intelligent Laboratory Systems, 2015, 142 : 172-183. [http://strijov.com/papers/Katrutsa2014TestGenerationEn.pdf PDF] | ||

+ | *# Strijov V.V. Error function in regression recovery problems // Factory laboratory. material diagnostics, 2013, 79(5) : 65-73. [http://strijov.com/papers/Strijov2012ErrorFn.pdf PDF] | ||

+ | *# Zaitsev A.A., Strijov V.V., Tokmakova A.A. Estimation of hyperparameters of regression models by the maximum likelihood method // Information technologies, 2013, 2 : 11-15. [http://strijov.com/papers/ZaytsevStrijovTokmakova2012Likelihood_Preprint.pdf PDF] | ||

+ | *# Kuznetsov M.P., Tokmakova A.A., Strijov V.V. Analytic and stochastic methods of structure parameter estimation // Informatica, 2016, 27(3) : 607-624. [http://strijov.com/papers/HyperOptimizationEng.pdf PDF] | ||

+ | *# Katrutsa A.M., Strijov V.V. The problem of multicollinearity in the selection of features in regression problems // Information technologies, 2015, 1 : 8-18. [http://strijov.com/papers/Katrutsa2014TestGeneration.pdf PDF] | ||

+ | *# Neichev Р.Г., Katrutsa A.M., Strijov V.V. Selection of the optimal set of features from a multicorrelated set in the forecasting problem. Zavodskaya Lab. material diagnostics, 2016, 82(3) : 68-74. [http://strijov.com/papers/Neychev2015FeatureSelection.pdf PDF] | ||

+ | * '''Base algorithm:''' Described in Reference 1: Quadratic Programming for QPFS Feature Selection. Code from Roman Isachenko. | ||

+ | * '''Solution:''' It is proposed to consider the structural parameters used in QPFS at the second level of Bayesian inference. Introduce informative a priori distributions of parameters and structural parameters. Compare different a priori assumptions. | ||

+ | * '''Novelty:''' Statistical Analysis of Structural Parameter Space and Visualization | ||

+ | * '''Authors:''' Alexander Aduenko consultant, Strijov V.V. | ||

+ | |||

+ | ===Problem 88.2021=== | ||

+ | *'''Name:''' Search for the boundaries of the iris by the method of circular projections | ||

+ | *'''Problem:''' Given a monochrome bitmap of the eye, [http://www.machinelearning.ru/wiki/images/1/16/Matveev2021project.pdf examples]. The approximate position of the center of the pupil is also known. The word "approximate" means that the calculated center of the pupil is no more than half of its true radius from the true one. It is necessary to determine the approximate positions of the circles approximating the pupil and iris. The algorithm must be very fast. | ||

+ | *'''Data:''' About 200 thousand eye images. For each, the position of the true circles is marked - for the purpose of training and testing the method being created. | ||

+ | *'''Basic algorithm:''' To speed up work with the image, it is proposed to aggregate data using circular projections of brightness. Circular projection is a function that depends on the radius, the value of which P(r) is equal to the integral of the directed image brightness gradient over a circle of radius r (or along an arc of a circle). Example for one arc (right quadrant) and for four arcs. Having built some circular projections, based on them, you can try to determine the position of the inner and outer borders of the iris (ring) using heuristics and / or a neural network. It is interesting to evaluate the capabilities of the neural network in this The problem. | ||

+ | *'''References:''' Matveev I.A. Detection of Iris in Image By Interrelated Maxima of Brightness Gradient Projections // Applied and Computational Mathematics. 2010. V.9. N.2. P.252-257 [https://www.researchgate.net/publication/228396639_Detection_of_iris_in_image_by_interrelated_maxima_of_brightness_gradient_projections PDF] | ||

+ | *'''Author:''' Matveev I.A. | ||

+ | |||

+ | ===Problem 53.2021=== | ||

+ | * '''Title:''' Solution of an optimization problem combining classification and regression to estimate the binding energy of a protein and small molecules. | ||

+ | * '''Problem description:''' The goal of the problem is to solve an optimization problem with classification and regression loss functions applied to biological data. | ||

+ | * '''Data:''' Approximately 12,000 complexes of proteins with small molecules. For classification, for each of them there is 1 correct position in space and 18 incorrect ones generated, for regression, each complex corresponds to the value of the binding constant (proportional to energy). The main descriptors are histograms of distributions of distances between different atoms. | ||

+ | * '''References:''' | ||

+ | *# https://www.overleaf.com/read/rjdnyyxpdkyj The problem details | ||

+ | *# http://cs229.stanford.edu/notes/cs229-notes3.pdf SVM | ||

+ | *# http://scikit-learn.org/stable/modules/linear_model.html#ridge-regression Ridge Regression | ||

+ | *# https://alex.smola.org/papers/2003/SmoSch03b.pdf SVR | ||

+ | * '''Base algorithm:''' In the classification The problem, we used an algorithm similar to linear SVM, whose relationship with the energy estimate, which is outside the scope of the classification The problem, is described in the article https://hal.inria.fr/hal-01591154/. For MSE, there is already a formulated dual The problem as a regression loss function, with the implementation of which we can start. | ||

+ | * '''Solution:''' The first step is to solve the problem with the MSE in the loss function using a solver that is convenient for you. The main difficulty may be the large dimensionality of the data, but they are sparse. Further it will be possible to change the wording of the problem. | ||

+ | * '''Novelty:''' Many models used to predict the interactions of proteins with ligands are "retrained" for some The problem. For example, models that are good at predicting binding energies may be poor at selecting a protein-binding molecule from a variety of non-binding ones, and models that are good at determining the correct geometry of the complex may be poor at predicting energies. In this problem, we propose to consider a new approach to combat such overfitting, since the combination of classification and regression loss functions seems to us to be a very natural regularization. | ||

+ | * '''Authors:''' Sergei Grudinin, Maria Kadukova. | ||

+ | |||

+ | ===Problem 75.2021=== | ||

+ | * '''Title:''' Alignment of image elements using metric models. | ||

+ | * '''Problem description:''' Character set specified. Each symbol is represented by one file - an image. Image pixel size may vary. All images are known to belong to the same class, such as faces, letters, flowers, or cars. (A more complicated option is to one class, which we are studying and noise classes.) It is known that each image can be combined with another with the help of an equalizing transformation up to noise, or up to some average image. (This image may or may not be present in the sample). This leveling transformation is specified in the base case by a neural network, and in the proposed case - by a parametric transformation from some given class (the first is a special case of the second). The aligned image is compared with the original one using the distance function. If the distance between two images is statistically significant, it is concluded that the images belong to the same class. It is required to 1) propose an adequate model of the alignment transformation that takes into account the assumptions about the nature of the image (for example, only rotation and proportional scaling), 2) propose a distance function, 3) propose a method for finding the average image. | ||

+ | * '''Data:''' Synthetic and real 1) pictures - faces and symbols with rotation and stretch transformation, 2) faces and cars with 3D rotation transformation with 2D projection. Synthetic images are proposed to be created manually using 1) photographs of a sheet of paper, 2) photographs of the surface of the drawing on a balloon. | ||

+ | * '''References:''' | ||

+ | *# support work - alignment of images using 2D DTW, | ||

+ | *# support work - alignment of images using neural networks, | ||

+ | *# DTW alignment work in 2D, | ||

+ | *# parametric alignment work. | ||

+ | * '''Base algorithm:''' from work 1. | ||

+ | * '''Solution:''' In the attached file pdf. | ||

+ | * '''Novelty:''' Instead of multidimensional image alignment, parametric alignment is proposed. | ||

+ | * '''Authors:''' Alexey Goncharov, Strijov V.V. | ||

+ | |||

+ | ===Problem 80.2021=== | ||

+ | * '''Title:''' Detection of correlations between activity in social networks and capitalization of companies | ||

+ | * '''Problem description:''' At present, the significant impact on stock quotes, company capitalization and the success or failure of an IPO depends on social factors such as public opinion expressed on social media. A recent notable example is the change in GameStore quotes caused by the surge in activity on Reddit. Our The problem at the first stage is to identify quotes between the shares of companies in different segments and activity in social networks. That is, it is necessary to identify correlations between significant changes in the company's capitalization and previous bursts (positive or negative) of its discussion in social networks. That is, it is necessary to find the minimum of the loss function when restoring the dependence in various classes of models (parametrics, neural networks, etc.). This The problem is part of a large project to analyze the analysis of markets and the impact of social factors on risks (within a team of 5-7 professors), which will lead to a series of publications sufficient to defend a dissertation. | ||

+ | * '''Data:''' The problem has a significant engineering context, the data is downloads from quotes on the Moscow Exchange, as well as NYT and reddit data (crawling and parsing is done by standard tools). The student working on this The problem must have strong engineering skills and a desire to engage in both the practice of machine learning and the engineering parts of The problem. | ||

+ | * '''References:''' | ||

+ | *# Paul S. Adler and Seok-Woo Kwon. Social Capital: Prospects for a new Concept. [https://journals.aom.org/doi/abs/10.5465/AMR.2002.5922314 LINK] | ||

+ | *# Kim and Hastak. Social network analysis: Characteristics of online social networks after a disaster [https://www.sciencedirect.com/science/article/pii/S026840121730525X?casa_token=JzqhHlll56IAAAAA:fQmNqxyErD4-VCCCFdJRA1WX0o4zdifj_zbm-vgwXDcmt26OBbAdu9gvgob0ntnlnCt_Y_ITD_g LINK] | ||

+ | *# Baumgartner, Jason, et al. "The pushshift reddit dataset." Proceedings of the International AAAI Conference on Web and Social Media. Vol. 14. 2020. [https://ojs.aaai.org/index.php/ICWSM/article/download/7347/7201/ LINK] | ||

+ | * '''Base algorithm:''' The basic algorithms are LSTM and Graph neural networks. | ||

+ | * '''Solution:''' Let's start by using LSTM, then try some of its standard extensions | ||

+ | * '''Novelty:''' In this area, there are a lot of economic, model solutions, but the accuracy of these solutions is not always high. The use of modern ML/DL models is expected to significantly improve the quality of the solution. | ||

+ | * '''Authors:''' Expert Yuri Maksimov, consultant Yuri Maksimov, student. | ||

+ | |||

+ | ===Problem 88b.2021=== | ||

+ | *'''Name:''' Finding a Pupil in an Eye Image Using the Luminance Projection Method | ||

+ | *'''Problem:''' Given a monochrome bitmap of the eye, [[Media:Matveev2021project.pdf|examples]]. It is necessary to determine the approximate coordinates of the center of the pupil. The word "approximate" means that the calculated pupil center must lie inside a circle centered at the pupil's true center and half the true radius. The algorithm must be very fast. | ||

+ | *'''Data:''' About 200 thousand eye images. For each, the position of the true circle is marked - for the purpose of training and testing the method being created. | ||

+ | '''Basic algorithm:''' To speed up work with the image, it is proposed to aggregate data using brightness projections. Image brightness is a function of two discrete arguments. Its projection on the horizontal axis is equal to. Similarly, projections are constructed on axes with an inclination. Having built several projections (two, four), based on them, you can try to determine the position of the pupil (compact dark area) using heuristics and / or a neural network. It is interesting to evaluate the capabilities of the neural network in this The problem. | ||

+ | *'''References:''' Zhi-Hua Zhou, Xin Geng Projection functions for eye detection // Pattern Recognition. 2004. V.37ю N.5. P.1049-1056. [https://doi.org/10.1016/j.patcog.2003.09.006 PDF] | ||

+ | *'''Author:''' Matveev I.A. | ||

+ | |||

+ | ===Problem 88c.2021=== | ||

+ | *'''Name:''' Searching for a century in an image as a parabolic contour using the projection method. | ||

+ | *'''Problem:''' Given a monochrome bitmap of the eye, [[Media:Matveev2021project.pdf|examples]]. It is necessary to find the contour of the upper eyelid as a parabola, that is, to determine the parameters. | ||

+ | *'''Data:''' About 200 thousand eye images. For some (about 2500), a human expert marked the position of a parabola that approximates the eyelid. | ||

+ | *'''Basic algorithm:''' The first step is pre-processing the image with a vertical gradient filter with further binarization, below is a typical result. There are various options for the next step. For example, if the coordinates of the pupil are known, you can set the region of interest (from above) and in it, using the selected points, construct a parabola by approximation using the least squares method. An example result is given below. More subtle methods are possible, such as finding a parabola using the Hough transform (see Wikipedia). Another way is to use projective methods (Radon transform). The main idea: after specifying the coefficient , apply a coordinate transformation to the image, as a result of which all parabolas of the form formula turn into lines of the form , then, given the coefficient , apply the coordinate transformation where , after which the oblique lines of the formula form become horizontal, which are easy to determine, for example, by horizontal projection (by summing the values in the rows of the matrix of the resulting image. If the coefficients are guessed correctly, the perabola representing the eyelid will give a clear maximum in the projection. By going through the formula (having a physical meaning), you can find those that give the maximum projection value, and consider that the desired parabola - eyelid. | ||

+ | *'''References:''' Wikipedia, articles "Hough Transform", "Radon Transform". | ||

+ | *'''Author:''' Matveev I.A. | ||

+ | |||

+ | ===Problem 62.2021=== | ||

+ | * '''Title:''' Construction of a method for dynamic alignment of multidimensional time series, resistant to local signal fluctuations. | ||

+ | * '''Problem description:''' In the process of working with multidimensional time series, the situation of the close proximity of sensors corresponding to different measurement channels is common. As a result, small signal shifts in space can lead to signal peak fixation by neighboring sensors, which leads to significant differences in measurements in terms of L2 distance.<br />Thus, small signal shifts lead to significant fluctuations in the readings of the sensors. The problem of constructing a distance function between points of time series that is resistant to noise generated by small spatial signal shifts is considered. It is necessary to consider the problem in the approximation of the presence of a map of the location of the sensors. | ||

+ | * '''Data:''' | ||

+ | *# [http://neurotycho.org/download Monkey brain activity measurements] | ||

+ | *# Artificially created data (several options must be proposed, for example signal movement in space clockwise and counterclockwise) | ||

+ | * '''References:''' | ||

+ | *# [https://www.cs.unm.edu/~mueen/DTW.pdf Reviriew DTW] | ||

+ | *# [https://www.researchgate.net/publication/228740947_Multi-dimensional_dynamic_time_warping_for_gesture_recognition Multi-Dimensional Dynamic Time Warping for Gesture Recognition] | ||

+ | *# [https://www.semanticscholar.org/paper/Multiple-Multidimensional-Sequence-Alignment-Using-Sanguansat/76d35bd5a52453ebde80faaa1467d7effd74426f Multiple Multidimensional Sequence Alignment Using Generalized Dynamic Time Warping] | ||

+ | * '''Base algorithm:''' L2 distance between a pair of measurements. | ||

+ | * '''Solution:''' Use the DTW distance function between two multidimensional time series. Two time axes are aligned, while inside the DTW functional, the distance between the i-th and j-th measurements is chosen such that it is resistant to local “shifts” of the signal. It is required to offer such functionality. The basic solution is L2, the improved solution is DTW between the i-th and j-th dimensions (dtw inside dtw).<br />You can suggest some modification, for example, the distance between the hidden layers of the autoencoder for points i and j. | ||

+ | * '''Novelty:''' A method for aligning multidimensional time series is proposed that takes into account small signal fluctuations in space. | ||

+ | * '''Authors:''' Expert Strijov V.V., consultants Gleb Morgachev, Alexey Goncharov. | ||

+ | |||

+ | ===Problem 58.2021=== | ||

+ | * '''Title:''' Transformation of the Gerchberg-Saxton algorithm using Bayesian neural networks. (or Neural network approach in the problem of phase search for images from the European synchrotron) | ||

+ | * '''Problem description:''' The aim of the project is to improve the quality of resolution of images of nanosized objects obtained in the laboratories of the European Synchrotron Radiation Foundation. | ||

+ | * '''Data:''' Contact an advisor for data (3GB). | ||

+ | '''References:''' | ||

+ | *# [https://arxiv.org/pdf/1809.04626.pdf] Iterative phase retrieval in coherent diffractive imaging: practical issues | ||

+ | *# [https://www.nature.com/articles/s41467-019-08635-x#Sec15] X-ray nanotomography of coccolithophores reveals that coccolith mass and segment number correlate with grid size | ||

+ | *# [https://www.nature.com/articles/s41598-018-34253-6#Sec14] Lens-free microscopy for 3D + time acquisitions of 3D cell culture | ||

+ | *# [https://arxiv.org/pdf/1904.11301.pdf] DEEP ITERATIVE RECONSTRUCTION FOR PHASE RETRIEVAL | ||

+ | *# https://docs.google.com/document/d/1K7bIzU33MSfeUvg3WITRZX0pe3sibbtH62aw42wxsEI/edit?ts=5e42f70e LinkReview | ||

+ | * '''Base algorithm:''' The transition from direct space to reciprocal space occurs using the Fourier transform. The Fourier transform is a linear transformation. Therefore, it is proposed to approximate it with a neural network. For example, an autoencoder for modeling forward and inverse Fourier transforms. | ||

+ | *'''Solution:''' Transformation of the Gerchberg-Saxton algorithm using Bayesian neural networks. Use of information on physical limitations and expertise. | ||

+ | *'''Novelty:''' Use of information about physical constraints and expert knowledge in the construction of the error function. | ||

+ | *'''Authors:''' Experts Sergei Grudinin, Yuri Chushkin, Strijov V.V., consultant Mark Potanin | ||

+ | |||

+ | ===Problem 63.2021=== | ||

+ | * '''Title:''' Hierarchical alignment of time sequences. | ||

+ | * '''Problem description:''' The problem of alignment of sequences of difficult events is considered. An example is the complex behavior of a person: when considering data from IMU sensors, one can put forward a hypothesis: there is an initial signal, there are aggregates of “elementary actions” and there are aggregates of “actions” of a person. Each of the indicated levels of abstraction can be distinguished and operated on exactly by it.<br />In order to accurately recognize the sequence of actions, it is possible to use metric methods (for example, DTW, as a method that is resistant to time shifts). For a more accurate quality of timeline alignment, it is possible to carry out alignment at different levels of abstraction.<br />It is proposed to explore such a hierarchical approach to sequence alignment, based on the possibility of applying alignment algorithms to objects of different structures, having a distance function on them. | ||

+ | * '''References:''' | ||

+ | *# [https://www.cs.unm.edu/~mueen/DTW.pdf Overview presentation about DTW] | ||

+ | *# [https://link.springer.com/article/10.1007/s00371-015-1092-0 DTW-based kernel and rank-level fusion for 3D gait recognition using Kinect Multi-Dimensional Dynamic Time Warping for Gesture Recognition] | ||

+ | *# [https://ieeexplore.ieee.org/abstract/document/8966048 Time Series Similarity Measure via Siamese Convolutional Neural Network] | ||

+ | *# [https://www.semanticscholar.org/paper/Multiple-Multidimensional-Sequence-Alignment-Using-Sanguansat/76d35bd5a52453ebde80faaa1467d7effd74426f Multiple Multidimensional Sequence Alignment Using Generalized Dynamic Time Warping] | ||

+ | * '''Base algorithm:''' classic DTW. | ||

+ | * '''Solution:''' It is proposed to perform the transition from one level of abstraction to another by using convolutional and recurrent neural networks. Then the object at the lower level of abstraction is the original signal. At the second level - a signal from the hidden layer of the model (built on the objects of the lower level), the dimension of which is much less, and the upper layer - a signal from the hidden layer of the model (built on the objects of the middle level).<br />In this case, DTW is calculated separately between the lower , between the middle and between the upper levels, but the formation of objects for calculating the distance is carried out taking into account the alignment path between the objects of the previous level.<br />This method is considered as a way to increase the interpretability of the alignment procedure and the accuracy of the action classification in connection with the transition to higher-level patterns. In addition, a significant increase in speed is expected. | ||

+ | * '''Novelty:''' The idea of aligning time sequences simultaneously at several levels of abstraction is proposed. The method should significantly improve the interpretability of alignment algorithms and increase their speed. | ||

+ | * '''Authors:''' Strijov V.V. Expert, Gleb Morgachev, Alexey Goncharov consultants. | ||

+ | |||

+ | ===Problem 57.2021=== | ||

+ | * '''Title:'''Additive Regularization and in The problems of Privileged Learning in Solving the Problem of Predicting the State of the Ocean | ||

+ | * '''Problem description:''' There is a sample of data from ocean buoys, it is required to predict the state of the ocean at different points in time. | ||

+ | * '''Data:''' The buoys provide data on wave height, wind speed, wind direction, wave period, sea level pressure, air temperature and sea surface temperature with a resolution of 10 minutes to 1 hour. | ||

+ | *# '''References:''' | ||

+ | *# [https://arxiv.org/pdf/1906.00195.pdf] | ||

+ | * '''Base algorithm:''' Using a simple neural network. | ||

+ | * '''Solution:'''Adding to the basic algorithm (a simple neural network) a system of differential equations. Explore the properties of the parameter space of teacher and student according to the preferred approach. | ||

+ | *'''Novelty:''' Investigation of the parameter space of the teacher and the student and their change. It is possible to set up separate teacher and student models and track the change in their parameters in the optimization process - variance, change in the quality of the student when adding teacher information, complexity. | ||

+ | * '''Authors:''' Strijov V.V., Mark Potanin | ||

+ | |||

+ | ===Problem 52. 2021=== | ||

+ | * '''Title:''' Predicting the quality of protein models using spherical convolutions on 3D graphs. | ||

+ | * '''Problem:''' The purpose of this work is to create and study a new convolution operation on three-dimensional graphs in the framework of solving the problem of assessing the quality of three-dimensional protein models (The problem regression on graph nodes). | ||

+ | * '''Data:''' Models generated by CASP competitors are used (http://predictioncenter.org). | ||

+ | * '''References:''' | ||

+ | *# [https://drive.google.com/file/d/1pXCED8XBcxbjwtg_1wZG0oAjvUCxFlua/view?usp=sharing] More about The problem. | ||

+ | *# [https://arxiv.org/abs/1806.01261] Relational inductive biases, deep learning, and graph networks. | ||

+ | *# [https://arxiv.org/abs/1611.08097] Geometric deep learning: going beyond euclidean data. | ||

+ | * '''Base algorithm:''' As a basic algorithm, we will use a neural network based on the graph convolution method, which is generally described in [https://arxiv.org/abs/1806.01261]. | ||

+ | * '''Solution:''' The presence of a peptide chain in proteins makes it possible to uniquely introduce local coordinate systems for all graph nodes, which makes it possible to create and apply spherical filters regardless of the graph topology. | ||

+ | * '''Novelty:''' In the general case, graphs are irregular structures, and in many graph learning The problems, the sample objects do not have a single topology. Therefore, the existing operations of convolutions on graphs are greatly simplified or do not generalize to different topologies. In this paper, we propose to consider a new method for constructing a convolution operation on three-dimensional graphs, for which it is possible to uniquely choose local coordinate systems associated with each node. | ||

+ | * '''Authors:''' Sergei Grudinin, Ilya Igashov. | ||

+ | |||

+ | ===Problem 44+. 2021=== | ||

+ | *'''Title:''' Early prediction of sufficient sample size for a generalized linear model. | ||

+ | *'''Deiscription''': The problem of experiment planning is investigated. The problem of estimating a sufficient sample size according to the data is solved. The sample is assumed to be simple. It is described by an adequate model. Otherwise, the sample is generated by a fixed probabilistic model from a known class of models. The sample size is considered sufficient if the model is restored with sufficient confidence. It is required, knowing the model, to estimate a sufficient sample size at the early stages of data collection. | ||

+ | * '''Goal''': On a small simple iid sample, predict the error on a replenished large one. The predictive model is smooth monotonic in two derivatives. The choice of model is a complete enumeration or genetics. The model depends on the reduced (explore) covariance matrix of the GLM parameters. | ||

+ | *'''Data:''' For the computational experiment, it is proposed to use classical samples from the UCI repository. Link to selections https://github.com/ttgadaev/SampleSizeEstimation/tree/master/datasets | ||

+ | *'''References:''' | ||

+ | *# [https://docs.google.com/document/d/1o2gtdV3nYeAsfW0JZ5fESlVPhCA4_lfUOVnWhRjg1ck/edit?usp=sharing Overview of Methods, Motivation and Problem Statement for Sample Size Estimation] | ||

+ | *# http://svn.code.sf.net/p/mlalgorithms/code/PhDThesis/. | ||

+ | *# Bootstrap method. https://projecteuclid.org/download/pdf_1/euclid.aos/1. | ||

+ | Bishop, C. 2006. Pattern Recognition and Machine Learning. Berlin: Springer. 758 p. | ||

+ | *'''Basic algorithm''': We will say that the sample size is sufficient if the log-likelihood has a small variance on a sample of size m calculated using the bootstrap. | ||

+ | We are trying to approximate the dependence of the average value of log-likelihood and its variance on the sample size. | ||

+ | *'''Solution:''' The methods described in the review are asymptotic or require a deliberately large sample size. The new method should be to predict volume in the early stages of experiment design, i.e. when data is scarce. | ||

+ | *'''Authors:''' expert Strijov V.V., consultant Malinovsky G. | ||

+ | |||

+ | ===Problem 12.2021=== | ||

+ | * '''Title:''' Machine translation training without parallel texts. | ||

+ | * '''Problem:''' The problem of building a text translation model without the use of parallel texts is considered, i.e. pairs of identical sentences in different languages. This The problem occurs when building translation models for low-resource languages (that is, languages for which there is not much data in the public domain). | ||

+ | * '''Data:''' A selection of articles from Wikipedia in two languages. | ||

+ | * '''References:''' | ||

+ | *# [https://arxiv.org/abs/1711.00043] Unsupervised Machine Translation Using Monolingual Corpora Only | ||

+ | *# [https://arxiv.org/pdf/1609.08144.pdf] Sequence to sequence. | ||

+ | *# [http://www.cs.toronto.edu/~larocheh/publications/icml-2008-denoising-autoencoders.pdf] Autoencoding. | ||

+ | *# [https://arxiv.org/pdf/1511.06709.pdf] Training with Monolingual Training Data. | ||

+ | * '''Basic algorithm''': Unsupervised Machine Translation Using Monolingual Corpora Only. | ||

+ | * '''Solution:''' As a translation model, it is proposed to consider a combination of two auto-encoders, each of which is responsible for presenting sentences in one of the languages. The models are optimized in such a way that the latent spaces of autoencoders for different languages match. As an initial representation of sentences, it is proposed to consider their graph description obtained using multilingual ontologies. | ||

+ | * '''Novelty:''' A method for constructing a translation model is proposed, taking into account graph descriptions of sentences. | ||

+ | * '''Authors:''' Oleg Bakhteev, Strijov V.V., | ||

+ | |||

+ | ===Problem 8.2021=== | ||

+ | * '''Title:''' Generation of features using locally approximating models (Classification of human activities according to measurements of fitness bracelets). | ||

+ | * '''Problem:''' It is required to check the feasibility of the hypothesis about the simplicity of sampling for the generated features. Features are the optimal parameters of approximating models. Moreover, the entire sample is not simple and requires a mixture of models to approximate it. Explore the information content of the generated features - the parameters of the approximating models trained on the segments of the original time series. According to the measurements of the accelerometer and gyroscope, it is required to determine the type of activity of the worker. It is assumed that the time series of measurements contain elementary movements that form clusters in the space of time series descriptions. The characteristic duration of the movement is seconds. Time series are labeled with activity type labels: work, leisure. The typical duration of activity is minutes. It is required to restore the type of activity according to the description of the time series and cluster. | ||

+ | * '''Data:''' WISDM accelerometer time series ([[Time series (library of examples)]], section Accelerometry). | ||

+ | *# WISDM (Kwapisz, J.R., G.M. Weiss, and S.A. Moore. 2011. Activity recognition using cell phone accelerometers. ACM SigKDD Explorations Newsletter. 12(2):74–82.), USC-HAD. Human activity recognition using smart phone embedded sensors: A Linear Dynamical Systems method, W Wang, H Liu, L Yu, F Sun - Neural Networks (IJCNN), 2014. | ||

+ | * '''References:''' | ||

+ | *# Motrenko A.P., Strijov V.V. Extracting fundamental periods to segment human motion time series // Journal of Biomedical and Health Informatics, 2016, Vol. 20, No. 6, 1466 - 1476. [http://sourceforge.net/p/mlalgorithms/code/HEAD/tree/Group874/Motrenko2014TSsegmentation/JBHI/MotrenkoStrijov2014RV2.pdf?format=raw URL] | ||

+ | *# Karasikov M.E., Strijov V.V. Classification of time series in the space of parameters of generating models // Informatics and its applications, 2016.[http://strijov.com/papers/Karasikov2016TSC.pdf URL] | ||

+ | *# Kuznetsov M.P., Ivkin N.P. Algorithm for Classifying Accelerometer Time Series by Combined Feature Description // Machine Learning and Data Analysis. 2015. T. 1, No. 11. C. 1471 - 1483. [http://jmlda.org/papers/doc/2015/no11/Ivkin2015TSclassification.pdf URL] | ||

+ | *# Isachenko R.V., Strijov V.V. Metric learning in The problem of multiclass classification of time series // Informatics and its applications, 2016, 10(2) : 48-57. [http://strijov.com/papers/Isachenko2016MetricsLearning.pdf URL] | ||

+ | *# Zadayanchuk A.I., Popova M.S., Strijov V.V. Choosing the optimal model for classifying physical activity based on accelerometer measurements // Information technologies, 2016. [http://strijov.com/papers/Zadayanchuk2015OptimalNN4.pdf URL] | ||

+ | *# Ignatov A., Strijov V. Human activity recognition using quasiperiodic time series collected from a single triaxial accelerometer // Multimedia Tools and Applications, 2015, 17.05.2015 : 1-14. [http://strijov.com/papers/Ignatov2015HumanActivity.pdf URL] | ||

+ | * '''Basic algorithm''': Basic algorithm described in [Karasikov, Strijov: 2016] and [Kuznetsov, Ivkin: 2014]. | ||

+ | * '''Solution:''' It is required to build a set of locally approximating models and choose the most adequate ones. Find the optimal segmentation method and the optimal description of the time series. Construct a metric space of descriptions of elementary motions. | ||

+ | * '''Novelty:''' A standard for building locally approximating models has been created. The connection of two characteristic times of the description of human life, the combined statement of the problem. | ||

+ | * '''Authors:''' Expert Strijov V.V., consultants Alexandra Galtseva, Danil Sayranov. | ||

+ | |||

+ | ==2020== | ||

+ | |||

+ | {|class="wikitable" | ||

+ | |- | ||

+ | ! Author | ||

+ | ! Topic | ||

+ | ! Links | ||

+ | ! Consultant | ||

+ | ! Letters | ||

+ | ! Reviewer | ||

+ | |- | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Magistrkoljan Grebenkova Olga] | ||

+ | |Variational optimization of deep learning models with model complexity control | ||

+ | |[https://docs.google.com/document/d/1gHyVeYgzFgco1vUTZRjxT2FbO03GsB27EVEstLWTzdM/edit?usp=sharing LinkReview] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020-Project60 GitHub] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020-Project60/raw/master/docs/Grebenkova2020Optimization.pdf Paper] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020-Project60/raw/master/slides/Grebenkova2020OptimizationSlides.pdf Slides] | ||

+ | [https://youtu.be/9ELhIqjFSE8 Video] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Oleg_Bakhteev Oleg Bakhteev] | ||

+ | |AILP+UXBR+HCV+TEDWS | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Vshokorov Shokorov Vyacheslav] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020_Project_9/raw/master/review%20Grebenkova.pdf Review] | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Vshokorov Shokorov Vyacheslav] | ||

+ | |Text recognition based on skeletal representation of thick lines and convolutional networks | ||

+ | |[https://docs.google.com/document/d/1zsk-tpd51axWfcYxpa4CWd1QZdOnr0Hv6b1_a34q28Y/edit?usp=sharing LinkReview] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020_Project_9 GitHub] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020_Project_9/raw/master/Image_classification_based_on_skeletonization_and_Graph_NN.pdf Paper] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020_Project_9/raw/master/slides_Image_classification_based_on_skeletonization_and_Graph_NN.pdf Slides] | ||

+ | [https://youtu.be/0je5wvaz_tQ Video] | ||

+ | |Denis Ozherelkov | ||

+ | |AIL | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Magistrkoljan Grebenkova Olga] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020-Project60/raw/master/docs/Shokorov2020ImageClassification_Review.pdf Review] | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Filatov Filatov Andrey] | ||

+ | |Intention forecasting. Investigation of the properties of local models in the spatial decoding of brain signals | ||

+ | |[https://docs.google.com/document/d/1UmRq34enjk7RpW2vpF5V88TaHKQd0Ne3LpwyoV0E6nA/edit?usp=sharing LinkReview] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020-Project-17 GitHub] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020-Project-17/raw/master/docs/Filatov2020LocalModel.pdf Paper] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020-Project-17/raw/master/slides/Filatov2020LocalModelSlides.pdf Slides] | ||

+ | [https://youtu.be/q5Skhl1H5cA Video] | ||

+ | |Valery Markin | ||

+ | |AILPHUXBRCVTEDWS | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Hristolubov_Maxim Hristolubov Maxim] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020_Project8/raw/master/docs/%D0%A0%D0%B5%D1%86%D0%B5%D0%BD%D0%B7%D0%B8%D1%8F.pdf Review] | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Rustem_Messi Islamov Rustem] | ||

+ | |Analysis of the properties of an ensemble of locally approximating models | ||

+ | |[https://docs.google.com/document/d/1wEYR3vXzZsYEv2L51wMCBFmP7UQwIBDPn3Gpz72MIyw/edit LinkReview] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020_Project-51 GitHub] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020_Project-51/raw/master/paper/Islamov2020EnsembleOfModels.pdf Paper] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020_Project-51/raw/master/slides/Islamov2020EnsembleOfModels_Presentation.pdf Slides] | ||

+ | [https://youtu.be/9yFRWsyj6zo Video] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Andriygav Andrey Grabovoi] | ||

+ | |AILPHUXBRCVTEDWS | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Gunaev_Ruslan Gunaev Ruslan] | ||

+ | [https://github.com/Gunaev/2020-Project-69/raw/master/paper/Islamov2020_Review.docx Review] | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Zholobov_Vladimir Zholobov Vladimir] | ||

+ | |Early prediction of sufficient sample size for a generalized linear model. | ||

+ | |[https://docs.google.com/document/d/1o2gtdV3nYeAsfW0JZ5fESlVPhCA4_lfUOVnWhRjg1ck/edit LinkReview] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020-Project44 GitHub] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020-Project44/raw/master/doc/paper/Zholobov2020SampleSizeEstimation.pdf Paper] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020-Project44/raw/master/slides/Zholobov2020Prezentation.pdf Slides] | ||

+ | [https://youtu.be/uWhaND3e1cw Video] | ||

+ | |Grigory Malinovsky | ||

+ | |AILPHUXBRCVTEWSF | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Vayser_Kirill Vayser Kirill] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020-Project_Regul/raw/master/docs/Zholobov2020EarlyForecast_Review.docx Review] | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Vayser_Kirill Vayser Kirill] | ||

+ | |Additive regularization and its meta parameters when choosing the structure of deep learning networks | ||

+ | |[https://docs.google.com/document/d/1LRVQ8dgRejQx8zdtk6dLMbHXdXwbAju6qD8NNSa1MgE/edit?usp=sharing LinkReview] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020-Project_Regul GitHub] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020-Project_Regul/raw/master/docs/Vayser2020AdditiveRegularization.pdf Paper] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020-Project_Regul/raw/master/docs/Vayser2020AdditiveRegularizationSlides.pdf Slides] | ||

+ | [https://youtu.be/tsMS1HTxVYU Video] | ||

+ | |Mark Potanin | ||

+ | |AILP+HUX+BRCV+TEDWS | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Zholobov_Vladimir Zholobov Vladimir] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020-Project44/blob/master/doc/review/Vaiser2020review.docx Review] | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Bishuk_Anton Bishuk Anton] | ||

+ | |Solution of an optimization problem combining classification and regression to estimate the binding energy of a protein and small molecules. | ||

+ | |[https://drive.google.com/file/d/1NPz05B6HceCdD1Q-P8xYCUkc15bka2Qz/view?usp=sharing LinkReview] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020_Project53_Class-Reg/ GitHub] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020_Project53_Class-Reg/blob/master/docs/Bishuk_2020_Cls_Rg_in_Mol_Docking.pdf Paper] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020_Project53_Class-Reg/blob/master/docs/Bishuk_2020_Cls_Rg_in_Mol_Docking_pres.pdf Slides] | ||

+ | [https://youtu.be/8sRcvKR2F-0 Video] | ||

+ | |Maria Kadukova | ||

+ | |AILPHUXBRCVTEDH | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Filippova_Anastasia Filippova Anastasia] | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Filippova_Anastasia Filippova Anastasia] | ||

+ | |Step detection for IMU navigation via deep learning | ||

+ | |[https://docs.google.com/spreadsheets/d/1XLDBM53bX_7_HwCYbmuZTY8IlbcE0A4B1BQ8EnIXJEo/edit?usp=sharing LinkReview] | ||

+ | [https://github.com/nastya236/Step-detection-for-IMU-navigation-via-deep-learning- GitHub] | ||

+ | [https://github.com/nastya236/Step-detection-for-IMU-navigation-via-deep-learning-/raw/master/docs/__.pdf Paper] | ||

+ | [https://github.com/nastya236/Step-detection-for-IMU-navigation-via-deep-learning-/raw/master/slides/Step-detection-for-IMU-navigation-via-deep-learning_slides.pdf Slides] | ||

+ | [https://github.com/nastya236/Step-detection-for-IMU-navigation-via-deep-learning-/blob/master/docs/Step_detection.pdf EnglishPaper] | ||

+ | [https://youtu.be/ox4llj_xz_c Video] | ||

+ | |Tamaz Gadaev | ||

+ | |AIL0PUXBRCVSF | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Bishuk_Anton Bishuk Anton] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020_Project53_Class-Reg/raw/master/Review.pdf Review] | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Savelev_Nickolay Savelev Nickolay] | ||

+ | |Distributed optimization under Polyak-Loyasievich conditions | ||

+ | |[https://docs.google.com/document/d/1tXEXnjv8F1CFYGSbdlp1Fd0fbU49N1E5bGnwo6XW3CU/edit?usp=sharing LinkReview] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020-Project59 GitHub] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020-Project59/raw/master/docs/Savelev2020PLoptimization.pdf Paper] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020-Project59/raw/master/docs/Savelev2020slides.pdf Slides] | ||

+ | [https://youtu.be/BefA7U_h1CI Video] | ||

+ | |A. N. Beznosikov | ||

+ | |AILPHUXBRCVTEDWS | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Lexakhar Khary Alexandra] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020-Project59/raw/master/docs/review.pdf Review] | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Lexakhar Khary Alexandra] | ||

+ | |Theoretical validity of the application of metric classification methods using dynamic alignment (DTW) to spatiotemporal objects. | ||

+ | |[https://docs.google.com/document/d/1B2INH2qRFHpUJWBMwn27kyQ6ySMI5i1-N322nzKUApY/edit?usp=sharing LinkReview] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020-Project64 GitHub] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020-Project64/raw/master/dosc/Khar2020DTWusing.pdf Paper] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020-Project64/raw/master/slides/SlidesKhar2020DTW.pdf Slides] | ||

+ | [https://youtu.be/_uXT3dVbEQQ Video] | ||

+ | |Gleb Morgachev, Alexey Goncharov | ||

+ | |AILPHUXBRCVTEDCWS | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Savelev_Nickolay Savelev Nickolay] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020-Project64/raw/master/dosc/Review.pdf Review] | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Hristolubov_Maxim Hristolubov Maxim] | ||

+ | |Generating features using locally approximating models (Classification of human activities by measurements of fitness bracelets) | ||

+ | |[https://drive.google.com/open?id=1j9NUd2r3rAmNlt_iobBcxHM8Nc1uXk51gCe4AAr1Evs LinkReview] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020_Project8 GitHub] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020_Project8/raw/master/docs/Hristolubov2020Accelerometer.pdf Paper] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020_Project8/raw/master/docs/Hristolubov2020slides.pdf Slides] | ||

+ | [https://youtu.be/fa-lipA-9G0 Video] | ||

+ | |Alexandra Galtseva, Danil Sayranov | ||

+ | |AILPH | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Filatov Filatov Andrey] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020-Project-17/raw/master/report/Hristolubov2020AccelerometerReview.pdf Review] | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Mamonov Mamonov Kirill] | ||

+ | |Nonlinear ranking of exploratory information search results. | ||

+ | |[https://docs.google.com/document/d/1PEIvEfvq_2Mo62M5jMN0Fgg_XTuWSoYMvdssnTlSXn4/edit?usp=sharing LinkReview] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020-Project73 GitHub] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020-Project73/raw/master/report/Mamonov2020Project73.pdf Paper] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020-Project73/raw/master/report/Mamonov2020Project73slides.pdf Slides] | ||

+ | [https://youtu.be/9Gr_YWYriww Video] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:MEremeev Maxim Eremeev] | ||

+ | |AILPHU+XBRC+V+TEDHWJSF | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Pavlichenko Pavlichenko Nikita] | ||

+ | | Predicting the quality of protein models using spherical convolutions on 3D graphs. | ||

+ | |[https://docs.google.com/document/d/1EaExQN9F94kt_JAJnglX1liuo-qS4C9Hee8pLOUWlL8/edit?usp=sharing LinkReview] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020-Project52 GitHub] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020-Project52/raw/master/report/main.pdf Paper] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020-Project52/raw/master/report/NVPavlichenkoPresentation.pdf Slides] | ||

+ | [https://youtu.be/Sw9KmvpuXFs Video] | ||

+ | |Sergei Grudinin, Ilya Igashov | ||

+ | |AILPUXBRHCVTEDH | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Sodikov Sodikov Mahmud], [http://www.machinelearning.ru/wiki/index.php?title=Участник:Skachkov Skachkov Daniel] | ||

+ | | Agnostic neural networks | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/WeightAgnosticNN/raw/master/WANN_modif.py Code] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/WeightAgnosticNN/raw/master/WANN_article.pdf Paper] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/WeightAgnosticNN/raw/master/WANN_presentation.pdf Slides] | ||

+ | [https://youtu.be/KHP5UXH0fSE Video] | ||

+ | | Radoslav Neichev | ||

+ | |AILPHUXBRC+VTEDHWJSF | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Kulagin Kulagin Petr] | ||

+ | [https://github.com/petr-kulagin/2020-Project62/blob/master/docs/SodikovSkachkov2020Project66_Review.pdf Review] | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Gunaev_Ruslan Gunaev Ruslan] | ||

+ | | Graph Neural Network in Reaction Yield prediction | ||

+ | |[https://docs.google.com/document/d/18-eJP3-bPs-aYGGR2PuD3tjJdaa7CF59JMJanwRQLJM/edit LinkReview] | ||

+ | [https://github.com/Gunaev/2020-Project-69 Github] | ||

+ | [https://github.com/Gunaev/2020-Project-69/raw/master/paper/Gunaev2020GCNN.pdf Paper] | ||

+ | [https://github.com/Gunaev/2020-Project-69/raw/master/slides/Gunaev2020GCNN_presentation_final_version.pdf Slides] | ||

+ | [https://youtu.be/JTmut-CpowE Video] | ||

+ | |Philip Nikitin | ||

+ | |AILPUXBRHCVTEDHWSF | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Rustem_Messi Islamov Rustem] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020_Project-51/raw/master/doc/Gunaev2020Project69_Review.pdf Review] | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Fyaush Yaushev Farukh] | ||

+ | | Investigation of ways to match models by reducing the dimension of space | ||

+ | |[https://docs.google.com/document/d/14T3fHZycMMtvd-1LROd5gDOtbI-johIPp_RdiW_Qd3c/edit?usp=sharing LinkReview] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020-Project-71 Github] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020-Project-71/raw/master/report/Yaushev2020Title.pdf Paper] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020-Project-71/raw/master/slides/Yaushev2020TitleSlides.pdf Slides] | ||

+ | [https://youtu.be/2c3DvTkFtDc Video] | ||

+ | |Roman Isachenko | ||

+ | |AILPUXBRHCVTEDHWJS | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Zholobov_Vladimir Zholobov Vladimir] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2020-Project44/blob/master/doc/review/Yaushev2020review.docx Review] | ||

+ | |} | ||

+ | |||

+ | ===51. 2020=== | ||

+ | *'''Name:''' Analysis of the properties of an ensemble of locally approximating models. | ||

+ | *'''Problem''': In this paper, we consider The problem of constructing a universal approximator --- a multimodel, which consists of a given finite set of local models. Each local model approximates a connected region in feature space. It is assumed that the set of local models cover the entire space of objects. A convex combination of local models is considered as an aggregating function. As the coefficients of the convex combination, we consider a function depending on the object --- the gate function. | ||

+ | *'''Required''': To construct an algorithm for optimizing the parameters of local models and parameters of the gate function. It is required to propose a metric in the space of objects, a metric in the space of models. | ||

+ | *'''Data:''' | ||

+ | *# Synthetically generated data. | ||

+ | *# Energy consumption forecasting data. It is proposed to use the following models as local models: working day, day off. (Energy Consumption, Turk Electricity Consumption German Spot Price). | ||

+ | *'''References:''' | ||

+ | *# [https://github.com/andriygav/EMprior/blob/master/paper/Grabovoy2019MixtureOfExpertEng.pdf Overview of methods for estimating sample size] | ||

+ | *# [http://www.machinelearning.ru/wiki/images/2/21/Voron-ML-Compositions-slides2.pdf Vorontsov's lectures on compositions] | ||

+ | *# [http://www.machinelearning.ru/wiki/images/0/0d/Voron-ML-Compositions.pdf Vorontsov's lectures on compositions] | ||

+ | *# Esen Y.S., Wilson J., Gader P.D. Twenty Years of Mixture of Experts. IEEE Transactions on Neural Networks and Learning Systems. 2012. Issues. 23. No 8. P. 1177-1193. | ||

+ | *# [https://sourceforge.net/p/mlalgorithms/code/HEAD/tree/MSThesis/Pavlov2012/ Pavlov K.V. Selection of multilevel models in The problems classification, 2012] | ||

+ | *'''Basic algorithm''': As a basic algorithm, it is proposed to use a two-level optimization problem, where local models are optimized at one iteration and at the next iteration, the parameters of the gate function are optimized. | ||

+ | *'''Authors:''' Grabovoi A.V. (consultant), Strijov V.V. (Expert) | ||

+ | |||

+ | ===54. 2020=== | ||

+ | * '''Title:''' Finding the pupil in the eye image using the brightness projection method. | ||

+ | * '''Problem:''' Given a monochrome bitmap of the eye, see examples (https://cloud.mail.ru/public/eaou/4JSamfmrh). | ||

+ | It is necessary to determine the approximate coordinates of the center of the pupil. The word "approximate" means that the calculated pupil center must lie inside a circle centered at the pupil's true center and half the true radius. The algorithm must be very fast. | ||

+ | * '''Data:''' About 200 thousand eye images. For each, the position of the true circle is marked - for the purpose of training and testing the method being created. | ||

+ | * '''Base algorithm:''' To speed up work with the image, it is proposed to aggregate data using brightness projections. Image brightness is a function of two discrete arguments I(x, y). Its projection onto the horizontal axis is P(x)=\sum \limits_y I(x,y). Similarly, projections are constructed on axes with an inclination. Having built several projections (two, four), based on them, you can try to determine the position of the pupil (compact dark area) using heuristics and / or a neural network. It is interesting to evaluate the capabilities of the neural network in this The problem. | ||

+ | * '''References:''' Zhi-Hua Zhou, Xin Geng Projection functions for eye detection // Pattern Recognition. 2004. V.37ю N.5. P.1049-1056. https://doi.org/10.1016/j.patcog.2003.09.006 | ||

+ | * '''Authors:''' Matveev I.A. | ||

+ | |||

+ | ===55. 2020=== | ||

+ | * '''Title:''' Search for the boundaries of the iris by the method of circular projections | ||

+ | * '''Problem:''' Given a monochrome bitmap of the eye, see examples (https://cloud.mail.ru/public/2DBu/5c6F6e3LC). The approximate position of the center of the pupil is also known. The word "approximate" means that the calculated center of the pupil is no more than half of its true radius from the true one. It is necessary to determine the approximate positions of the circles approximating the pupil and iris. The algorithm must be very fast. | ||

+ | * '''Data:''' About 200 thousand eye images. For each, the position of the true circle is marked - for the purpose of training and testing the method being created. | ||

+ | * '''Base algorithm:''' To speed up work with the image, it is proposed to aggregate data using circular projections of brightness. Circular projection is a function that depends on the radius, the value of which P(r) is equal to the integral of the directed image brightness gradient over a circle of radius r (or along an arc of a circle). Example for one arc (right quadrant) and for four arcs. Having built some circular projections, based on them, you can try to determine the position of the inner and outer borders of the iris (ring) using heuristics and / or a neural network. It is interesting to evaluate the capabilities of the neural network in this The problem. | ||

+ | * '''References:''' Matveev I.A. Detection of Iris in Image By Interrelated Maxima of Brightness Gradient Projections // Applied and Computational Mathematics. 2010. V.9. N.2. P.252-257. https://www.researchgate.net/publication/228396639_Detection_of_iris_in_image_by_interrelated_maxima_of_brightness_gradient_projections | ||

+ | * '''Authors:''' Matveev I.A. | ||

+ | |||

+ | ===56. 2020=== | ||

+ | * '''Title:''' Construction of local and universal interpretable scoring models | ||

+ | * '''Problem:''' Build a simple and interpretable scoring system as a superposition of local models, taking into account the requirements for the system to retain knowledge about key customers and features (in other words, take into account new economic phenomena). The model must be a superposition, and each element must be controlled by its own quality criterion. Introduce a schedule for optimizing the structure and parameters of the model: the system must work in a single optimization chain. Propose an algorithm for selecting features and objects. | ||

+ | * '''Data:''' | ||

+ | # Data from OTP Bank. The sample contains records of 15,223 clients classified into two classes: 1 - there was a response (1812 clients), 0 - there was no response (13411 clients). Feature descriptions of clients consist of 50 features, which include, in particular, age, gender, social status in relation to work, social status in relation to pension, number of children, number of dependents, education, marital status, branch of work. The data are available at the following addresses: www.machinelearning.ru/wiki/images/2/26/Contest_MMRO15_OTP.rar (sample A), www.machinelearning.ru/wiki/images/5/52/Contest_MMRO15_OTP_(validation).rar (sample B). | ||

+ | # Data from Home Credit: https://www.kaggle.com/c/home-credit-default-risk/data | ||

+ | * '''References:''' | ||

+ | *# Strijov V.V. Error function in regression analysis // Factory Laboratory, 2013, 79(5) : 65-73 | ||

+ | *# Bishop C. M. Linear models for classification / В кн.: Pattern Recognition and Machine Learning. Под ред.: M. Jordan, J. Kleinberg, B. Scholkopf. – New York: Springer Science+Business Media, 2006, pp--203 – 208 | ||

+ | *# Tokmakova A.A. Obtaining Stable Hyperparameter Estimates for Linear Regression Models // Machine Learning and Data Analysis. — 2011. — № 2. — С. 140-155 | ||

+ | *# S. Scitovski and N. Sarlija. Cluster analysis in retail segmentation for credit scoring // CRORR 5. 2014. 235–245 | ||

+ | *# Goncharov A.V. Building Interpretable Deep Learning Models in the Social Ranking Problem | ||

+ | * '''Base algorithm:''' Iterative weighted least squares (described in (2)) | ||

+ | * '''Solution:''' It is proposed to build a scoring system containing such a preprocessing block as a block for generating metric features. It is proposed to investigate the influence of the non-equivalence of objects on the selection of features for the model, to investigate the joint selection of features and objects when building a model. It is required to implement a schedule for optimizing the model structure using an algorithm based on the analysis of covariance matrices of model hyperparameters. The schedule includes a phased replenishment of the set of features and objects. The feature sample size will be determined by controlling the error variance. The main criterion for the quality of the system: ROC AUC (Gini). | ||

+ | * '''Novelty:''' | ||

+ | # The model structure optimization schedule must satisfy the requirement to rebuild the model at any time without losing its characteristics. | ||

+ | # Accounting for the unequal value of objects in the selection of features | ||

+ | * '''Authors:''' Pugaeva I.V. (consultant), Strijov V.V. (Expert) | ||

+ | |||

+ | ===59. 2020=== | ||

+ | * Name: Distributed optimization under Polyak-Loyasievich conditions | ||

+ | * '''Problem description:''' The problem is to efficiently solve large systems of nonlinear equations using a network of calculators. | ||

+ | * '''Solution:''' A new method for decentralized distributed solution of systems of nonlinear equations under Polyak-Loyasievich's conditions is proposed. The approach is based on the fact that the distributed optimization problem can be represented as a composite optimization problem (see 2 from the literature), which in turn can be solved by analogs of the similar triangles or sliding method (see 2 from the literature). | ||

+ | * Basic algorithm: The proposed method is compared with gradient descent and accelerated gradient descent | ||

+ | * '''References:''' | ||

+ | *# Linear Convergence of Gradient and Proximal-GradientMethods Under the Polyak- Lojasiewicz Condition https://arxiv.org/pdf/1608.04636.pdf | ||

+ | *# Linear Convergence for Distributed Optimization Under the Polyak-Łojasiewicz Condition https://arxiv.org/pdf/1912.12110.pdf | ||

+ | *# Optimal Decentralized Distributed Algorithms for Stochastic ConvexOptimization https://arxiv.org/pdf/1911.07363.pdf | ||

+ | *# Modern numerical optimization methods, universal gradient descent method https://arxiv.org/ftp/arxiv/papers/1711/1711.00394.pdf | ||

+ | * '''Novelty:''' Reduction of a distributed optimization problem to a composite optimization problem and its solution under Polyak-Loyasievich conditions | ||

+ | * '''Authors:''' Expert A.B. Gasnikov, consultant A.N. Beznossikov | ||

+ | * '''Comment: it is important to set up a computational experiment in this The problem, otherwise The problem will be poorly compatible with the course.''' | ||

+ | |||

+ | ===17. 2020=== | ||

+ | * '''Title:''' Intention forecasting. Investigation of the properties of local models in the spatial decoding of brain signals | ||

+ | * '''Problem:''' When building brain-computer interface systems, simple, stable models are used. An important stage in the construction of such a model is the construction of an adequate feature space. Previously, such the problem was solved by extracting features from the frequency characteristics of signals. | ||

+ | * '''Data:''' ECoG/EEG brain signal data sets. | ||

+ | * '''References:''' | ||

+ | *# Motrenko A.P., Strijov V.V. Multi-way feature selection for ECoG-based brain-computer Interface // Expert systems with applications. - 2018. | ||

+ | *# Eliseyev A., Aksenova T. Stable and artifact-resistant decoding of 3D hand trajectories from ECoG signals using the generalized additive model //Journal of neural engineering. – 2014. | ||

+ | * '''Basic algorithm''': The comparison is proposed to be made with the partial least squares algorithm. | ||

+ | * '''Solution:''' In this paper, it is proposed to take into account the spatial dependence between sensors that read data. To do this, it is necessary to locally model the spatial impulse/signal and build a predictive model based on the local description. | ||

+ | * '''Novelty:''' An essentially new way of constructing a feature description in the problem of signal decoding is proposed. Bonus: analysis of changes in the structure of the model, adaptation of the structure when the sample changes. | ||

+ | * '''Authors:''' Strijov V.V., Roman Isachenko - Experts, consultants – Valery Markin, Alina Samokhina | ||

+ | |||

+ | ===9. 2020=== | ||

+ | * '''Title:''' Text recognition based on skeletal representation of thick lines and convolutional networks | ||

+ | * '''Problem:''' It is required to build two CNNs, one recognizes a raster representation of an image, the other a vector one. | ||

+ | * '''Data:''' Fonts in raster representation. | ||

+ | * '''References:'''List of works [http://www.machinelearning.ru/wiki/images/a/a2/Morozov2017Synthesis_of_medicines.pdf], in particular arXiv:1611.03199 and | ||

+ | *# Goyal P., Ferrara E. Graph embedding techniques, applications, and performance: A survey. arXiv:1705.02801, 2017. | ||

+ | *# Cai H., Zheng V.W., Chang K.C.-C. A comprehensive survey of graph embedding: Problems, techniques and applications. arXiv:1709.07604, 2017. | ||

+ | *# Grover A., Leskovec J. node2vec: Scalable Feature Learning for Networks. arXiv:1607.00653, 2016. | ||

+ | *# Mestetskiy L., Semenov A. Binary Image Skeleton - Continuous Approach // Proceedings 3rd International Conference on Computer Vision Theory and Applications, VISAPP 2008. P. 251-258. [https://www.researchgate.net/publication/221415333_Binary_Image_Skeleton_-_Continuous_Approach URL] | ||

+ | *# Kushnir O.A., Seredin O.S., Stepanov A.V. Experimental study of regularization parameters and approximation of skeletal graphs of binary images // Machine Learning and Data Analysis. 2014. Т. 1. № 7. С. 817-827. [http://jmlda.org/papers/doc/2014/no7/Kushnir2014ParametersResearch.pdf URL] | ||

+ | *# Zhukova K.V., Reyer I.A. Basic Skeleton Connectivity and Parametric Shape Descriptor // Machine Learning and Data Analysis.2014. Т. 1. № 10. С. 1354-1368. [http://jmlda.org/papers/doc/2014/no10/Reyer2014SkeletonConnectivity.pdf URL] | ||

+ | *# Kushnir O., Seredin O. Shape Matching Based on Skeletonization and Alignment of Primitive Chains // Communications in Computer and Information Science. 2015. V. 542. P. 123-136. [https://link.springer.com/chapter/10.1007/978-3-319-26123-2_12 URL] | ||

+ | * '''Basic algorithm''': Convolution network for bitmap. | ||

+ | * '''Solution:''' It is required to propose a method for collapsing graph structures, which allows generating an informative description of the thick line skeleton. | ||

+ | * '''Novelty:''' A method is proposed for improving the quality of recognition of thick lines due to a new method for generating their descriptions. | ||

+ | * '''Authors:''' Experts Reyer I.A., Strijov V.V., Mark Potanin, consultant Denis Ozherelkov | ||

+ | |||

+ | ===60. 2020=== | ||

+ | * '''Title:''' Variational optimization of deep learning models with model complexity control | ||

+ | * '''Problem:''' The problem of optimizing a deep learning model with a predetermined model complexity is considered. It is required to propose a model optimization method that allows generating new models with a given complexity and low computational costs. | ||

+ | * '''Data:'''MNIST, CIFAR | ||

+ | * '''References:''' | ||

+ | *# [1] variational inference for neural networks https://papers.nips.cc/paper/4329-practical-variational-inference-for-neural-networks.pdf | ||

+ | *# [2] hypernets https://arxiv.org/abs/1609.09106 | ||

+ | *# [3] network factories https://papers.nips.cc/paper/6304-convolutional-neural-fabrics.pdf | ||

+ | * '''Base algorithm:''' Random search | ||

+ | * '''Solution:''' The proposed method is to represent a deep learning model as a hypernet (a network that generates the parameters of another network) using a Bayesian approach. Probabilistic assumptions about the parameters of deep learning models are introduced, and a variational lower estimate of the Bayesian validity of the model is maximized. The variation estimate is considered as a conditional value depending on the external parameter of complexity. | ||

+ | * '''Novelty:''' The proposed method allows generating models in one-shot mode (practically without retraining) with the required model complexity, which significantly reduces the cost of optimization and retraining. | ||

+ | * '''Authors:''' Oleg Bakhteev, Strijov V.V. | ||

+ | |||

+ | ===61. 2020=== | ||

+ | * '''Title:''' Selecting a deep learning model based on the triplet relationship of model and sample | ||

+ | * '''Problem:''' The problem one-shot of choosing a deep learning model is considered: choosing a model for a specific sample, issued from some general population, should not be computationally expensive. | ||

+ | * '''Data:'''MNIST, synthetic data | ||

+ | * '''References:''' | ||

+ | *# [1] learning model predictions on pairs <sample, model> https://www.ri.cmu.edu/pub_files/2016/10/yuxiongw_eccv16_learntolearn.pdf | ||

+ | *# [2] Bayesian choice for two domains https://arxiv.org/abs/1806.08672 | ||

+ | * '''Base algorithm:''' Random search | ||

+ | * '''Solution:''' It is proposed to consider the space of parameters and models as two domains with their own generative models. To obtain a connection between domains, a generalization of the variational derivation to the case of triplet constraints is used. | ||

+ | * '''Novelty:''' New one-shot model training method | ||

+ | * '''Authors:''' Oleg Bakhteev, Strijov V.V. | ||

+ | |||

+ | ===64. 2020=== | ||

+ | * '''Title:''' Theoretical validity of the application of metric classification methods using dynamic alignment (DTW) to spatiotemporal objects. | ||

+ | * '''Problem description:''' It is necessary to study the existing theoretical justifications for applying dynamic alignment methods to various objects, and explore the use of such methods for space-time series.<br />When proving the applicability of alignment methods, it is proved that the function generated by the dynamic alignment algorithm is the core. Which, in turn, justifies the use of metric classification methods. | ||

+ | * '''References:''' | ||

+ | *# [https://www.cs.unm.edu/~mueen/DTW.pdf Overview presentation about DTW] | ||

+ | *# [http://www.machinelearning.ru/wiki/index.php?title=Теорема_Мерсера Mercer's theorem] | ||

+ | *# [https://www.researchgate.net/profile/Vincent_Wan/publication/221478420_Polynomial_dynamic_time_warping_kernel_support_vector_machines_for_dysarthric_speech_recognition_with_sparse_training_data/links/09e4150b7256b621ac000000/Polynomial-dynamic-time-warping-kernel-support-vector-machines-for-dysarthric-speech-recognition-with-sparse-training-data.pdf Polynomial dynamic time warping kernel support vector machines for dysarthric speech recognition with sparse training data] | ||

+ | *# [https://link.springer.com/content/pdf/10.1007/11608288_67.pdf Online Signature Verification with New Time Series Kernels for Support Vector Machines] | ||

+ | * '''Solution:''' For different formulations of the DTW method (when the internal function of the distance between time series samples is different) - find and collect evidence that the function is the kernel in one place.<br />For a basic set of datasets with time series (on which the accuracy of distance functions is checked ) check the fulfillment of the conditions from the Mercer theorem (positive definiteness of the matrix). Do this for various modifications of the DTW distance function. (Sakoe-Chiba band, Itakura band, weighted DTW.) | ||

+ | * '''Novelty:''' Investigation of theoretical justifications for applying the dynamic alignment algorithm (DTW) and its modifications to space-time series. | ||

+ | * '''Authors:''' Strijov V.V. - Expert, [[Участник:Morgachev.gleb|Gleb Morgachev]], Alexey Goncharov - consultants. | ||

+ | |||

+ | ===66. 2020=== | ||

+ | * '''Title:''' Agnostic neural networks | ||

+ | * '''Problem description:''' Introduce a metric space into the problem of automatic construction (selection) of agnostic networks. | ||

+ | * '''Data:''' Data from the Reinforcement learning area. Preferably the type of cars on the track. | ||

+ | * '''References:''' | ||

+ | *# (!) Kulunchakov A.S., Strijov V.V. Generation of simple structured Information Retrieval functions by genetic algorithm without stagnation // [http://strijov.com/papers/Kulunchakov2014RankingBySimpleFun.pdf Expert Systems with Applications, 2017, 85 : 221—230.] | ||

+ | *# A. A. Varfolomeeva The choice of features when marking bibliographic lists by methods of structural learning, 2013, [http://www.machinelearning.ru/wiki/images/f/f2/Varfolomeeva2013Diploma.pdf?format=raw] | ||

+ | *# Bin Cao, Ying Li and Jianwei Yin Measuring Similarity between Graphs Based on the Levenshtein Distance, 2012, [http://naturalspublishing.com/files/published/92cn7jm44d8wt1.pdf?format=raw] | ||

+ | *# https://habr.com/ru/post/465369/ | ||

+ | *# https://weightagnostic.github.io/ | ||

+ | * '''Base algorithm:''' Networks from an archived article. Symbolic regression from an article in ESwA (you need to restore the code). | ||

+ | * '''Solution:''' We create a model generator in the framework of symbolic regression. We create a model generator as a variational autoencoder (we won’t have time during the course). We study the metric properties of sample spaces (Euclidean) and models (Banach). We create a GAN pair - a generator-discriminator for predicting the structures of predictive models. | ||

+ | * '''Novelty:''' So far, no one has succeeded. Here they discussed Tommi Yaakkola, how he came to us in Yandex. He hasn't succeeded yet either. | ||

+ | * '''Authors:''' Expert Strijov V.V., Radoslav Neichev - consultant | ||

+ | |||

+ | ===13. 2020=== | ||

+ | * '''Title:''' Deep learning for RNA secondary structure prediction | ||

+ | * '''Problem:''' RNA secondary structure is an important feature which defines RNA functional properties. Its importance can be illustrated by the fact, that it is evolutionary preserved and some types of functional RNAs always * have the same secondary structure, for example all tRNAs fold into cloverleaf. As secondary structure often defines functions, knowing RNAs secondary structure may help investigate functions of novel RNA molecules. RNA folding is not as easy as DNA folding, because RNA is single stranded molecule which forms complicated base-pairing interactions, while DNA mostly exists as fully base paired double helices. Current methods of RNA structure prediction rely on experimentally evaluated thermodynamic rules, but with thermodynamics alone only 80% of structures can be accurately predicted. We propose an AI-driven method for predicting RNA secondary structure inspired by neural machine translation model. | ||

+ | * '''Data:''' RNA sequences in form of strings of characters | ||

+ | * '''References:''' https://arxiv.org/abs/1609.08144 | ||

+ | * '''Base algorithm:''' https://www.ncbi.nlm.nih.gov/pubmed/16873527 | ||

+ | * '''Solution:''' Deep learning recurrent encoder-decoder model with attention | ||

+ | * '''Novelty:''' Currently RNA secondary structure prediction still remains unsolved problem and to the best of our knowledge DL approach has never been introduced in the literature before | ||

+ | * '''Authors:''' consultant Maria Popova, Alexander Isaev (we are waiting for a response from them, without a response The problem is removed) | ||

+ | |||

+ | ===65. 2020=== | ||

+ | * '''Title:''' Approximation of low-dimensional samples by heterogeneous models | ||

+ | * '''Problem description:''' The problem of knowledge transfer (Hinton's distillation, Vapnik's privileged learning) from one network to another is investigated. | ||

+ | * '''Data:''' UCI samples, see what samples are used in papers on this topic | ||

+ | * '''References:''' | ||

+ | *# Neichev's Diploma [http://www.machinelearning.ru/wiki/images/3/36/NeyhevMS_Thesis.pdf Informative a priori assumptions in the privileged learning problem], [http://www.machinelearning.ru/wiki/images/1/1c/NeychevMS_Slides.pdf presentation] | ||

+ | *# Works Hinton Knowledge distilling, pay attention to error functions | ||

+ | * '''Base algorithm:''' described in the work of Neichev | ||

+ | * '''Novelty:''' Exploring different sampling methods | ||

+ | * '''Solution:'''Try different models that are in the lectures, from non-parametric to deep ones, compare and visualize the likelihood functions | ||

+ | * '''Authors:''' consultants Mark Potanin, (ask Andrey Grabovoi for help) Strijov V.V. | ||

+ | |||

+ | ===67. 2020=== | ||

+ | * '''Title:''' Selection of topics in topic models for exploratory information retrieval. | ||

+ | * '''Problem description:''' Test the hypothesis that when searching for similar documents by their topic vectors, not all topics are informative, so discarding some topics can increase the accuracy and completeness of the search. Consider the alternative hypothesis that instead of discarding topics, one can compare vectors by a weighted cosine proximity measure with adjustable weights. | ||

+ | * '''Data:''' Text collections of sites habr.com and techcrunch.com. Labeled selections: queries and related documents. | ||

+ | * '''References:''' | ||

+ | *# Vorontsov K. V. [[Media:voron17survey-artm.pdf|Probabilistic Topic Modeling: An Overview of Models and Additive Regularization]]. | ||

+ | *# Ianina A., Vorontsov K. [https://fruct.org/publications/fruct25/files/Ian.pdf Regularized Multimodal Hierarchical Topic Model for Document-by-Document Exploratory Search] // FRUCT ISMW, 2019. | ||

+ | * '''Base algorithm:''' The topic model with regularizers and modalities described in the article (source code available). | ||

+ | * '''Novelty:'''The question of informativeness of topics for vector search of thematically related documents has not been studied before. | ||

+ | * '''Solution:''' Evaluate the individual informativeness of topics by throwing them out one at a time; then sort the topics by individual informativeness and determine the threshold for cutting off non-informative topics. A suggestion as to why this should work: background themes are not informative, and discarding them increases search accuracy and recall by a few percent. | ||

+ | * '''Authors:''' [http://www.machinelearning.ru/wiki/index.php?title=Участник:Vokov Vorontsov K. V.], consultant Anastasia Yanina. | ||

+ | |||

+ | ===68. 2020=== | ||

+ | * '''Title:''' Meta-learning of topic classification models. | ||

+ | * '''Problem description:''' Develop universal heuristics for a priori assignment of modality weights in thematic models of text classification. | ||

+ | * '''Data:''' [https://docs.google.com/spreadsheets/d/1dhiz7ecgWH7lWi1wM4OkhlDI2r1D_OvcGUXaP8CDHEI/edit#gid=0 Description of datasets], [https://drive.google.com/drive/folders/1PPnw6aZOJAJoLRYuwdGm437RssV-XQx0?usp=sharing Folder with datasets]. | ||

+ | * '''References:''' | ||

+ | *# Vorontsov K. V. [[Media:voron17survey-artm.pdf|Probabilistic Topic Modeling: An Overview of Models and Additive Regularization]]. | ||

+ | * '''Base algorithm:''' Thematic classification models for several datasets. | ||

+ | * '''Novelty:'''In topic modeling, the problem of automatic selection of modality weights has not yet been solved. | ||

+ | * '''Solution:''' Optimize the weights of modalities according to the quality criterion of text classification. Investigate the dependence of the optimal relative weights of modalities on the dimensional characteristics of the problem. Find formulas for estimating the initial values of modality weights without explicitly solving the problem. To reproduce datasets, apply sampling of fragments of source documents. | ||

+ | * '''Authors:''' [http://www.machinelearning.ru/wiki/index.php?title=Участник:Vokov Vorontsov K. V.], consultant Yulian Serdyuk. | ||

+ | |||

+ | ===70. 2020=== | ||

+ | * Name: Investigation of the structure of the target space when building a predictive model | ||

+ | * The problem:The problem of forecasting a complex target variable is studied. Complexity means the presence of dependencies (linear or non-linear). It is assumed that the initial data are heterogeneous: the spaces of the independent and target variables are of different nature. It is required to build a predictive model that would take into account the dependence in the source space of the independent variable, as well as in the space of the target variable. | ||

+ | * Data: Heterogeneous data: picture - text, picture - speech and so on. | ||

+ | * Basic algorithm: As basic algorithms, it is proposed to use a linear model, as well as a nonlinear neural network model. | ||

+ | * '''Authors:''' Strijov V.V. - Expert, consultant: Isachenko Roman. | ||

+ | |||

+ | ===71. 2020=== | ||

+ | * Name: Investigation of ways to match models by reducing the dimension of space | ||

+ | * '''Problem description:''' The problem of predicting a complex target variable is investigated. Complexity means the presence of dependencies (linear or non-linear). It is proposed to study ways to take into account dependencies in the space of the target variable, as well as the conditions under which these dependencies affect the quality of the final predictive model. | ||

+ | * Data: Synthetic data with known data generation hypothesis. | ||

+ | * Basic algorithm: As basic algorithms, it is proposed to use space dimensionality reduction methods (PCA, PLS, autoencoder) and linear matching models. | ||

+ | * '''Authors:''' Strijov V.V. - Expert, consultant: Isachenko Roman. | ||

+ | |||

+ | ===72. 2020=== | ||

+ | * Name: Construction of a single latent space in the problem of modeling heterogeneous data. | ||

+ | * '''Problem description:''' The problem of predicting a complex target variable is investigated. Complexity means the presence of dependencies (linear or non-linear). It is proposed to build a single latent space for the independent and target variables. Model matching is proposed to be carried out in the resulting low-dimensional space. | ||

+ | * Data: Heterogeneous data: picture - text, picture - speech and so on. | ||

+ | * Basic algorithm: As basic algorithms, it is proposed to use space dimensionality reduction methods (PCA, PLS, autoencoder) and linear matching models. | ||

+ | * '''Authors:''' Strijov V.V. - Expert, consultant: Isachenko Roman. | ||

+ | |||

+ | ===73. 2020=== | ||

+ | * '''Title:''' Nonlinear ranking of exploratory information search results. | ||

+ | * '''Problem description:''' Develop an algorithm for recommending the reading order of documents (reading order, reading list) found using exploratory information retrieval. Documents should be ranked from simple to complex, from general to specific, that is, in the order in which it will be easier for the user to understand a new subject area for him. The algorithm must build a reading graph - a partial order relation on the set of found documents; in particular, it can be a collection of trees (document forest). | ||

+ | * '''Data:''' Part of Wikipedia and reference reading graph derived from Wikipedia categories. | ||

+ | * '''References:''' | ||

+ | *# Vorontsov K. V. [[Media:voron17survey-artm.pdf|Probabilistic Topic Modeling: An Overview of Models and Additive Regularization]]. | ||

+ | *# Georgia Koutrika, Lei Liu, and Steven Simske. [https://www.hpl.hp.com/techreports/2014/HPL-2014-5R1.pdf Generating reading orders over document collections]. HP Laboratories, 2014. | ||

+ | *# James G. Jardine. [https://www.cl.cam.ac.uk/techreports/UCAM-CL-TR-848.pdf Automatically generating reading lists]. Cambridge, 2014. | ||

+ | * '''Base algorithm:''' described in the article G.Koutrika. | ||

+ | * '''Novelty:''' The problem has been little studied in the literature. Regularized multimodal topic models (ARTM, BigARTM) have never been applied to this problem. | ||

+ | * '''Solution:''' The use of ARTM topic models in conjunction with estimates of the cognitive complexity of the text. | ||

+ | * '''Authors:''' [http://www.machinelearning.ru/wiki/index.php?title=Участник:Vokov Vorontsov K. V.], consultant Maxim Eremeev. | ||

+ | |||

+ | ==2019== | ||

+ | |||

+ | {|class="wikitable" | ||

+ | |- | ||

+ | ! Author | ||

+ | ! Topic | ||

+ | ! Links | ||

+ | ! Consultant | ||

+ | ! Reviewer | ||

+ | ! | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Severilov.pa Severilov Pavel] | ||

+ | |The problem of searching characters in texts | ||

+ | |[https://docs.google.com/document/d/1FljjnPqYXNj9u7zjLCMf8eKYcbTmsSUmZbs0BDvzI84/edit?usp=sharing LinkReview] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2019-Project-46/tree/master/code code] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2019-Project-46/raw/master/Severilov2019SymbolsInTexts/Severilov2019SymbolsInTexts.pdf paper] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2019-Project-46/raw/master/report/final_slides/Severilov_Pr46.pdf slides] [https://www.youtube.com/watch?v=vaE1vLoPFVk video] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Mapishev Murat Apishev] | ||

| | | | ||

| | | | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Grigorev.ad Grigoriev Alexey] | ||

+ | |Text recognition based on skeletal representation of thick lines and convolutional networks | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2019-Project-9/blob/master/Grigorev2019Project9/LinkReview.pdf LinkReview] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2019-Project-9/tree/master/Grigorev2019Project9/code code], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2019-Project-9/raw/master/Grigorev2019Project9/report/Image_classification_based_on_skeletonization_and_Graph_NN.pdf paper], [https://github.com/Intelligent-Systems-Phystech/2019-Project-9/raw/master/Grigorev2019Project9/report/skeletons_presentation.pdf slides] | ||

+ | [https://www.youtube.com/watch?v=j0I1w8htPZA video] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Ilyazharikov Ilya Zharikov] | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2019-Project-9/raw/master/Grigorev2019Project9/report/Grigorev_review.docx review] [http://www.machinelearning.ru/wiki/index.php?title=Участник:Varenik.nv Varenyk Natalia] | ||

| | | | ||

|- | |- | ||

- | | | + | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Grishanov Grishanov Alexey] |

+ | |Automatic configuration of BigARTM parameters for a wide class of The problems | ||

+ | |[https://docs.google.com/document/d/1UFvURCZloCHlnLTTJmpXFr_-GWCo4t8fTOJl4FygtJk/edit?usp=sharing LinkReview] [https://github.com/Intelligent-Systems-Phystech/2019-Project-4/tree/master/code code], [https://github.com/Intelligent-Systems-Phystech/2019-Project-4/raw/master/Grishanov2019Project4/Grishanov2019Project4.pdf paper][https://github.com/Intelligent-Systems-Phystech/2019-Project-4/raw/master/report/Grishanov2019Presentation.pdf slides] | ||

+ | [https://www.youtube.com/watch?v=OVGUuHUvNjc video] | ||

+ | |Viktor Bulatov | ||

+ | |[https://github.com/Nikolay-Gerasimenko/Experiment/raw/master/Рецензия%20на%20рукопись.docx review][http://www.machinelearning.ru/wiki/index.php?title=Участник:Nikolay-Gerasimenko Gerasimenko Nikolay] | ||

| | | | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Yusupov_igor Yusupov Igor] | ||

+ | |Dynamic alignment of multivariate time series | ||

+ | |[https://docs.google.com/document/d/1RHAdwtvDZU5JS6cTVKEWYkSI-6KgwDd3aefpBAw9Ujw/edit LinkReview] code [https://github.com/igor-yusupov/2018-Project-3/raw/patch-1/Yusupov2019Title/Yusupov2019.pdf paper] [https://github.com/igor-yusupov/2018-Project-3/raw/patch-1/Yusupov2019Title/presentation.pdf slides] [https://www.youtube.com/watch?v=wtnGACpmU8k video] | ||

+ | |Alexey Goncharov | ||

| | | | ||

| | | | ||

|- | |- | ||

- | | | + | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Varenik.nv Varenyk Natalia] |

+ | |Spherical CNN for QSAR prediction | ||

+ | |[https://docs.google.com/document/d/13L7JHa3H19lSuJKRgq2novuzSnaMv3MpwwGpcW5rRZc/edit LinkReview], [https://github.com/Natalia-Varenik/s2cnn code], [https://github.com/Intelligent-Systems-Phystech/2019-Project-47/raw/master/Varenik2019Project47/Varenik2019Project47.pdf paper], [https://github.com/Intelligent-Systems-Phystech/2019-Project-47/raw/master/report/Varenik2019Project47Presentation.pdf slides] [https://www.youtube.com/watch?v=0kJW898HPqM video] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Mpopova Maria Popova] | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2019-Project-47/raw/master/report/review.pdf review] [http://www.machinelearning.ru/wiki/index.php?title=Участник:Grigorev.ad Grigoriev Alexey] | ||

| | | | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Beznosikov.an Beznosikov Alexander] | ||

+ | |Z-learning of linearly-solvable Markov Decision Processes | ||

+ | |[https://docs.google.com/document/d/1Ef25ueOxzBkbcAFV24fuCEHAApwxspGRAPq_r2hw0EM/edit?usp=sharing LinkReview] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2019-Project-42/raw/master/Beznosikov2019Z-learning%20of%20linearly-solvable%20Markov%20Decision%20Processes/Beznosikov2019Z_learning_of_linearly_solvable_Markov_Decision_Processes.pdf paper] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2019-Project-42/tree/master/code code] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2019-Project-42/raw/master/report/Beznosikov.pdf slides] [https://www.youtube.com/watch?v=--rEzR4VGKg video] | ||

+ | |Yury Maximov | ||

| | | | ||

| | | | ||

|- | |- | ||

- | | | + | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:PanchenkoSviatoslav Panchenko Svyatoslav] |

+ | |Obtaining a simple sample at the output of the neural network layer | ||

+ | |[https://docs.google.com/document/d/1CPgyqyaM4pv_6jxFio5NwU_Ncgu6tazFxl_jgH4gSWQ/edit?usp=sharing LinkReview], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2019-Project-43/tree/master/code code], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2019-Project-43/raw/master/Panchenko2019Project43/Panchenko2019Project43.pdf paper], slides | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Tamaz Gadaev Tamaz] | ||

| | | | ||

+ | | | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:VeselovaER Veselova Evgeniya] | ||

+ | |Deep Learning for reliable detection of tandem repeats in 3D protein structures | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2019-Project-14 Code] [https://docs.google.com/document/d/1_BtCiAihPg9ON-2PlxORkcmwL80pgqC4gOE7A03rQjg link review] [https://github.com/Intelligent-Systems-Phystech/2019-Project-14/raw/master/Veselova2019Project14/Veselova2019Project14.pdf paper] [https://github.com/Intelligent-Systems-Phystech/2019-Project-14/raw/master/Veselova2019Project14/Veselova2019Slides.pdf slides] [https://www.youtube.com/watch?v=XGLT5BGYTek video] | ||

+ | |Guillaume Pages, Sergei Grudinin | ||

| | | | ||

| | | | ||

|- | |- | ||

- | | | + | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Aminov.tv Aminov Timur] |

+ | |Quality Prediction for a Feature Selection Procedure | ||

+ | |[https://docs.google.com/document/d/1HLo0fNei0KoTrFQNgkdubFCM39PRpEYOyeF1WilibpY/edit LinkReview] code [https://github.com/Intelligent-Systems-Phystech/2019-Project-40/raw/master/doc/Aminov2019FSPP.pdf paper] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2019-Project-40/raw/master/doc/pres%20(1).pdf slides] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Isachenkoroma Roman Isachenko] | ||

| | | | ||

+ | | | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Vmarkin Markin Valery] | ||

+ | |Investigation of the properties of local models in the spatial decoding of brain signals | ||

+ | |[https://docs.google.com/document/d/17rXnTPT9M6nYEkoxwfv5XDE8LIBt-mR1wv2vzrQSljw/edit?usp=sharing LinkReview] | ||

+ | [https://github.com/Intelligent-Systems-Phystech code] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/ECoG_Project/raw/master/Markin2019SpatialDecoding.pdf paper] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/ECoG_Project/raw/master/Markin2019Slides.pdf slides] [https://www.youtube.com/watch?v=l_4AJ-Xb5cs video] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Isachenkoroma Roman Isachenko] | ||

| | | | ||

| | | | ||

|- | |- | ||

- | | | + | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Sadiev1998 Abdurahmon Sadiev] |

+ | |Generation of features using locally approximating models | ||

+ | |[https://docs.google.com/document/d/1A_rWU-2DnvD3ZVCOPLQcAEqB3Iw2YyWOqb9YspByh9o/edit LinkReview] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2019-Project-8/tree/master/code code], [https://github.com/Intelligent-Systems-Phystech/2019-Project-8/raw/master/paper/Feature_gen.pdf paper], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2019-Project-8/raw/master/slides_Sadiev.pdf slides] [https://www.youtube.com/watch?v=bDpvKQRZA7w video] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Anastasiya Anastasia Motrenko] | ||

| | | | ||

+ | | | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Tagirschik Tagir Sattarov] | ||

+ | |Machine translation training without parallel texts. | ||

+ | |[https://docs.google.com/document/d/1ORgDN1bVeIduWTdcmjl9R346MNIgpe0_T3G-aUtrxlo/edit?usp=sharing LinkReview] [https://github.com/Intelligent-Systems-Phystech/2019-project-12/blob/master/monogolingual_mt_example.ipynb code] [https://github.com/Intelligent-Systems-Phystech/2019-project-12/blob/master/paper.pdf paper], [https://github.com/Intelligent-Systems-Phystech/2019-project-12/raw/master/Sattarov_presentation.pdf slides] [https://www.youtube.com/watch?v=wduZgu6ym-0 video] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Oleg_Bakhteev Oleg Bakhteev] | ||

| | | | ||

| | | | ||

|- | |- | ||

- | | | + | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Nikolay-Gerasimenko Gerasimenko Nikolay] |

+ | |Thematic search for similar cases in the collection of acts of arbitration courts. | ||

+ | |[https://docs.google.com/document/d/1D1fOYNCne6sU5oqgET4s9WKmSj84-Ra8pSRKoi215kc/edit?usp=sharing LinkReview] [https://github.com/Intelligent-Systems-Phystech/2019-Project-50/tree/master/code code] [https://github.com/Intelligent-Systems-Phystech/2019-Project-50/raw/master/Gerasimenko2019Project50/Russian/Gerasimenko2019Project50.pdf paper] [https://github.com/Intelligent-Systems-Phystech/2019-Project-50/raw/master/report/Gerasimenko2019Project50Presentation.pdf slides] [https://www.youtube.com/watch?v=EhgQexs2yIQ video] | ||

+ | |Ekaterina Artyomova | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2019-Project-50/raw/master/Gerasimenko2019Project50/Russian/Review.docx review][http://www.machinelearning.ru/wiki/index.php?title=Участник:Grishanov Grishanov Alexey] | ||

| | | | ||

+ | |} | ||

+ | |||

+ | ===40. 2019=== | ||

+ | * '''Title:''' Quality prediction for the feature selection procedure. | ||

+ | * '''Problem description:''' The solution of the feature selection problem is reduced to enumeration of binary cube vertices. This procedure cannot be performed for a sample with a large number of features. It is proposed to reduce this problem to optimization in a linear space. | ||

+ | * '''Data:''' Synthetic data + simple samples | ||

+ | * '''References:''' | ||

+ | *# Bertsimas D. et al. Best subset selection via a modern optimization lens //The annals of statistics. – 2016. – Т. 44. – №. 2. – С. 813-852. | ||

+ | *# Luo R. et al. Neural architecture optimization //Advances in Neural Information Processing Systems. – 2018. – С. 7827-7838. | ||

+ | * '''Base algorithm:''' Popular feature selection methods. | ||

+ | * '''Solution:''' In this paper, it is proposed to build a model that, based on a set of features, predicts the quality on a test sample. To do this, a mapping of a binary cube into a linear space is constructed. After that, the quality of the model in linear space is maximized. To reconstruct the solution of the problem, the model of inverse mapping into a binary cube is used. | ||

+ | * '''Novelty:''' A constructively new approach to solving the problem of choosing models is proposed. | ||

+ | * '''Authors:''' Strijov V.V., Tetiana Aksenova, consultant – Roman Isachenko | ||

+ | |||

+ | ===42. 2019=== | ||

+ | * '''Title:''' Z-learning of linearly-solvable Markov Decision Processes | ||

+ | * '''Problem:''' Adapt Z-learning from [1] to the case of Markov Decision Process discussed in [2] in the context of energy systems. Compare it with standard (in reinforcement learning) Q-learning. | ||

+ | * '''Data:''' We consider a Markov Process described via transition probability matrix. Given initial state vector (probability of being in a state at time zero), we generate data for the time evolution of the state vector. See [2] for an exemplary process describing evolution of an ensemble of energy consumers. | ||

+ | * '''References:''' | ||

+ | *# E. Todorov. Linearly-solvable Markov decision problems https://homes.cs.washington.edu/~todorov/papers/TodorovNIPS06.pdf | ||

+ | *# Ensemble Control of Cycling Energy Loads: Markov Decision Approach. Michael Chertkov, Vladimir Y. Chernyak, Deepjyoti Deka. https://arxiv.org/abs/1701.04941 | ||

+ | *# Csaba Szepesvári. Algorithms for Reinforcement Learning. https://sites.ualberta.ca/~szepesva/papers/RLAlgsInMDPs.pdf | ||

+ | * '''Base algorithm:''' Principal comparison should be made with Q learning described in [3] | ||

+ | * '''Solution:''' We suppose that plugging in algorithm from [1] directly into [2] gives faster and more reliable solution. | ||

+ | * '''Novelty:''' In the area of power systems there is a huge demand on fast reinforcement learning algorithms, but there is still a lack of that (in particular the ones respect the physics/underlying graph) | ||

+ | * '''Authors:''' Yury Maximov (consultant, expert), Michael Chertkov (expert) | ||

+ | |||

+ | ===1. 2019=== | ||

+ | * '''Title:''' Forecasting the direction of movement of the price of exchange instruments according to the news flow. | ||

+ | * '''Problem description:''' Build and explore a model for predicting the direction of price movement. Given a set of news S and a set of timestamps T corresponding to the time of publication of news from S. 2. Time series P, corresponding to the price of an exchange instrument, and time series V, corresponding to the volume of sales for this instrument, for a period of time T'. 3. The set T is a subset of the time period T'. 4. Time intervals w=[w0, w1], l=[l0, l1], d=[d0, d1], where w0 < w1=l0 < l1=d0 < d1. It is required to predict the direction of movement of the price of an exchange instrument at the time t=d0 according to the news released in the period w. | ||

+ | * '''Data:''' | ||

+ | *# Financial data: data on quotes (at one tick interval) of several financial instruments (GAZP, SBER, VTBR, LKOH) for the 2nd quarter of 2017 from the Finam.ru website; for each point of the series, the date, time, price and volume are known. | ||

+ | *# Text data: economic news for the 2nd quarter of 2017 from Forexis; each news is a separate html file. | ||

+ | * '''References:''' | ||

+ | *# Usmanova K.R., Kudiyarov S.P., Martyshkin R.V., Zamkovoy A.A., Strijov V.V. Analysis of relationships between indicators in forecasting cargo transportation // Systems and Means of Informatics, 2018, 28(3). | ||

+ | *# Kuznetsov M.P., Motrenko A.P., Kuznetsova M.V., Strijov V.V. Methods for intrinsic plagiarism detection and author diarization // Working Notes of CLEF, 2016, 1609 : 912-919. | ||

+ | *# Aysina Roza Munerovna, Thematic modeling of financial flows of corporate clients of a bank based on transactional data, final qualification work. | ||

+ | *# Lee, Heeyoung, et al. "On the Importance of Text Analysis for Stock Price Prediction." LREC. 2014. | ||

+ | * '''Base algorithm:''' Method used in the article (4). | ||

+ | * '''Solution:''' Using topic modeling (ARTM) and local approximation models to translate a sequence of texts corresponding to different timestamps into a single feature description. Quality criterion: F1-score, ROC AUC, profitability of the strategy used. | ||

+ | * '''Novelty:''' To substantiate the connection of time series, the Converging cross-mapping method is proposed. | ||

+ | * '''Authors:''' Ivan Zaputlyaev (consultant), Strijov V.V., K.V. Vorontsov (Experts) | ||

+ | |||

+ | ===3. 2019=== | ||

+ | * '''Title:''' Dynamic alignment of multidimensional time series. | ||

+ | * '''Problem description:''' A characteristic multidimensional time series is the trajectory of a point in 3-dimensional space. The two trajectories need to be optimally aligned with each other. For this, the distance DTW between two time series is used. In the classical representation, DTW is built between one-dimensional time series. It is necessary to introduce various modifications of the algorithm for working with high-dimensional time series: trajectories, corticograms. | ||

+ | * '''Data:''' The data describes 6 classes of time series from the mobile phone's accelerometer. https://sourceforge.net/p/mlalgorithms/code/HEAD/tree/Group274/Goncharov2015MetricClassification/data/ | ||

+ | * '''References:''' | ||

+ | *# Multidimensional DTW: https://pdfs.semanticscholar.org/76d3/5bd5a52453ebde80faaa1467d7effd74426f.pdf | ||

+ | * '''Base algorithm:''' Using L_p distances between two dimensions of a time series, their modifications. | ||

+ | * '''Solution:''' Investigation of distances resistant to change of coordinate order, studies of distances unstable to change of coordinate order. Experiments with other types of distances (cosine, RBF, others). | ||

+ | * '''Novelty:''' There is no complete review and study of methods for working with multivariate time series. The dependence of the quality of the solution on the selected distances between measurements has not been studied. | ||

+ | * '''Authors:''' Alexey Goncharov - consultant, Expert, Strijov V.V. - Expert | ||

+ | |||

+ | ===43. 2019=== | ||

+ | * '''Title:''' Getting a simple sample at the output of the neural network layer | ||

+ | * '''Problem:''' The output of the neural network is usually a generalized linear model over the outputs of the penultimate layer. It is necessary to propose a way to test the simplicity of the sample and its compliance with the generalized linear model (linear regression, logistic regression) using a system of statistical criteria. | ||

+ | * '''Data:''' For the computational experiment, it is proposed to use classical samples from the UCI repository. Link to samples https://github.com/ttgadaev/SampleSize/tree/master/datasets | ||

+ | * '''References:''' http://www.ccas.ru/avtorefe/0016d.pdf c 49-63 Bishop, C. 2006. Pattern Recognition and Machine Learning. Berlin: Springer. $758 | ||

+ | * '''Base algorithm:''' White test, Wald test, Goldfeld-Quantum test, Durbin-Watson, Chi-square, Fry-Behr, Shapiro-Wilk | ||

+ | * '''Solution:''' The system of tests for checking the simplicity of the sample (and the adequacy of the model), the independent variables are not random, the dependent variables are distributed normally or binomially, there are no gaps and outliers, the classes are balanced, the sample is approximated by a single model. The variance of the error function does not depend on the independent variable. The study is based on synthetic and real data. | ||

+ | * '''Authors:''' Gadaev T. T. (consultant) Strijov V.V., Grabovoi A.V. (Experts) | ||

+ | |||

+ | ===14. 2019=== | ||

+ | * '''Title:''' Deep Learning for reliable detection of tandem repeats in 3D protein structures [[Media:Strijov_3D_CNN.pdf|more in PDF]] | ||

+ | * '''Problem:''' Deep learning algorithms pushed computer vision to a level of accuracy comparable or higher than a human vision. Similarly, we believe that it is possible to recognize the symmetry of a 3D object with a very high reliability, when the object is represented as a density map. The optimization problem includes i) multiclass classification of 3D data. The output is the order of symmetry. The number of classes is ~10-20 ii) multioutput regression of 3D data. The output is the symmetry axis (a 3-vector). The input data are typically 24x24x24 meshes. The total amount of these meshes is of order a million. Biological motivation : Symmetry is an important feature of protein tertiary and quaternary structures that has been associated with protein folding, function, evolution, and stability. Its emergence and ensuing prevalence has been attributed to gene duplications, fusion events, and subsequent evolutionary drift in sequence. Methods to detect these symmetries exist, either based on the structure or the sequence of the proteins, however, we believe that they can be vastly improved. | ||

+ | * '''Data:''' Synthetic data are obtained by ‘symmetrizing’ folds from top8000 library (http://kinemage.biochem.duke.edu/databases/top8000.php). | ||

+ | * '''References:''' Our previous 3D CNN: [https://arxiv.org/abs/1801.06252] Invariance of CNNs (and references therein): [https://hal.inria.fr/hal- 01630265/document], [https://arxiv.org/pdf/1706.03078.pdf] | ||

+ | * '''Basic algorithm''': A prototype has already been created using the Tensorflow framework [4], which is capable of detecting the order of cyclic structures with about 93% accuracy. The main goal of this internship is to optimize the topology of the current neural network prototype and make it rotational and translational invariant with respect to input data. [4] [https://www.tensorflow.org/] | ||

+ | * '''Solution:''' The network architecture needs to be modified according to the invariance properties (most importantly, rotational invariance). Please see the links below [https://hal.inria.fr/hal-01630265/document], [https://arxiv.org/pdf/1706.03078.pdf] The code is written using the Tensorflow library, and the current model is trained on a single GPU (Nvidia Quadro 4000)of a desktop machine. | ||

+ | * '''Novelty:''' Applications of convolutional networks to 3D data are still very challenging due to large amount of data and specific requirements to the network architecture. More specifically, the models need to be rotationally and transnationally invariant, which makes classical 2D augmentation tricks loosely applicable here. Thus, new models need to be developed for 3D data. | ||

+ | * '''Authors:''' Expert Sergei Grudinin, consultants Guillaume Pages | ||

+ | |||

+ | ===46. 2019=== | ||

+ | * Name: The problem of searching characters in texts | ||

+ | * '''Problem description:''' In the simplest case, this The problem is reduced to the Sequence Labeling The problem on a labeled selection. The difficulty lies in obtaining a sufficient amount of training data, that is, it is required to obtain a larger sample from the existing small Expert markup (automatically by searching for patterns or by compiling a simple and high-quality markup instruction, for example, in Toloka). The presence of markup allows you to start experimenting with the selection of the optimal model, various neural network architectures (BiLSTM, Transformer, etc.) may be of interest here. | ||

+ | * Data: Dictionary of symbols, Marked artistic texts | ||

+ | * '''References:''' http://www.machinelearning.ru/wiki/images/0/05/Mmta18-rnn.pdf | ||

+ | * Basic algorithm: HMM, RNN | ||

+ | * '''Solution:''' It is proposed to compare the work of several state-of-the-art algorithms. Propose a classifier quality metric for characters (character/non-character). Determine applicability of methods. | ||

+ | * '''Novelty:''' The proposed approach to text analysis is used by Experts in manual mode and has not been automated | ||

+ | * '''Authors:''' M. Apishev (consultant), D. Lemtyuzhnikova | ||

+ | |||

+ | ===47. 2019=== | ||

+ | * '''Title:''' Deep learning for RNA secondary structure prediction | ||

+ | * '''Problem:''' RNA secondary structure is an important feature which defines RNA functional properties. Its importance can be illustrated by the fact, that it is evolutionary preserved and some types of functional RNAs always * have the same secondary structure, for example all tRNAs fold into cloverleaf. As secondary structure often defines functions, knowing RNAs secondary structure may help investigate functions of novel RNA molecules. RNA folding is not as easy as DNA folding, because RNA is single stranded molecule which forms complicated base-pairing interactions, while DNA mostly exists as fully base paired double helices. Current methods of RNA structure prediction rely on experimentally evaluated thermodynamic rules, but with thermodynamics alone only 80% of structures can be accurately predicted. We propose an AI-driven method for predicting RNA secondary structure inspired by neural machine translation model. | ||

+ | * '''Data:''' RNA sequences in form of strings of characters | ||

+ | * '''References:''' https://arxiv.org/abs/1609.08144 | ||

+ | * '''Base algorithm:''' https://www.ncbi.nlm.nih.gov/pubmed/16873527 | ||

+ | * '''Solution:''' Deep learning recurrent encoder-decoder model with attention | ||

+ | * '''Novelty:''' Currently RNA secondary structure prediction still remains unsolved problem and to the best of our knowledge DL approach has never been introduced in the literature before | ||

+ | * '''Authors:''' consultant Maria Popova Chapel-Hill | ||

+ | |||

+ | ===4. 2019=== | ||

+ | * '''Title:''' Automatic setting of ARTM parameters for a wide class of The problems. | ||

+ | * '''Problem description:''' The bigARTM open library allows you to build topical models using a wide class of possible regularizers. However, this flexibility makes The problem of setting the coefficients very difficult. This tuning can be greatly simplified by using the relative regularization coefficients mechanism and automatic selection of N-grams. We need to test the hypothesis that there is a universal set of relative regularization coefficients that gives "reasonably good" results on a wide class of problems. Several datasets are given with some external quality criterion (for example, classification of documents into categories or ranking). We find the best parameters for a particular dataset, giving the "locally the best model". We find the bigARTM initialization algorithm that produces thematic models with quality comparable to the "locally best model" on its dataset. Comparability criterion in quality: on this dataset, the quality of the "universal model" is no more than 5% worse than that of the "locally best model". | ||

+ | *'''Data:''' [https://archive.ics.uci.edu/ml/datasets/Victorian+Era+Authorship+Attribution Victorian Era Authorship Attribution Data Set], [https://archive.ics. uci.edu/ml/datasets/Twenty+Newsgroups 20 Newsgroups], ICD-10, search/ranking triplets. | ||

+ | * '''References:''' | ||

+ | *# WRC by Nikita Doykov: http://www.machinelearning.ru/wiki/images/9/9f/2015_417_DoykovNV.pdf | ||

+ | *# Presentation by Viktor Bulatov at a scientific seminar: https://drive.google.com/file/d/19pJ21LRPeeOxY4mkcSnQCRm93zOO4J5b/view | ||

+ | *# Draft with formulas: https://drive.google.com/open?id=1AqS7snUsSJ18ZYBtC-6uP_2dMTDJSGeD | ||

+ | * '''Base algorithm:''' PLSA / LDA / logregression. | ||

+ | * '''Solution:''' bigARTM with background themes and smoothing, sparseness and decorrelation regularizers (coefficients picked up automatically), as well as automatically selected N-grams. | ||

+ | * '''Novelty:''' The need for automated tuning of model parameters and the lack of such implementations in the scientific community. | ||

+ | * '''Authors:''' consultant Viktor Bulatov, Expert[http://www.machinelearning.ru/wiki/index.php?title=Участник:Vokov Vorontsov K. V.]. | ||

+ | |||

+ | ===50. 2019=== | ||

+ | * '''Title:''' Thematic search for similar cases in the collection of acts of arbitration courts. | ||

+ | * '''Problem description:''' Build an information retrieval algorithm for a collection of acts of arbitration courts. The request can be an arbitrary document of the collection (the text of the act). The search result should be a list of documents in the collection, ranked in descending order of relevance. | ||

+ | *'''Data:''' collection of text documents — acts of arbitration courts http://kad.arbitr.ru. | ||

+ | * '''References:''' | ||

+ | *# Anastasia Yanina. [[Media:ianina18msc.pdf|Thematic exploratory information search]]. 2018. FIVT MIPT. | ||

+ | *# Ianina A., Golitsyn L., Vorontsov K. [[Media:ianina17exploratory.pdf|Multi-objective topic modeling for exploratory search in tech news]]. AINL-2017. CCIS, Springer, 2018. | ||

+ | *# Ahmed El-Kishky, Yanglei Song, Chi Wang, Clare Voss, Jiawei Han. [http://hanj.cs.illinois.edu/pdf/vldb15_ael-kishky.pdf Scalable Topical Phrase Mining from Text Corpora]. 2015. | ||

+ | * '''Base algorithm:''' BigARTM with decorrelation, smoothing, sparse regularizers. Search by TF-IDF of words, by TF-IDF of UPA links, by thematic vector representations of documents, using a cosine proximity measure. TopMine algorithm for collocation detection. | ||

+ | * '''Solution:''' Add modality of links to legal acts. Add modality of legal terms. Choose the optimal number of topics and regularization strategy. Organize the process of marking pairs of documents. Implement the evaluation of the quality of the search for a labeled sample of pairs of documents. | ||

+ | * '''Novelty:''' The first attempt to use ARTM for thematic search of legal texts. | ||

+ | * '''Authors:''' consultant Ekaterina Artyomova, Expert [http://www.machinelearning.ru/wiki/index.php?title=Участник:Vokov Vorontsov K. V.]. | ||

+ | |||

+ | ==2019 Group 2== | ||

+ | {|class="wikitable" | ||

+ | |- | ||

+ | ! Author | ||

+ | ! Topic | ||

+ | ! Links | ||

+ | ! Consultant | ||

+ | ! Reviewer | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Ninavishn Vishnyakova Nina] | ||

+ | |Optimal Approximation of Non-linear Power Flow Problem | ||

+ | |[https://docs.google.com/document/d/1TvMgA1ytOMrCm1Fx35UsrnMSASvECnr249x0Nvy7TaY/edit LinkReview] [https://github.com/Intelligent-Systems-Phystech/2019-Project-41/raw/master/report/Optimal_Approximation_of_Non_linear_Power_Flow_Problem.pdf paper] [https://github.com/Intelligent-Systems-Phystech/2019-Project-41 code] [https://github.com/Intelligent-Systems-Phystech/2019-Project-41/raw/master/report/Vishnyakova_nina_2019_41_Talk.pdf presentation] [https://youtu.be/QINA00S1_Bo video] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Yury.maximov Yury Maximov] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Loginov-ra Loginov Roman] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2019-Project-41/raw/master/report/Vishnyakova2019Project41_Review.pdf review] | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Polinakud Kudryavtseva Polina] | ||

+ | |Intention forecasting. Building an optimal signal decoding model for modeling a brain-computer interface. | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2019-Project-18/tree/master/code code] | ||

+ | [https://docs.google.com/document/d/1sRFisJeQ7QVNtlIh7k1CX47bAk7peuneiPZRxHeFigM/edit?usp=sharing LinkReview] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2019-Project-18/raw/master/doc/Kudryavtseva2019Project18.pdf paper] [https://www.youtube.com/watch?v=wo-nJU3uG1I video] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2019-Project-18/raw/master/doc/Kudryavtseva2019Slides.pdf presentation] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Isachenkoroma Roman Isachenko] | ||

+ | |Nechepurenko Ivan | ||

+ | [https://docs.google.com/document/d/1i6WuDNEozojFYMkJHu5DcaItE5qrsr_Tt3ubBE298DQ/edit review] | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Loginov-ra Loginov Roman] | ||

+ | |Multi-simulation as a universal way to describe a general sample | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2019-Project-28/tree/master/code code] | ||

+ | [https://docs.google.com/document/d/1cCEttJpkGTtB10QieS2TWHI0COv_BUKgCckd4refcFE/edit LinkReview] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2019-Project-28/raw/master/Loginov2019MultimodellingAdaptive/Loginov2019MultimodellingAdaptive.pdf paper] | ||

+ | [https://hangouts.google.com/group/rRyggcQjYKF81nrE2 ChatInvite] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2019-Project-28/raw/master/report/Loginov2019MultimodellingTime.pdf presentation] | ||

+ | [https://www.youtube.com/watch?v=GCl7VSAz-Xg video] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Aduenko Alexander Aduenko] | ||

+ | |Makarov Mikhail [http://www.machinelearning.ru/wiki/images/9/92/Loginov2019Project28_Review.rtf review] | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Makarov.mv Mikhail Makarov] | ||

+ | |Location determination by accelerometer signals | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-26/tree/master/code code] | ||

+ | [https://docs.google.com/document/d/1er3SgPu9bBBWkLk1yVev-9Ue42BOPapOkLn6sL0GAGA/edit LinkReview] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-26/raw/master/Makarov2019Project26/Makarov2019Project26.pdf paper] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-26/raw/master/pres/Project26presentation.pdf presentation] | ||

+ | [https://www.youtube.com/watch?v=OEe9xmoNUNQ video] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Anastasiya Anastasia Motrenko] | ||

+ | |Cherepkov Anton: [https://github.com/Intelligent-Systems-Phystech/2018-Project-26/raw/master/Makarov2019Project26/Makarov2019_review.pdf review] | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Alex-kozinov Kozinov Alexey] | ||

+ | |The problem of finding characters in images | ||

+ | |[https://docs.google.com/document/d/1P_osIW236MTBPe_aMJUI-EEHgUhheQR9bqlKCN97e8M/edit?usp=sharing LinkReview] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2019-Project-45/raw/master/Kozinov2019Project45/Kozinov2019Project45.pdf paper] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2019-Project-45/tree/master/code code] | ||

+ | | M. Apishev, | ||

+ | D. Lemtyuzhnikova | ||

+ | |Gracheva Anastasia [https://github.com/Intelligent-Systems-Phystech/2019-Project-15/raw/master/review.pdf review] | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Buchnev.valentin Buchnev Valentin] | ||

+ | |Early prediction of sufficient sample size for a generalized linear model. | ||

+ | |[https://docs.google.com/document/d/1-xpsWSbI-hlX8PQXdVZ5gMOQC03LH0oM8u4dpTDMSKs/edit?usp=sharing LinkReview] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2019-Project-44/raw/master/Buchnev2019Project44/Buchnev2019Project44.pdf paper] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2019-Project-44/ code] [https://github.com/Intelligent-Systems-Phystech/2019-Project-44/raw/master/report/Buchnev2019Project44presentation.pdf presentation] | ||

+ | [https://www.youtube.com/watch?v=0SJL6Xx5VnU video] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Andriygav Grabovoi Andrey] | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Ivan.nechepurenco Nechepurenko Ivan] | ||

+ | |Multisimulation, privileged training | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2019-Project-48/tree/master/code code], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2019-Project-48/raw/master/Nechepurenco2019Project48/Nechepurenco2019Project48.pdf paper], | ||

+ | [https://docs.google.com/document/d/1DJNwFfFXCipPictxTUWd8dBfj_Zv6zrfp86L5p_cfTI/edit?usp=sharing LinkReview] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2019-Project-48/raw/master/slides/Nechepurenco2019.pdf presentation] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Neychev R. G. Neichev] | ||

+ | |Kudryavtseva Polina | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Gracheva.as Gracheva Anastasia] | ||

+ | |Estimation of binding energy of protein and small molecules | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2019-Project-15 code] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2019-Project-15/raw/master/Gracheva2019Project15/Gracheva2019Title.pdf paper] | ||

+ | [https://docs.google.com/document/d/1INJAFAXNjEyvqDME6KiGiCnRJ6qQ9b_3dM_fzePgU7U/edit LinkReview] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2019-Project-15/raw/master/presentations/Gracheva_presentation.pdf presentation] | ||

+ | [https://www.youtube.com/watch?v=smj4XwMnE-4 video] | ||

+ | |Sergei Grudinin, | ||

+ | Maria Kadukova | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Anthonycherepkov Cherepkov Anton] | ||

+ | |Privileged learning in the problem of iris boundary approximation | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2019-Project-7/raw/master/report/Cherepkov_2019_Iris_circle_problem.pdf paper], [https://github.com/Intelligent-Systems-Phystech/2019-Project-7/raw/master/slides/Cherepkov_2019_Iris_circle_problem.pdf slides], [https://github.com/Intelligent-Systems-Phystech/2019-Project-7/tree/master/code code], [https://docs.google.com/document/d/140k6Qrf63iOHUqHcG9IO8cCa1PXEypY5zgboQ3S0LoU/edit?usp=sharing LinkReview] | ||

+ | [https://www.youtube.com/watch?v=cI3x-vjOAIo video] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Neychev R. G. Neichev] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Mlepekhin Lepekhin Mikhail] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2019-Project-7/raw/master/review/Cherepkov2019_review.pdf preliminary review] | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Mlepekhin Lepekhin Mikhail] | ||

+ | |Creation of ranking models for information retrieval systems. Algorithm for Predicting the Structure of Locally Optimal Models | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2019-Project-27/blob/master/code code] | ||

+ | [https://docs.google.com/document/d/1oVnIwD6T1VEegE1Pieo8-b5JyBPbIzrh0Cdk3V-BlO4/edit?usp=sharing LinkReview] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2019-Project-27/raw/master/Lepekhin2019Project27/Lepekhin2019Project27.pdf paper] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2019-Project-27/raw/master/slides/report.pdf presentation] | ||

+ | [https://www.youtube.com/watch?v=AL6Q7u3daPw video] | ||

+ | |Andrey Kulunchakov | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Ninavishn Vishnyakova Nina], [https://github.com/Intelligent-Systems-Phystech/2019-Project-41/raw/master/report/Рецензия%20на%20статью%20Лепехина%20Михаила.pdf review] | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Gridasovii Gridasov Ilya] | ||

+ | |Automatic construction of a neural network of optimal complexity | ||

+ | |[https://docs.google.com/document/d/1RcUfc9dKu-hO9r9sqS9hXUu7QofHeDfvHTuJqM8BgU4/edit?usp=sharing LinkReview] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2019-Project-11/raw/master/Gridasov2019Project11/paper/Gridasov2019Project11.pdf paper] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2019-Project-11/raw/master/Gridasov2019Project11/presentation/Gridasov2019Project11Presentation.pdf Presentation] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2019-Project-11/tree/master/Gridasov2019Project11/code code] | ||

+ | |O. Yu. Bakhteev, Strijov V.V. | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Buchnev.valentin Buchnev Valentin] | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Telenkov-Dmitry Telenkov Dmitry] | ||

+ | |Brain signal decoding and intention prediction | ||

+ | |[https://docs.google.com/document/d/1pTzCafRueWf1hTYCY2uwatNEAFia_nbZSlsgYGYoWnY LinkReview] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2019-Project-49 git] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2019-Project-49/raw/master/Telenkov2019Article49/Telenkov2019Article49.pdf The paper] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2019-Project-49/raw/master/report/Presentation.pdf Presentation] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2019-Project-49/blob/master/report/Experiment.ipynb code] | ||

+ | |Andrey Zadayanchuk | ||

+ | |- | ||

+ | |} | ||

+ | |||

+ | ===18. 2019=== | ||

+ | * '''Title:''' Forecasting intentions. Building an optimal signal decoding model for modeling a brain-computer interface. | ||

+ | * '''Problem:''' The Brain Computer Interface (BCI) allows you to help people with disabilities regain their mobility. According to the available description of the device signal, it is necessary to simulate the behavior of the subject. | ||

+ | * '''Data:''' Data sets of ECoG/EEG brain signals. | ||

+ | * '''References:''' | ||

+ | #* Motrenko A.P., Strijov V.V. Multi-way feature selection for ECoG-based brain-computer Interface // Expert systems with applications. - 2018. | ||

+ | * '''Basic algorithm''': It is proposed to compare with the partial least squares algorithm. | ||

+ | * '''Solution:''' In this work, it is proposed to build a single system that solves the problem of signal decoding. As stages of building such a system, it is proposed to solve the problems of data preprocessing, feature space extraction, dimensionality reduction and selection of a model of optimal complexity. It is proposed to use the tensor version of PLS with feature selection. | ||

+ | * '''Novelty:''' In the formulation of the problem, the complex nature of the signal is taken into account: a continuous trajectory of movement, the presence of discrete structural variables (fingers or joint movement), the presence of continuous variables (position of a finger or limb). | ||

+ | * '''Authors:''' Strijov V.V., Tetiana Aksenova, consultant – Roman Isachenko | ||

+ | |||

+ | ===41. 2019=== | ||

+ | * '''Title:''' Optimal Approximation of Non-linear Power Flow Problem | ||

+ | * '''Problem:''' Our goal is to approximate the solution of non-linear non-convex optimal power flow problem by solving a sequence of convex optimization problems (aka trust region approach). On this way we propose to compare various approaches for an approximate solution of this problem with adaptive approximation of the power flow non-linearities with a sequence of quadratic and/or piece-wise linear functions | ||

+ | * '''Data:''' Matpower module from MATLAB contains all necessary test cases. Start considering IEEE 57 bus case. | ||

+ | * '''References:''' | ||

+ | *# Molzahn, D. K., & Hiskens, I. A. (2019). A survey of relaxations and approximations of the power flow equations. Foundations and Trends in Electric Energy Systems, 4(1-2), 1-221. https://www.nowpublishers.com/article/DownloadSummary/EES-012 | ||

+ | *# The QC Relaxation: A Theoretical and Computational Study on Optimal Power Flow. Carleton Coffrin ; Hassan L. Hijazi; Pascal Van Hentenryck https://ieeexplore.ieee.org/abstract/document/7271127/ | ||

+ | *# Convex Relaxations in Power System Optimization: A Brief Introduction. Carleton Coffrin and Line Roald. https://arxiv.org/pdf/1807.07227.pdf | ||

+ | *# Optimal Adaptive Linearizations of the AC Power Flow Equations. Sidhant Misra, Daniel K. Molzahn, and Krishnamurthy Dvijotham https://molzahn.github.io/pubs/misra_molzahn_dvijotham-adaptive_linearizations2018.pdf | ||

+ | * '''Base algorithm:''' A set of algorithms described in [1] should be considered to compare with, details behind the proposed method would be shared by the consultant (a draft of the paper) | ||

+ | * '''Solution:''' to figure out the quality of the solution we propose to compare it with the ones given by IPOPT and numerous relaxations, and do some reverse engineering regarding to our method | ||

+ | * '''Novelty:''' The OPF is a truly hot topic in power systems, and is of higher interest by the discrete optimization community (as a general QCQP problem). Any advance in this area is of higher interest by the community | ||

+ | * '''Authors:''' Yury Maximov (consultant and expert), Michael Chertkov (expert) | ||

+ | * '''Notes''': the problem has both the computational and the theoretical focuses, so 2 students are ok to work on this topic | ||

+ | |||

+ | ===2. 2019=== | ||

+ | * '''Title:''' Investigation of reference objects in the problem of metric classification of time series. | ||

+ | * '''Problem description:''' The DTW function is the distance between two time series that can be non-linearly warped relative to each other. It looks for the best alignment between two objects, so it can be used in a metric object classification problem. One of the methods for solving the problem of metric classification is measuring distances to reference objects and using the vector of these distances as an indicative description of the object. The DBA method is an algorithm for constructing centroids (reference objects) for time series based on the DTW distance. When plotting the distance between the time series and the centroid, different pairs of values (eg peak values) are more specific to one of the classes, and the impact of such coincidences on the distance value should be higher. | ||

+ | It is necessary to explore various ways of constructing reference objects, as well as determining their optimal number. The criterion is the quality of the metric classifier in The problem. In the DBA method, for each centroid, it is proposed to create a weight vector that demonstrates the "significance" of the measurements of the centroid, and use it in the modified weighted-DTW distance function. | ||

+ | * '''Data:''' The data describes 6 classes of time series from the mobile phone's accelerometer. https://sourceforge.net/p/mlalgorithms/code/HEAD/tree/Group274/Goncharov2015MetricClassification/data/ | ||

+ | * '''References:''' | ||

+ | *# DTW: http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.215.7850&rep=rep1&type=pdf | ||

+ | *# DBA: https://hal.sorbonne-universite.fr/hal-01630288/document | ||

+ | *# weighted DTW: http://www.mathnet.ru/php/archive.phtml?wshow=paper&jrnid=ia&paperid=414&option_lang=rus | ||

+ | * '''Base algorithm:''' Implement basic methods: | ||

+ | *# Selection of a subset of training sample objects as reference | ||

+ | *# Pre-processing of anomalous objects | ||

+ | *# Clustering training sample objects to build centroids within the cluster | ||

+ | *# Using the DBA method to build reference objects | ||

+ | *# Using numerical optimization methods to find the optimal vector of weights with given constraints | ||

+ | * '''Solution:''' Extension of constraint types to weight vector type: binary vector, same vector for all centroids, binary same vector for all centroids. Such a solution will save energy costs during the operation of sensors of a mobile device. | ||

+ | Literature research and a combination of up-to-date methods. | ||

+ | * '''Novelty:''' There has not been a comprehensive study of various methods of constructing centroids and reference elements along with the choice of their optimal number. | ||

+ | * '''Authors:''' Alexey Goncharov - consultant, Expert, Strijov V.V. - Expert | ||

+ | |||

+ | ===7. 2019=== | ||

+ | * '''Title:''' Privileged learning in the iris boundary approximation problem | ||

+ | * '''Problem:''' Based on the image of the human eye, determine the circles approximating the inner and outer border of the iris. | ||

+ | * '''Data:''' Bitmap monochrome images, typical size 640*480 pixels (however other sizes are possible)[http://www.bath.ac.uk/elec-eng/research/sipg/irisweb/ ], [http://www.cb-sr.ia.ac.cn/IrisDatabase.htm]. | ||

+ | * '''References:''' | ||

+ | *# Aduenko A.A. Selection of multi-models in The problems classification (supervisor Strijov V.V.). Moscow Institute of Physics and Technology, 2017. [http://www.frccsc.ru/sites/default/files/docs/ds/002-073-05/diss/11-aduenko/11-Aduenko_main.pdf?626] | ||

+ | *# K.A. Gankin, A.N. Gneushev, I.A. Matveev Segmentation of the iris image based on approximate methods with subsequent refinements // Izvestiya RAN. Theory and control systems, 2014, no. 2, p. 78–92. | ||

+ | *# Duda, R. O. Use of the Hough transformation to detect lines and curves in pictures / R. O. Duda, P. E. Hart // Communications of the ACM. 1972 Vol. 15, no. 1.Pp. | ||

+ | * '''Basic algorithm''': Efimov Yury. Search for the outer and inner boundaries of the iris in the eye image using the paired gradient method, 2015. | ||

+ | * '''Solution:''' See [[Media:Iris_circle_problem.pdf | iris_circle_problem.pdf]] | ||

+ | * '''Novelty:''' A fast non-enumerative algorithm for approximating boundaries using linear multimodels is proposed. Additionally, capsule neural networks. | ||

+ | * '''consultant''': Radoslav Neichev (by Strijov V.V., Expert Matveev I.A.) | ||

+ | |||

+ | ===44. 2019=== | ||

+ | *'''Name:''' Early prediction of sufficient sample size for a generalized linear model. | ||

+ | *'''Problem''': The problem of designing an experiment is being investigated. The problem of estimating a sufficient sample size according to the data is solved. The sample is assumed to be simple. It is described by an adequate model. Otherwise, the sample is generated by a fixed probabilistic model from a known class of models. The sample size is considered sufficient if the model is restored with sufficient confidence. It is required, knowing the model, to estimate a sufficient sample size at the early stages of data collection. | ||

+ | *'''Data:''' For the computational experiment, it is proposed to use classical samples from the UCI repository. Link to samples https://github.com/ttgadaev/SampleSize/tree/master/datasets | ||

+ | *'''References:''' | ||

+ | *# [Overview of methods for estimating sample size] | ||

+ | *# http://svn.code.sf.net/p/mlalgorithms/code/PhDThesis/. | ||

+ | *# Bootstrap method. https://projecteuclid.org/download/pdf_1/euclid.aos/1. | ||

+ | Bishop, C. 2006. Pattern Recognition and Machine Learning. Berlin: Springer. $758 | ||

+ | *'''Basic algorithm''': We will say that the sample size is sufficient if the log-likelihood has a small variance, on a sample of size m calculated using bootstrap. | ||

+ | We are trying to approximate the dependence of the average value of log-likelihood and its variance on the sample size. | ||

+ | *'''Solution:''' The methods described in the review are asymptotic or require a deliberately large sample size. The new method should be to predict volume in the early stages of experiment design, i.e. when data is scarce. | ||

+ | *'''Authors:''' Grabovoi A.V. (consultant), Gadaev T. T. Strijov V.V. (Experts) | ||

+ | * Note: to determine the simplicity of the sample, a new definition of complexity is proposed ([http://www.machinelearning.ru/wiki/images/3/37/Ivanychev18BachelorThesis_%28merged%29.pdf Sergey Ivanychev]). This is a separate work, +1 The problem 44a (? Katruza). | ||

+ | |||

+ | ===15. 2019=== | ||

+ | * '''Title:''' Formulation and solution of an optimization problem combining classification and regression to estimate the binding energy of a protein and small molecules. The problem description [https://www.overleaf.com/read/rjdnyyxpdkyj] | ||

+ | * '''Problem:''' From a bioinformatics point of view, The problem is to estimate the free energy of protein binding to a small molecule (ligand): the best ligand in its best position has the ''lowest free energy'' of interaction with the protein. (Following a large text, see the file at the link above.) | ||

+ | * '''Data:''' | ||

+ | *# Data for binary classification. Approximately 12,000 protein-ligand complexes: for each of them there is 1 native position and 18 non-native ones. The main descriptors are histograms of distributions of distances between different atoms of the protein and ligand, the dimension of the vector of descriptors is ~ 20,000. In the case of continued research and publication in a specialized journal, the set of descriptors can be expanded. The data will be provided as binary files with a python script to read. | ||

+ | *# Data for regression. For each of the presented complexes, the value of the quantity is known, which can be interpreted as the binding energy. | ||

+ | * '''References:''' | ||

+ | *# SVM [http://cs229.stanford.edu/notes/cs229-notes3.pdf] | ||

+ | *# Ridge Regression [http://scikit-learn.org/stable/modules/linear_model.html#ridge-regression] | ||

+ | *# [https://alex.smola.org/papers/2003/SmoSch03b.pdf] (section 1) | ||

+ | * '''Basic algorithm''': [https://hal.inria.fr/hal-01591154/] In the classification problem, we used an algorithm similar to linear SVM, whose relationship with the energy estimate is beyond the scope of the classification problem, described in the above article. Various loss functions can be used in a regression problem. | ||

+ | * '''Solution:''' It is necessary to connect the previously used optimization problem with the regression problem and solve it using standard methods. Cross-validation will be used to check the operation of the algorithm. There is a separate test set consisting of (1) 195 complexes of proteins and ligands, for which it is necessary to find the best ligand pose (the algorithm for obtaining ligand positions differs from that used in training), (2) complexes of proteins and ligands, for which native poses it is necessary to predict the energy binding, and (3) 65 proteins for which the most strongly binding ligand is to be found. | ||

+ | * '''Novelty:''' First of all, the interest is ''combining classification and regression problems. The correct assessment of the quality of protein and ligand binding is used in drug development to search for molecules that interact most strongly with the protein under study. Using the classification problem described above to predict the binding energy results in an insufficiently high correlation of predictions with experimental values, while using the regression problem alone leads to overfitting. | ||

+ | * '''Authors''' Sergei Grudinin, Maria Kadukova | ||

+ | |||

+ | ===27. 2019=== | ||

+ | * '''Title:''' Creation of ranking models for information retrieval systems. Algorithm for Predicting the Structure of Locally Optimal Models | ||

+ | * '''Problem:''' It is required to predict a time series using some parametric superposition of algebraic functions. It is proposed not to cost the prognostic model, but to predict it, that is, to predict the structure of the approximating superposition. A class of considered superpositions is introduced, and on the set of such structural descriptions, a search is made for a locally optimal model for the problem under consideration. The problem consists in 1) searching for a suitable structural description of the model 2) describing the search algorithm for the structure that will correspond to the optimal model 3) describing the algorithm for inverse construction of the model according to its structural description. For an already existing example of the answer to questions 1-3, see the works of A. A. Varfolomeeva. | ||

+ | * '''Data:''' | ||

+ | *# Collection of text documents TREC (!) | ||

+ | *# A set of time series, which implies the restoration of functional dependencies. It is proposed to first use synthetic data or immediately apply the algorithm to forecasting time series 1) electricity consumption 2) physical activity with subsequent analysis of the resulting structures. | ||

+ | * '''References:''' | ||

+ | *# (!) Kulunchakov A.S., Strijov V.V. Generation of simple structured Information Retrieval functions by genetic algorithm without stagnation // [http://strijov.com/papers/Kulunchakov2014RankingBySimpleFun.pdf Expert Systems with Applications, 2017, 85: 221–230.] | ||

+ | *# A. A. Varfolomeeva Selection of features when marking up bibliographic lists using structural learning methods, 2013, [http://www.machinelearning.ru/wiki/images/f/f2/Varfolomeeva2013Diploma.pdf?format=raw] | ||

+ | *# Bin Cao, Ying Li and Jianwei Yin Measuring Similarity between Graphs Based on the Levenshtein Distance, 2012, [http://naturalspublishing.com/files/published/92cn7jm44d8wt1.pdf?format=raw] | ||

+ | * '''Base algorithm:''' Described in [1]. Developed in the work of the 974 group team. It is proposed to use their code and experiment. | ||

+ | * '''Solution:''' It is proposed to try to repeat the experiment of A. A. Varfolomeeva for a different structural description in order to understand what is happening. The superposition of algebraic functions defines an ortree, on the vertices of which the labels of the corresponding algebraic functions or variables are given. Therefore, the structural description of such a superposition can be its DFS-code. This is a string consisting of vertex labels, written in the order in which the tree is traversed by depth-first search. Knowing the arities of the corresponding algebraic functions, we can restore any such DFS-code in O(n) and get back the superposition of functions. On the set of similar string descriptions, it is proposed to search for the string description that will correspond to the optimal model. | ||

+ | * '''Authors:''' consultant [https://www.inria.fr/centre/grenoble Andrey Kulunchakov (Inria Montbonnot)], Expert Strijov V.V. | ||

+ | |||

+ | ===26. 2019=== | ||

+ | * '''Title:''' Accelerometer positioning | ||

+ | * '''Problem:''' Given initial coordinates, accelerometer signals, additional information (gyroscope, magnetometer signals). Possibly inaccurate map given (The problem [https://en.wikipedia.org/wiki/Simultaneous_localization_and_mapping SLAM]) | ||

+ | * '''Data:''' from [1], self-collected data. | ||

+ | * '''References:''' | ||

+ | *# https://arxiv.org/pdf/1712.09004.pdf | ||

+ | *# https://ieeexplore.ieee.org/document/1528431 | ||

+ | * '''Basic algorithm''': from [1]. | ||

+ | * '''Solution:''' Search for a priori and additional information that improves positioning accuracy. | ||

+ | * '''Novelty:''' Statement of the problem in terms of Projection to Latent Spaces | ||

+ | * '''Authors:''' consultant [http://www.forecsys.ru/ru/site/projects/solut2/ Anastasia Motrenko], Expert [https://www.huawei.com/en/ Ilya Gartseev] , Strijov V.V. | ||

+ | |||

+ | ===45. 2019=== | ||

+ | * Name: The problem of searching characters in images | ||

+ | * '''Problem description:''' This The problem in one of the formulation options can be reduced to two sequential operations: 1) searching for objects in the image and determining their class 2) searching the database for information about the symbolic meaning of the found objects. The main difficulty in solving the problem lies in the search for objects in the image. However, the following classification may also be difficult due to the fact that the image of the object may be incomplete, unusually stylized, and the like. | ||

+ | * Data: Dictionary of Symbols Museum Sites Image-net | ||

+ | * '''References:''' | ||

+ | *# http://www.machinelearning.ru/wiki/images/e/e2/IDP18.pdf (p. 116) | ||

+ | *# http://www.image-net.org | ||

+ | * Basic algorithm: CNN | ||

+ | * '''Solution:''' It is proposed to compare the work of several state-of-the-art algorithms. Suggest a quality metric for searching and classifying objects. Determine applicability of methods. | ||

+ | * '''Novelty:''' The proposed image analysis approach is used by Experts in manual mode and has not been automated | ||

+ | * '''Authors:''' M. Apishev (consultant), D. Lemtyuzhnikova | ||

+ | |||

+ | ===28. 2019=== | ||

+ | * Name: Multi-simulation as a universal way to describe a general sample | ||

+ | * '''Problem description:''' Build a method for incremental refinement of the multimodel structure when new objects appear. Development and comparison of different algorithms for updating the structure of multimodels. Construction of an optimal scheme for refining the structure of a multimodel depending on the total sample size. | ||

+ | * Data: At the initial stage of work, synthetic data with a known statistical structure is used. Testing of the developed methods is carried out on real data from the UCI repository. | ||

+ | * '''References:''' | ||

+ | *# Bishop, Christopher M. "Pattern recognition and machine learning." Springer, New York (2006). | ||

+ | *# Gelman, Andrew, et al. Bayesian data analysis, 3rd edition. Chapman and Hall/CRC, 2013. | ||

+ | *# MacKay, David JC. "The evidence framework applied to classification networks." Neural computation 4.5 (1992): 720-736. | ||

+ | *# Aduenko A. A. "Choice of multimodels in The problem classification" Ph.D. thesis | ||

+ | *# Motrenko, Anastasiya, Strijov V.V., and Gerhard-Wilhelm Weber. "Sample size determination for logistic regression." Journal of Computational and Applied Mathematics 255 (2014): 743-752. | ||

+ | * Basic algorithm: Algorithm for constructing adequate multi-models from #4. | ||

+ | * '''Solution:''' Bayesian approach to the problem of choosing models based on validity. Analysis of the properties of validity and its relationship with statistical significance. | ||

+ | * '''Novelty:''' A method is proposed for constructing an optimal scheme for updating the structure of a multimodel when new objects appear. The relationship between validity and statistical significance for some classes of models has been studied. | ||

+ | * '''Authors:''' Strijov Vadim Viktorovich, Aduenko Alexander Alexandrovich (GMT-5) | ||

+ | |||

+ | ===11. 2019=== | ||

+ | * '''Title:''' Automatic construction of a neural network of optimal complexity | ||

+ | * '''Problem:''' The problem of finding a stable (and not redundant in terms of parameters) neural network structure is considered. The neural network is considered as a computational graph, the edges of which are primitive functions, and the vertices are intermediate representations of the sample obtained under the action of these functions. It is required to choose a subgraph of the model, in which the final neural network will give an acceptable classification quality with a small number of parameters. | ||

+ | * '''Data:''' Samples Boston, MNIST, CIFAR-10 | ||

+ | * '''References:''' | ||

+ | *# [http://strijov.com/papers/BakhteevEvidenceArticle3.pdf Oleg Bakhteev Yu., Strijov V.V. Selection of deep learning models of suboptimal complexity using variational likelihood estimation // Avtomatika and telemechanika, 2018.] | ||

+ | *# [http://strijov.com/papers/SmerdovBakhteevStrijov_Paraphrase2017.pdf Smerdov A.N., Oleg Bakhteev Yu., Strijov V.V. Choosing the optimal model of the recurrent network in the Paraphrase Search The problems // Informatics and its applications, 2018.] | ||

+ | *# [https://papers.nips.cc/paper/4329-practical-variational-inference-for-neural-networks] Variational inference. | ||

+ | *# [https://arxiv.org/abs/1611.00712] Relaxation based on variational inference. | ||

+ | *# [https://arxiv.org/abs/1806.09055] DARTS. | ||

+ | * '''Base algorithm:''' random search and DARTS algorithm (model selection using relaxation without variational inference). | ||

+ | * '''Decision'''It is proposed to choose the structure of the neural network based on the variational inference. To select the optimal structure, relaxation is used: from a strict choice of one of several considered submodels of the neural network, it is proposed to move to the composition of these models with different weights for each of them. | ||

+ | * '''Novelty:''' A method of automatic model building is proposed, which takes into account inaccuracies in the optimization of model parameters and allows finding the most stable models. | ||

+ | * '''Authors:''' Oleg Bakhteev, Strijov V.V. | ||

+ | |||

+ | ===48. 2019=== | ||

+ | * '''Title:''' Multi-simulation, privileged training | ||

+ | * '''Problem:''' Considers The problem of learning one model from another | ||

+ | * '''Data:''' Time series samples | ||

+ | * '''References:''' | ||

+ | *# https://github.com/neychev/distillation_n_privileged_info_torch | ||

+ | *# https://github.com/neychev/MultiThe problem_forecast_code | ||

+ | *# Article by Mixture Experts | ||

+ | *# Neychev's diploma http://www.machinelearning.ru/wiki/images/3/36/NeyhevMS_Thesis.pdf | ||

+ | * '''Base algorithm:''' Blend of Experts, privileged training, distillation | ||

+ | * '''Solution''' Run an experiment illustrating these approaches | ||

+ | * '''Novelty:''' A forecasting method is proposed that uses a priori information about the membership of the model sample (publish the results). | ||

+ | * '''Authors:''' R.G. Neichev (consultant), Strijov V.V. | ||

+ | |||

+ | ===49. 2019=== | ||

+ | * Name: Brain signal decoding and intention prediction | ||

+ | * '''Problem description:''' It is required to build a model that restores the movement of the limbs according to the corticogram. | ||

+ | * Data: neurotycho.org [9] (or fingers) | ||

+ | * '''References:''' | ||

+ | *# Neichev R.G., Katrutsa A.M., Strijov V.V. Selection of the optimal set of features from a multicorrelated set in the forecasting problem. Zavodskaya Lab. Materials Diagnostics, 2016, 82(3) : 68-74. [10] | ||

+ | *# Isachenko R.V., Strijov V.V. Quadratic Programming Optimization with Feature Selection for Non-linear Models // Lobachevskii Journal of Mathematics, 2018, 39(9) : 1179-1187. article | ||

+ | * Basic algorithm: Partial Least Squares[11] | ||

+ | * '''Solution:''' Create a feature selection algorithm alternative to PLS and taking into account the non-orthogonal feature interdependence structure. | ||

+ | * '''Novelty:''' A feature selection method is proposed that takes into account the regularities of both the and independent variable and the dependent variable. Bonus: Explore changes in model structure as the nature of the sample changes. | ||

+ | * '''Authors:''' Andrey Zadayanchuk, Strijov V.V. | ||

+ | |||

+ | ==2018== | ||

+ | |||

+ | {|class="wikitable" | ||

+ | |- | ||

+ | ! Title | ||

+ | ! Links | ||

+ | ! Team | ||

+ | |- | ||

+ | |(Example) Metric classification of time series | ||

+ | |[http://svn.code.sf.net/p/mlalgorithms/code/Group274/Goncharov2015MetricClassification/code Code], | ||

+ | [https://docs.google.com/document/d/1fx7fVlmnwdTesElt-lbaHvoGEjJC5t_9e-X0ZpUzEcQ/edit?usp=sharing LinkReview], | ||

+ | [https://t.me/joinchat/Ak0SzkfYN_boA3eRtfPKvg Discussion] | ||

+ | |[http://svn.code.sf.net/p/mlalgorithms/code/Group274/Goncharov2015MetricClassification/doc/Goncharov2015MetricClassification.pdf Alexey Goncharov]*, [http://svn.code.sf.net/p/mlalgorithms/code/Group274/Goncharov2015MetricClassification/doc/Goncharov2015MetricClassification.pdf Maxim Savinov] | ||

+ | |- | ||

+ | |Forecasting the direction of movement of the price of exchange instruments according to the news flow | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-1 Code], | ||

+ | [https://docs.google.com/document/d/1qa6PO_3AXcXPkJKNjQgihBXWkmBpspFWi3Ct34FYonw/edit LinkReview], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-1/raw/master/Presentation.pdf Slides], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-1/raw/master/Paper.pdf Report] | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-1/raw/master/Borisov2018Project1/Borisov2018Project1.pdf Alexander Borisov], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-1/blob/master/Drobin2018Project1/Drobin2018Project1.pdf Drobin Maxim], [https://github.com/Intelligent-Systems-Phystech/2018-Project-1/raw/master/Govorov2018Project1/Govorov2018Project1.pdf Govorov Ivan], [https://github.com/Intelligent-Systems-Phystech/2018-Project-1/raw/master/Mukhitdinova2018Project1/Mukhitdinova2018Project1.pdf Mukhitdinova Sofia], [https://github.com/Intelligent-Systems-Phystech/2018-Project-1/raw/master/Rodionov2018Project1/Rodionov2018Project1.pdf Valentin Rodionov], [https://github.com/Intelligent-Systems-Phystech/2018-Project-1/raw/master/Akhiarov2018Project1/Akhiarov2018Project1.pdf Valentin Akhiyarov] | ||

+ | |- | ||

+ | |Construction of reference objects for a set of multidimensional time series | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-2 Code] | ||

+ | [https://docs.google.com/document/d/1ruVHmEMgBXcULWsy-mYg2KgAV2SyC5si4T4UHVPMu2E/edit LinkReview] | ||

+ | |[https://raw.githubusercontent.com/Intelligent-Systems-Phystech/2018-Project-2/master/Iskhakov2018Project2/test.pdf Iskhakov Rishat], | ||

+ | [https://raw.githubusercontent.com/Intelligent-Systems-Phystech/2018-Project-2/master/Korepanov2018Project2/test.pdf Korepanov Georgy], | ||

+ | [https://raw.githubusercontent.com/Intelligent-Systems-Phystech/2018-Project-2/master/Solodnev2018Project2/test.pdfStepan Solodnev] | ||

+ | [https://raw.githubusercontent.com/Intelligent-Systems-Phystech/2018-Project-2/master/Solodnev2018Project2/test.pdf Samirkhanov Danil] | ||

+ | |- | ||

+ | |Dynamic alignment of multivariate time series | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-3 Code] | ||

+ | [https://docs.google.com/document/d/1ruVHmEMgBXcULWsy-mYg2KgAV2SyC5si4T4UHVPMu2E/edit LinkReview] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-3/raw/master/Morgachev2018Title/presentation/MorgachevSmirnovLipnitckaia2019SpatialTsSlides.pdf Slides] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-3/raw/master/Morgachev2018Title/paper/Morgachev2018Title.pdf Report] | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-3/raw/master/Morgachev2018Title/Morgachev2018Title.pdf Gleb Morgachev], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-3/blob/master/Smirnov2018Title/Smirnov2018Title.pdf Vladislav Smirnov], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-3/blob/master/Lipnitckaia2018Title/Lipnitckaia2018Title.pdf Tatiana Lipnitskaya] | ||

+ | |- | ||

+ | |Automatic adjustment of ARTM parameters for a wide class of The problems | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-4 Code], | ||

+ | [https://docs.google.com/document/d/1RidglPMH1-Yb1rx7V7QayDDuM-HfL-pF2kkGBWbWrxk/edit LinkReview], | ||

+ | [https://docs.google.com/presentation/d/1WpCbs7Rf9i7oCT25mSTcbBCLlN_tXwdjdv1VQ6Y8bVs/edit#slide=id.p Presentation] | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-4/raw/master/Golubeva2018Problem4/Golubeva2018Problem4.pdf Golubeva Tatiana], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-4/raw/master/Ivanova2018Problem4/Ivanova2018Problem4.pdf Ivanova Ekaterina], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-4/raw/master/Matveeva2018Problem4/Matveeva2018Problem4.pdf Matveeva Svetlana], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-4/raw/master/Trusov2018Problem4/Trusov2018Problem4.pdf Trusov Anton], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-4/raw/master/Tsaritsyn2018Problem4/Tsaritsyn2018Problem4.pdf Tsaritsyn Mikhail], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-4/raw/master/Chernonog2018Problem4/Chernonog2018Problem4.pdf Chernonog Vyacheslav] | ||

+ | |- | ||

+ | |Finding paraphrases | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-5 Code], | ||

+ | [https://docs.google.com/document/d/1rTEFOVCDVNPHss09IRG-C95yovUE4XTyryOnpb8DWFA LinkReview] | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-5/raw/master/Okrug2018Paraphrases/report.pdf Stas Okrug], [https://github.com/Intelligent-Systems-Phystech/2018-Project-5/raw/master/Mokrov2018Paraphrases/report.pdf Nikita Mokrov] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-5/raw/master/Kitashov2018Paraphrases/report.pdf Fedor Kitashov], [https://github.com/Intelligent-Systems-Phystech/2018-Project-5/raw/master/Proskura2018Paraphrases/report.pdf Polina Proskura], [https://github.com/Intelligent-Systems-Phystech/2018-Project-5/raw/master/Basimova2018Paraphrases/report.pdf Natalia Basimova], [https://github.com/Intelligent-Systems-Phystech/2018-Project-5/raw/master/Krasnikov2018Paraphrases/report.pdf Roman Krasnikov], [https://github.com/Intelligent-Systems-Phystech/2018-Project-5/raw/master/Shabanov2018Paraphrases/report.pdf Akhmedkhan Shabanov] | ||

+ | |- | ||

+ | |On conformational changes of proteins using collective motions in torsion angle space and L1 regularization | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-6 Code], | ||

+ | [https://docs.google.com/document/d/1XKrjFj5TSIV6GxW1kZj6DthHC8jc9YbGN44Tr67TdQI LinkReview] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-6/raw/master/report/Proteins.pdf Presentation] | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-6/raw/master/Ryabinina2018Project6/report.pdf Ryabinina Raisa], [https://github.com/Intelligent-Systems-Phystech/2018-Project-6/raw/master/Emtsev2018Project6/report.pdf Emtsev Daniil] | ||

+ | |- | ||

+ | |Privileged training in the problem of approximating the borders of the iris | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-7 Code], | ||

+ | [https://docs.google.com/document/d/1QNm2L98o-yz_LuwqBaC-XqpX49Rhvpzoli9WcTUQrH8/edit LinkReview] | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-7/raw/master/Learning_Pashtet_Crew/Fedosov2018Project7/Privileged_training_in_the_problem_of_approximating_the_borders_of_the_iris.pdf Pavel Fedosov], [https://github.com/Intelligent-Systems-Phystech/2018-Project-7/raw/master/Learning_Pashtet_Crew/Gladkov2018Project7/Privileged_training_in_the_problem_of_approximating_the_borders_of_the_iris.pdf Alexey Gladkov], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-7/raw/master/Learning_Pashtet_Crew/Kenigsberger2018Project7/Privileged_training_in_the_problem_of_approximating_the_borders_of_the_iris.pdf Genrikh Kenigsberger], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-7/raw/master/Learning_Pashtet_Crew/Korostelev2018Project7/Privileged_training_in_the_problem_of_approximating_the_borders_of_the_iris.pdf Ivan Korostelev], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-7/raw/master/Learning_Pashtet_Crew/Balakin2018Project7/Privileged_training_in_the_problem_of_approximating_the_borders_of_the_iris.pdf Nikolay Balakin] | ||

+ | |- | ||

+ | |Generation of features using locally approximating models | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-8/tree/master/code Code], | ||

+ | [https://docs.google.com/document/d/1e65opLey0Yxo_kAZ4cKTcjMIIYxR1jVPCQrpmr4k29w/edit?usp=sharing LinkReview] | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-8/raw/master/Kurashov2018Project8/Kurashov2018Project8.pdf Ibrahim Kurashov], [https://github.com/Intelligent-Systems-Phystech/2018-Project-8/raw/master/Gilmutdinov2018Project8/Gilmutdinov2018Project8.pdf Nail Gilmutdinov], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-8/raw/master/Mulyukov2018Project8/Mulyukov2018Project8.pdf Albert Mulyukov], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-8/raw/master/Spivak2018Project8/Spivak2018Project8.pdf Valentin Spivak] | ||

+ | |- | ||

+ | |Text recognition based on skeletal representation of thick lines and convolutional networks | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-9 Code], [https://docs.google.com/document/d/1vvOqLwLJSelbKBglc4LKh6XUWS5c72L0XMzyeJ20XBM/edit LiteratureReview], [https://drive.google.com/file/d/1pzfKkjVe1aP1-5ab1ewN0NMF60RJ26IA/view?usp=drivesdk Slides], [https://github.com/Intelligent-Systems-Phystech/2018-Project-9/raw/master/Lukoyanov2018Project9/main.pdf report] | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-9/raw/master/Kutsevol2018Project9/Kutsevol_Article.pdf Kutsevol Polina] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-9/raw/master/Lukoyanov2018Project9/main.pdf Lukoyanov Artem] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-9/raw/master/Korobov2018Project9/ArticleKorobov.pdf Korobov Nikita] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-9/raw/master/Boyko2018Project9/Boyko2018Project9.pdf Boyko Alexander] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-9/raw/master/Litovchenko2018Project9/Litovchenko.pdf Litovchenko Leonid] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-9/blob/master/ValukovAlex2018Project9/AlexandrValukov.pdf Valukov Alexandr] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-9/blob/master/Badrutdinov2018Project9/KamilBadrutdinov.pdf Badrutdinov Kamil] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-9/blob/master/Yakushevskiy2018Project9/main.pdf Yakushevskiy Nikita] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-9/blob/master/ValukovKolya2018Project9/main.pdf Valyukov Nikolay] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-9/blob/master/Tushin2018Project9/Tushin.pdf Tushin Kirill] | ||

+ | |- | ||

+ | |Comparison of neural network and continuous-morphological methods in the problem of text detection | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-10 Code], [https://docs.google.com/document/d/1Gocn0x-FfYkD_L7ZLZdULxNTBfo25OMMKPBr2-otw-w/edit?usp=sharing LinkReview], [https://t.me/joinchat/DEQDKU-oqyt8FRG4SoFh3w Discussion], [https://docs.google.com/presentation/d/17_7i0KFELxyaL-MtvVmu2ed07sg331hiMagYqNpq9Ek/edit?usp=sharing Presentation] | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-10/blob/master/report/Gaiduchenko2018Project10/Gaiduchenko2018Project10.pdf Gaiduchenko Nikolay] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-10/tree/master/report/Torlak2018Project10 Torlak Artyom] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-10/tree/master/report/Akimov2018Project10 Akimov Kirill] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-10/tree/master/report/Mironova2018Project10 Mironova Lilia] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-10/tree/master/report/Gonchar2018Project10 Gonchar Daniel] | ||

+ | |- | ||

+ | |Automatic construction of a neural network of optimal complexity | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-11 Code], [https://docs.google.com/document/d/131-9Uxl4tTIMKBh7WNJuZR5MI1pHypvcb5qsYl-bAnI/edit?usp=sharing LinkReview], [https://github.com/Intelligent-Systems-Phystech/2018-Project-11/raw/master/report/report.pdf report], [https://github.com/Intelligent-Systems-Phystech/2018-Project-11/raw/master/report/pres.pdf slides] | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-11/blob/master/Goryan2018Project11/Goryan2018Project11.pdf Nikolai Goryan] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-11/tree/master/Ulitin2018Project11/Ulitin2018Project11.pdf Alexander Ulitin] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-11/blob/master/Tovkes2018Project11/Abstract.pdf Tovkes Artem] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-11/raw/master/Taranov2018Project-11/Taranov2018Project11.pdf Taranov Sergey] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-11/blob/master/Gubanov2018Project11/Gubanov2018Project11.pdf Gubanov Sergey] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-11/blob/master/Krinitskiy2018Project11/Abstract.pdf Krinitsky Konstantin] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-11/blob/master/Zabaznov2018Project11/Zabaznov2018Project11.pdf Zabaznov Anton] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-11/blob/master/Markin2018Project11/Markin2018Project11%20(1).pdf Valery Markin] | ||

+ | |- | ||

+ | |Machine translation training without parallel texts. | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-12 Code], | ||

+ | [https://docs.google.com/document/d/1_5lrNNecgpiW3yObDglUAkTepVGj8ucreMhhcDV60qc/edit LinkReview], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-12/raw/master/report/result.pdf Report], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-12/raw/master/report/pres.pdf Slides] | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-12/raw/master/Artemenkov2018Title/Artemenkov2018Title.pdf Alexander Artemenkov] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-12/raw/master/Yaroshenko2018Title/Yaroshenko2018Title.pdf Angelina Yaroshenko] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-12/blob/master/Stroganov2018Title/Stroganov2018Title.pdf Andrey Stroganov] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-12/blob/master/Skidnov2018Title/Skidnov2018Title.pdf Egor Skidnov] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-12/raw/master/Borisova2018Title/Borisova2018Title.pdf Anastasia Borisova] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-12/blob/master/Ryabov2018Title/Ryabov2018Title.pdf Ryabov Fedor] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-12/tree/master/Mazurov2018Title/Abstract.pdf Mazurov Mikhail] | ||

+ | |- | ||

+ | |Deep learning for RNA secondary structure prediction | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-13/tree/master/code Code] | ||

+ | [https://docs.google.com/document/d/1RrIPcrVb0mEdA_hc7Ttk8thIDnDvtBXgyriIxwpYzzM/edit Link Review] | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-13/blob/master/Dorokhin2018Problem13/Dorokhin2018Problem13.pdf Dorokhin Semyon] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-13/tree/master/Pastukhov2018Project13 Pastukhov Sergey] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-13/raw/master/Pikunov2018Problem13/first.pdf Pikunov Andrey] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-13/blob/master/Nesterova2018Project13/tutorial.pdf Nesterova Irina] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-13/blob/master/Kurilovich2018Problem13/Kurilovich2018Problem13.pdfKurilovich Anna] | ||

+ | [https://t.me/joinchat/DE_WxRAo9v0lIKxGyc07Kg chat] | ||

+ | |- | ||

+ | |Deep Learning for reliable detection of tandem repeats in 3D protein structures | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2019-Project-14 Code] | ||

+ | [https://docs.google.com/document/d/1_BtCiAihPg9ON-2PlxORkcmwL80pgqC4gOE7A03rQjg Link Review] | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2019-Project-14/raw/master/Veselova2019Project14/Veselova2019Project14.pdf Veselova Evgeniya] | ||

+ | |- | ||

+ | |Formulation and solution of an optimization problem combining classification and regression to estimate the binding energy of a protein and small molecules | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-15/Code Code] | ||

+ | [https://docs.google.com/document/d/1Be2O0My8KWwOKLo8bFMmF8tPMCFGCK4zUVArurrPeNQ/edit Link Review] | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-15/tree/master/Merkulova2018Title Merkulova Anastasia] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-15/tree/master/Plumite2018Title Plumite Elvira] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-15/tree/master/Zhiboedova2018Title Zhiboyedova Anastasia] | ||

+ | [https://vk.me/join/AJQ1d2J3jQq0jJ50G5VAoioS chat] | ||

+ | |- | ||

+ | |Estimation of the optimal sample size for research in medicine | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-16 Code] | ||

+ | [https://docs.google.com/document/d/1yqnjgMUheHQUp8AAQPqqy9jTJhhzzd_6wvnHY7GF1Fk/edit?usp=sharing Link Review] | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-16/blob/master/report/Kharatyan2018Project16/report.pdf Artemy Kharatyan], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-16/raw/master/Mikheev2018Project16 Mikhail Mikheev], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-16/tree/master/Evgin2018Project16 Evgin Alexander], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-16/tree/master/Seppar2018Project16 Seppar Alexander], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-16/tree/master/Konoplev2018Project16 Konoplyov Maxim], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-16/tree/master/Murlatov2018Project16 Murlatov Stanislav], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-16/tree/master/Makarenko2018Project16 Makarenko Stepan] | ||

+ | |- | ||

+ | |Intention forecasting. Investigation of the properties of local models in the spatial decoding of brain signals | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-17/tree/master/code Code], | ||

+ | [https://docs.google.com/document/d/1j6laGt-zTP3lTm1v0Ozev3dKxivYciq9TOWfmn5sAIU/edit?usp=sharing LinkReview], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-17/raw/master/report/Presentation.pdf Presentation] | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-17/blob/master/Bolobolova2018Project17/Bolobolova2018Project17.pdf Natalia Bolobolova], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-17/raw/master/Samokhina2018Project17/Samokhina2018Problem17.pdf Alina Samokhina], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-17/raw/master/Shiyanov2018Project17/Shiyanov2018Project17.pdf Shiyanov Vadim] | ||

+ | |- | ||

+ | |Intention forecasting. Building an optimal signal decoding model for modeling a brain-computer interface. | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-18 Code], | ||

+ | [https://docs.google.com/document/d/1b-CjunKY5nkZUK0Zfur0nKyQPaY2eWqht7kMcMQd-J8/edit LinkReview], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-18/raw/master/Presentation-v1.pdf Presentation], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-18/raw/master/_________________________.pdf Article] | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-18/raw/master/Nasedkin2018Project18/Nasedkin2018Project18.pdf Ivan Nasedkin], [https://github.com/Intelligent-Systems-Phystech/2018-Project-18/raw/master/Latypova2018Project18/Latypova.pdf Galiya Latypova], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-18/raw/master/Sukhodolskiy2018Project18/Sukhodolskiy2018Project18.pdf Nestor Sukhodolsky], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-18/raw/master/Shemenev2018Project18/Shemenev2018Project18.pdf Alexander Shemenev] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-18/raw/master/Borodulin2018Project18/Borodulin2018Project18.pdf Ivan Borodulin], | ||

+ | |- | ||

+ | |Investigation of the dependence of the quality of recognition of ontological objects on the depth of hyponymy. | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-19 Code], | ||

+ | [https://github.com/ddvika/2018-Project-19/raw/master/report/final_report.pdf Report], | ||

+ | [https://docs.google.com/document/d/1OeMPgVMi72AbHOKsKsUDs6ggMdNL2UT0liycgmYrnLk/edit LinkReview], [https://github.com/ddvika/2018-Project-19/raw/master/report/presentation19project.pdf Presentation] | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-19/raw/master/Rezyapkin2018Project19/RezyapkinPaper.pdf Vyacheslav Rezyapkin], [https://github.com/Intelligent-Systems-Phystech/2018-Project-19/raw/master/Russkin2018Project19/Russkin2018Project19.pdf Alexey Russkin], | ||

+ | [https://github.com/ddvika/2018-Project-19/raw/master/Dochkina2018Project19/Dochkina2018Project19.pdf Victoria Dochkina], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-19/raw/master/Kuznetsov2018Project19/KuznetsovMiron.pdf Miron Kuznetsov], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-19/raw/master/Yarmoshik2018Project19/Yarmoshik_article.pdf Yarmoshyk Demyan] | ||

+ | |- | ||

+ | |Comparison of the quality of end-to-end trainable models in The problem of answering questions in a dialogue, taking into account the context | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-20 Code] | ||

+ | [https://docs.google.com/document/d/1GQmJ6I2fIBchikR-44DcmMD4H-58j3_wuIchNK49Zrs/edit LinkReview] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-20/blob/master/Ryakin2018problem20/Ryakin2018project20.pdf Report], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-20/blob/master/presentation/QuAC.pdf Presentation] | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-20/raw/master/Agafonov2018probem20/article/Agafonov2018project20.pdf Agafonov Alexey], [https://github.com/Intelligent-Systems-Phystech/2018-Project-20/blob/master/Ryakin2018problem20/Ryakin2018project20.pdf Ryakin Ilya],[https://github.com/Intelligent-Systems-Phystech/2018-Project-20/blob/master/Litvinenko2018problem20/Litvinenko2018project20.pdf Litvinenko Vladimir], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-20/blob/master/Khokhlov2018problem20/Khokhlov2018project20.pdf Khokhlov Ivan], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-20/blob/master/Velikovsky2018project20/Velikovsky2018project20.pdf Velikovsky Nikita], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-20/blob/master/Anufrienko2018project20/Anufrienko2018project20.pdf Anufrienko Oleg] | ||

+ | |- | ||

+ | |High order convex optimization methods | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-21/tree/master/code Code], | ||

+ | [https://docs.google.com/document/d/1jF1Hkqbn2e7BnuguTzYuRPp43Y5MbMP36MlWwFVkf6U/edit LinkReview], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-21/blob/master/report/presentation_results.pdf Slides] | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-21/raw/master/Selikhanovych2018Title/Selikhanovych2018Title.pdf Selikhanovich Daniel], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-21/blob/master/Sokolov2018Title/Sokolov2018Title.pdf Sokolov Igor] | ||

+ | |- | ||

+ | |Fractal analysis and synthesis of optical images of sea waves | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-23/tree/master/code code], | ||

+ | [https://docs.google.com/document/d/1g-8H-i8vyThkWUTvthebbr4-qSd8c-kE4B_bieykF7c/edit LinkReview], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-23/blob/master/Kanygin2018/Projecte23_presentation.pdf Presentation] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-23/raw/master/Kanygin2018/Kanygin2018Project23.pdf Report] | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-23/raw/master/Kanygin2018/Kanygin2018Project23.pdf Kanygin Yuri] | ||

+ | |- | ||

+ | |Entropy maximization for various types of image transformations | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-24/tree/master/code code], | ||

+ | [https://docs.google.com/document/d/1FtOjEcx7S0PJ7ASP0V_5zM2nQDTSl0c9I61r0SYAWVc/edit LinkReview], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-24/raw/master/report/report2018Project24.pdf report], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-24/raw/master/slides/slides2018Project24.pdf slides] | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-24/raw/master/Voskresenskiy2018Project24/Voskresenskiy2018Project24.pdf Nikita Voskresensky], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-24/raw/master/Shabalina2018Project24/Shabalina2018Project24.pdf Alisa Shabalina], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-24/raw/master/Murzaev2018Project24/Murzaev2018Project24.pdf Yaroslav Murzaev], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-24/raw/master/Khokhlov2018Project24/Khokhlov2018Project24.pdf Alexey Khokhlov], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-24/raw/master/Kazakov2018Project24/Kazakov2018Project24.pdf Alexey Kazakov], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-24/raw/master/Gribova2018Project24/Gribova2018Project24.pdf Olga Gribova], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-24/raw/master/Belozertsev2018Project24/Belozertsev2018Project24.pdf Alexander Belozertsev] | ||

+ | |- | ||

+ | |Automatic detection and recognition of objects in images | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-25 code], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-25a code_A], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-25/raw/master/Baymakov2018/25_Project_demo.pdf Slides_for_demo], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-25/raw/master/Proshutinskii2018/!%20Article/Proshutinskii2018Project25_30.pdf Report2018Project25_30] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-25/raw/master/report/Report2018Project25_31.pdf Report2018Project25_31] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-25/raw/master/Proshutinskii2018/Project30%20Proshutinskii%20Razumov.pdf slides_30] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-25/raw/master/report/slides_last.pdf slides_25_31] | ||

+ | [https://docs.google.com/document/d/1s7QlihPkamecuVXXLVc5V76cBQn3HBo47HdbAOD0xBI/edit LinkReview] | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-25/raw/master/Demidova2018Title/Demidova2018Project25_31.pdf Julia Demidova] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-25/blob/master/Razumov2018Title/Razumov2018Project25_30.pdf Ivan Razumov] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-25/raw/master/report/Report2018Project25_31.pdf Vladislav Tominin] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-25/raw/master/TomininY2018Title/final/TomininY2018Project25_31.pdf Yaroslav Tominin] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-25/blob/master/Dudorov2018Title/Dudorov2018Project25_31.pdf Nikita Dudorov] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-25/raw/master/Erlygin2018Title/jmlda-example-students.pdf Leonid Erlygin] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-25/raw/master/Proshutinskii2018/!%20Article/Proshutinskii2018Project25_30.pdf Proshutinsky Dmitry] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-25/blob/master/Baymakov2018/25_Project.pdf Baimakov Vladimir] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-25/blob/master/Zubkov2018/Zubkov2018Problem25.pdf Zubkov Alexander] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-25/blob/master/Chernenkova2018/Chernenkova2018Problem25.pdf Chernenkova Elena] | ||

+ | |- | ||

+ | |Location determination by accelerometer signals | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-26 Code], | ||

+ | [https://docs.google.com/document/d/1er3SgPu9bBBWkLk1yVev-9Ue42BOPapOkLn6sL0GAGA/edit?usp=sharing LinkReview], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-26/raw/master/Project26.pdf Slides], | ||

+ | [https://github.com/Vitaly-Protasov/Project26/raw/master/text.pdf Text] | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-26/raw/master/Zainulina2018Project26/Zainulina2018Project26.pdf Elvira Zainulina] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-26/raw/master/Fateev2018Project26/Fateev2018Project26.pdf Fateev Dmitry] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-26/raw/master/ProtasovKing2018Project26/Article.pdf Vitaly Protasov] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-26/raw/master/Bozhedomov2018Project26/Bozhedomov2018Project26.pdf Nikita Bozhedomov] | ||

+ | |- | ||

+ | |Multimodelling as a universal way to describe a general sample | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-28 Code], | ||

+ | [https://docs.google.com/document/d/1w8KoJqcppcsjjtQ_MNd4JTdxmCgerllRRkqvJHWhpX4/edit Linkreview], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-28/blob/master/Slides.pdf Slides], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-28/blob/master/report/report.pdf Report] | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-28/raw/master/Kachanov2018Project28/Kachanov2018Project28.pdf Vladimir Kachanov] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-28/raw/master/Strelkova2018Project28/Strelkova2018Project28.pdf Evgenia Strelkova] | ||

+ | |- | ||

+ | |Cross-Language Document Extractive Summarization with Neural Sequence Model | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-29/tree/master/code Code], | ||

+ | [https://docs.google.com/spreadsheets/d/1mDOp2KnXI9dH8_QYdj4fY-pMBWnqXfECkFUEg244O38/edit#gid=0 Linkreview], [https://github.com/Intelligent-Systems-Phystech/2018-Project-29/raw/master/report/The problem29_Report.pdf Report], [https://github.com/Intelligent-Systems-Phystech/2018-Project-29/raw/master/report/CrossLang_Summa.pdf Slides] | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-29/raw/master/Zakharov2018Title/Zakharov2018Article.pdf Pavel Zakharov] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-29/blob/master/Kvasha2018Title/article.pdf Pavel Kvasha] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-29/tree/master/Dyachkov2018Title/article.pdf Evgeny Dyachkov] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-29/raw/master/Petrov2018Title/article.pdf Evgeny Petrov] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-29/blob/master/Selnitskiy2018Title/article.pdf Ilya Selnitsky] | ||

+ | |- | ||

+ | |Pairwise energy matrix construction for inverse folding problem | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-31/tree/master/code Code], | ||

+ | [https://docs.google.com/document/d/1mq1fNJmwnxeuTJLVmfF9unYP85sdK7FDXGslGFiYZMc/edit LinkReview] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-31/blob/master/Rubinstein2018Project31/Rubinstein2018Project31.pdf Report] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-31/raw/master/report/RubinsteinAR.pdf Slides] | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-31/raw/master/Rubinstein2018Project31/Rubinstein2018Project31.pdf Rubinshtein Alexander] | ||

+ | |- | ||

+ | |Smooth orientation-dependent scoring function | ||

+ | |[https://gitlab.inria.fr/grudinin/sbrod Code] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-SBROD | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/2018-Project-SBROD/blob/master/Noskova/report.pdf Noskova Elizaveta] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-SBROD/blob/master/Kachkov/report.pdf Kachkov Sergey] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/2018-Project-SBROD/blob/master/Sidorenko/report.pdf Sidorenko Anton] | ||

+ | |- | ||

+ | |} | ||

+ | |||

+ | ===5. 2018=== | ||

+ | * '''Title:''' Finding paraphrases. | ||

+ | * '''Problem description:''' Paraphrases are different variations of the same and the same text, identical in meaning, but differing lexically and grammatically, for example: "Where did the car go" and "Which direction did the car go". The problem of detecting paraphrases is to select clusters in a set of texts, such that each cluster contains only paraphrases of the same and the same sentence. The easiest way to extract paraphrases is to cluster texts, where each text is represented by a "bag of words". | ||

+ | * '''Data:''' There are open datasets of questions for testing and training on kaggle.com, there are open datasets for testing from semeval conferences. | ||

+ | * '''Base algorithm:''' Use one of the document clustering algorithms to extract paraphrases, where each document is represented by a bag of words or tf-idf. | ||

+ | * '''Solution:''' Use neural network architectures to search for paraphrases, use phrases extracted with parsers as features, use multilevel clustering. | ||

+ | * '''Novelty:''' Lack of implementations for the Russian language that will use parsers for a similar The problem, all current solutions are quite "simple". | ||

+ | * '''Authors:''' Artyom Popov. | ||

+ | |||

+ | ===6. 2018=== | ||

+ | * '''Title:''' On conformational changes of proteins using collective motions in torsion angle space and L1 regularization. | ||

+ | * '''Problem description:''' Torsion angles are the most natural degrees of freedom for describing motions of polymers, such as proteins. This is because bond lengths and bond angles are heavily constrained by covalent forces. Thus, multiple attempts have been done to describe protein dynamics in the torsion angle space. For example, one of us has developed an elastic network model (ENM) [1] in torsion angle space called Torsional Network Model (TNM) [2]. Functional conformational changes in proteins can be described in the Cartesian space using just a subset of collective coordinates [3], or even a sparse representation of these [4]. The latter requires a solution of a LASSO optimization problem [5]. The goal of the current project is to study if a sparse subset of collective coordinates in the torsion subspace can describe functional conformational changes in proteins. This will require a solution of a ridge regression problem with a L1 regularization constraint. The starting point will be the LASSO formulation. | ||

+ | * '''Data:''' Experimental conformations will be extracted from the Protein Docking Benchmark v5 (https://zlab.umassmed.edu/benchmark/) and a few others. The TNM model can be downloaded from https://ub.cbm.uam.es/tnm/tnm_soft_main.php | ||

+ | * '''References:''' | ||

+ | *# Tirion MM. (1996) Large Amplitude Elastic Motions in Proteins from a Single-Parameter, Atomic Anal- ysis. Phys Rev Lett. 77:1905–1908. | ||

+ | *# Mendez R, Bastolla U. (2011) Torsional network model: normal modes in torsion angle space better correlate with conformation changes in proteins. Phys Rev Lett. 2010 104:228103. | ||

+ | *# SwarmDock and the use of normal modes in protein-protein docking. IH Moal, PA Bates - International journal of molecular sciences, 2010 | ||

+ | *# Modeling protein conformational transition pathways using collective motions and the LASSO method. TW Hayes, IH Moal - Journal of chemical theory and computation, 2017 | ||

+ | *# https://en.wikipedia.org/wiki/Lasso_(statistics) | ||

+ | *# E. Frezza, R. Lavery, Internal normal mode analysis (iNMA) applied to protein conformational flexibility, Journal of Chemical Theory and Computation 11 (2015) 5503–5512. | ||

+ | * '''Base algorithm:''' The starting point will be a combination of methods from references 2 and 4. It has to be a LASSO formulation with the direction vectors reconstructed from the internal coordinates. The quality will be computed based on the RMSD measure between the prediction and the solution on several benchmarks. Results will be presented with statistical plots (see examples in references 3-4. | ||

+ | * '''Novelty:''' This is an important and open question in computational structural bioinformatics - how to efficiently represent transitions between protein structures. Not much has been done in the torsional angle subspace (internal coordinates)[6] and nearly nothing has been done using L1 regularization [4]. | ||

+ | * '''Authors:''' Ugo Bastolla on the torsional subspace (https://ub.cbm.uam.es/home/ugo.php), Sergei Grudinin on L1 minimization (https://team.inria.fr/nano-d/team-members/sergei-grudinin/) | ||

+ | |||

+ | ===10. 2018=== | ||

+ | * '''Title:''' Comparison of neural network and continuous-morphological methods in the problem of text detection (Text Detection). | ||

+ | * '''Problem:''' Automatically Detect Text in Natural Images. | ||

+ | * '''Data:''' Synthetic generated data + prepared sample of photos + [https://vision.cornell.edu/se3/coco-text-2/ COCO-Text dataset] + [http://www.machinelearning.ru/ Competition Avito 2014]. | ||

+ | * '''References:''' [https://vision.cornell.edu/se3/wp-content/uploads/2016/01/1601.07140v1.pdf COCO benchmark], [https://vision.cornell.edu/se3/wp-content/uploads/2016/01/1601.07140v1.pdf One of a state-of-the-art architecture] | ||

+ | * '''Base algorithm:''' [https://github.com/eragonruan/text-detection-ctpn code] + morphological methods, [http://www.machinelearning.ru/wiki/images/f/f1/Avito.ru-2014_Ulyanov_presentation.pdf Avito 2014 winner’s solution]. | ||

+ | * '''Solution:''' It is proposed to compare the performance of several state-of-the-art algorithms that need a large training set with morphological methods that require a small amount of data. It is proposed to determine the limits of applicability of certain methods. | ||

+ | * '''Novelty:''' propose an algorithm based on the use of both neural network and morphological methods (solution of the word detection problem). | ||

+ | * '''Authors:''' I. N. Zharikov. | ||

+ | * '''Expert''': L. M. Mestetsky (morphological methods). | ||

+ | |||

+ | ===16. 2018=== | ||

+ | * '''Title:''' Estimate of the optimal sample size for research in medicine | ||

+ | * '''Problem:''' In conditions of an insufficient number of expensive measurements, it is required to predict the optimal size of the replenished sample. | ||

+ | * '''Data:''' Samples of measurements in medical diagnostics, in particular, a sample of immunological markers. | ||

+ | * '''References:''' | ||

+ | *# Motrenko A.P. Materials on algorithms for estimating the optimal sample size in the MLAlgorithms repository [http://svn.code.sf.net/p/mlalgorithms/code/PhDThesis/Motrenko/doc/], [http://svn.code.sf.net/ p/mlalgorithms/code/Group874/Motrenko2014KL/]. | ||

+ | * '''Basic algorithm''': A series of empirical sample size estimation algorithms. | ||

+ | * '''Solution:''' Investigation of the properties of the parameter space when replenishing the sample. | ||

+ | * '''Novelty:''' A new methodology for sample size forecasting is proposed, justified in terms of classical and Bayesian statistics. | ||

+ | * '''Authors:''' A.M. Katrutsa, Strijov V.V., coordinator Tamaz Gadaev | ||

+ | |||

+ | ===19. 2018=== | ||

+ | * Name: Study of the dependence of the quality of recognition of ontological objects on the depth of hyponymy. | ||

+ | * '''Problem description:''' It is necessary to investigate the dependence of the quality of recognition of ontological objects at different levels of concept hyponymy. The classic formulation of the problem of named entity recognition: https://en.wikipedia.org/wiki/Named-entity_recognition | ||

+ | * Data: Hyponyms from https://wordnet.princeton.edu/ , texts from different domains presumably from WebOfScience. | ||

+ | * '''References:''' Relevant articles for classical staging http://arxiv-sanity.com/search?q=named+entity+recognition | ||

+ | * Basic algorithm: https://arxiv.org/pdf/1709.09686.pdf or its simplified version can be used as an algorithm, studies are performed using the DeepPavlov library. | ||

+ | * '''Solution:''' It is necessary to collect a dataset of hyponymy (nesting of concepts) of objects using WordNet, to automatically mark up ontological objects of texts of various domains for several levels of generalization of concepts, to conduct a series of experiments to determine the quality of recognition of ontological objects for different levels of nesting. | ||

+ | * '''Novelty:''' Similar studies have not been carried out, there are no ready-made datasets with a hierarchical markup of objects. Recognition of ontological objects at various levels of hyponymy can be used to produce additional features when solving various NLP (Natural language processing) The problems, as well as determining whether objects are a hyponym-hypernym pair. | ||

+ | * '''Authors:''' Burtsev Mikhail Sergeevich (Expert), Baimurzina Dilyara Rimovna (consultant). | ||

+ | |||

+ | ===21. 2018=== | ||

+ | * '''Title:''' High order convex optimization methods | ||

+ | * '''Problem description:''' High-order methods are effectively (up to n ~ 10^3 sometimes even up to n ~ 10^4) used for convex problems of not very large dimensions. Until recently, it was generally accepted that these are second-order methods (using the second derivatives of the function being optimized). However, at the beginning of 2018 Yu.E. Nesterov [1] proposed an efficient third-order method in the theory, which works according to almost optimal estimates. In the manual [3] in exercise 1.3, an example of a "bad" convex function proposed by Yu.E. Nesterov, on which I would like to compare the Nesterov method of the second and third order [1], the method from [2] of the second and third order and the usual fast gradient methods (of the first order). It is worth comparing both by the number of iterations and by the total running time. | ||

+ | * '''References:''' | ||

+ | # https://alfresco.uclouvain.be/alfresco/service/guest/streamDownload/workspace/SpacesStore/aabc2323-0bc1-40d4-9653-1c29971e7bd8/coredp2018_05web.pdf?guest=true | ||

+ | # https://arxiv.org/pdf/1809.00382.pdf | ||

+ | # https://arxiv.org/pdf/1711.00394.pdf | ||

+ | * '''Author:''' Evgenia Alekseevna Vorontsova (Associate Professor of Far Eastern Federal University, Vladivostok), Alexander Vladimirovich Gasnikov | ||

+ | |||

+ | ===22. 2018=== | ||

+ | * '''Title:''' Cutting plane methods for copositive optimization | ||

+ | * '''Problem:''' Conic program over the copositive cone (copositive program) min <C,X> : <A_i,X> = b_i, X \in \Pi_i C^k_i, k_i <= 5 A linear function is minimized over the intersection of an affine subspace with a product of copositive cones of orders k_i <= 5. | ||

+ | * '''Data:''' The algorithm will be tested on randomly generated instances | ||

+ | * '''References:''' | ||

+ | *# [1] Peter J. C. Dickinson, Mirjam Dür, Luuk Gijben, Roland Hildebrand. Scaling relationship between the copositive cone and Parrilo’s first level approximation. Optim. Lett. 7(8), 1669—1679, 2013. | ||

+ | *# [2] Stefan Bundfuss, Mirjam Dür. Algorithmic copositivity detection by simplicial partition. Linear Alg. Appl. 428, 1511—1523, 2008. | ||

+ | *# [3] Mirjam Dür. Copositive programming — a Survey. In Recent advances in Optimization and its Applications in Engineering, Springer, pp. 3-20, 2010. | ||

+ | * '''Base algorithm:''' The reference algorithm is described in [4] Stefan Bundfuss, Mirjam Dür. An Adaptive Linear Approximation Algorithm for Copositive Programs. SIAM J. Optim., 20(1), 30-53, 2009. | ||

+ | * '''Solution:''' The copositive program will be solved by a cutting plane algorithm. The cutting plane (in the case of an infeasible iterate) will be constructed from the semidefinite representation of the diagonal 1 section of the cone proposed in [1]. The algorithm will be compared to a simplicial division method proposed in [2], [4]. General information about copositive programs and their applications in optimization can be found in [3] . | ||

+ | * '''Novelty:''' The proposed algorithm for optimization over copositive cones up to order 5 uses an exact semi-definite representation. In contrast to all other algorithms existing today the generation of cutting planes is non-iterative. | ||

+ | * '''Author''': [http://www-ljk.imag.fr/membres/Roland.Hildebrand/ Roland Hildebrand] | ||

+ | |||

+ | ===23. 2018=== | ||

+ | * '''Title:''' Fractal analysis and synthesis of optical images of sea waves | ||

+ | * '''Problem description:''' A variety of physical processes and phenomena are studied with the help of images obtained remotely. An important The problem is to obtain adequate information about the processes and phenomena of interest by measuring certain image characteristics. Lines of equal brightness (isolines) on the images of many natural objects are fractal, that is, they are sets of points that cannot be represented by lines of finite length and occupy an intermediate position between lines and two-dimensional flat figures. Such sets are characterized by the fractal dimension D, which generalizes the classical concept of the dimension of a set and can take fractional values. For a solitary point on the image D=0, for a smooth curve D=1, for a flat figure D=2. The fractal isoline has the dimension 1<D<2. The algorithm for calculating D is given, for example, in [1]. The fractal dimension of the sea surface isolines can serve to estimate the spatial spectra of sea waves according to remote sensing data [1]. The problem is as follows. It is necessary to conduct a numerical study of the relationship between the characteristics of the spatial spectra of sea waves and the fractal dimension of satellite images of the Earth in the solar glare region. For the study, the method of numerical synthesis of optical images of sea waves, described in [2], should be used. Numerical modeling should be done with different characteristics of sea waves, as well as with different positions of the Sun and spatial resolution of images. | ||

+ | * '''References:''' | ||

+ | *# Lupyan E. A., Murynin A. B. Possibilities of fractal analysis of optical images of the sea surface. // Preprint of the Space Research Institute of the Academy of Sciences of the USSR Pr.-1521, Moscow, 1989, 30 p. | ||

+ | *# Murynin A. B. Reconstruction of the spatial spectra of the sea surface from optical images in a nonlinear model of the brightness field // Research of the Earth from Space, 1990. No. 6. P. 60-70. | ||

+ | * '''Author:''' Ivan Alekseevich Matveev | ||

+ | |||

+ | ===24. 2018=== | ||

+ | * '''Name''' Entropy maximization for various types of image transformations | ||

+ | * '''Problem description:''' Pansharpening is an algorithm for upscaling multispectral images using a reference image. The problem of pansharpening is formulated as follows: having a panchromatic image of the required resolution and a multispectral image of reduced resolution, it is required to restore the multispectral image in the spatial resolution of the panchromatic one. From empirical observations based on a large number of high-resolution images, it is known that the spatial variability of the reflected radiation intensity for objects of the same nature is much greater than the variability of their spectrum. In other words, one can observe that the spectrum of reflected radiation is homogeneous within the boundaries of one object, while even within one object the intensity of reflected radiation varies. In practice, good results can be achieved using a simplified approach, in which it is assumed that if the intensity of neighboring regions differ significantly, then these regions probably belong to different objects with different reflected spectra. This is the basis for the developed probabilistic algorithm for increasing the resolution of multispectral images using a reference image [1] | ||

+ | * '''It is necessary''' to conduct a study on maximizing the entropy for various types of transformations on the image. Show that entropy can serve as an indicator of the loss of information contained in the image during transformations over it. Formulation of the inverse problem for image restoration: Condition 1: Correspondence of the intensity (at each point) of the restored image with the intensity of the panchromatic image. Condition 2: Correspondence of the low-frequency component of the reconstructed image with the original multispectral image. Condition 3: Homogeneity (similarity) of the spectrum within one object and the assumption of an abrupt change in the spectrum at the border of two homogeneous regions. Condition 4: Under the first three conditions, the local entropy of the reconstructed image must be maximized. | ||

+ | * '''References:''' | ||

+ | *# Gorohovsky K. Yu., Ignatiev V. Yu., Murynin A. B., Rakova K. O. Search for optimal parameters of a probabilistic algorithm for increasing the spatial resolution of multispectral satellite images // Izvestiya RAN. Theory and control systems, 2017, No. 6. | ||

+ | * '''Author:''' Ivan Alekseevich Matveev | ||

+ | |||

+ | ===25. 2018=== | ||

+ | * '''Title:''' Automatic detection and recognition of objects in images | ||

+ | * '''Problem description:''' Automatic detection and recognition of objects in images and videos is one of the main The problems of computer vision. As a rule, these The problems are divided into several subThe problems: preprocessing, extraction of the characteristic properties of the object image and classification. The pre-processing stage usually includes some operations on the image such as filtering, brightness equalization, geometric corrective transformations to facilitate robust feature extraction. | ||

+ | The characteristic properties of an image of an object are understood as a set of features that approximately describe the object of interest. Features can be divided into two classes: local and integral. The advantage of local features is their versatility, invariance with respect to uneven changes in brightness and illumination, but they are not unique. Integral features that characterize the image of the object as a whole are not resistant to changes in the structure of the object and difficult lighting conditions. There is a combined approach - the use of local features as elements of an integral description, when the desired object is modeled by a set of areas, each of which is characterized by its own set of features - a local texture descriptor. The totality of such descriptors characterizes the object as a whole. | ||

+ | Classification is understood as determining whether an object belongs to a particular class by analyzing the feature vector obtained at the previous stage, dividing the feature space into subdomains indicating the corresponding class. There are many approaches to classification: neural network, statistical (Bayesian, regression, Fisher, etc.), decision trees and forests, metric (nearest K-neighbors, Parzen windows, etc.) and nuclear (SVM, RBF, method of potential functions), compositional (AdaBoost). For The problem of detecting an object in an image, membership in two classes is evaluated - the class of images containing the object, and the class of images that do not contain the object (background images). | ||

+ | * [[Media:ThemesIS2018Video.pdf| References and more details here]] | ||

+ | * '''Author:''' Ivan Alekseevich Matveev | ||

+ | |||

+ | ===29. 2018=== | ||

+ | * Name: Cross-Language Document Extractive Summarization with Neural Sequence Model. | ||

+ | * '''Problem description:''' It is proposed to solve the transfer learning problem for the text reduction model by extractive summarization and to investigate the dependence of the quality of text reduction on the quality of training of the translation model. Having data for training the abbreviation model in English and a parallel English-Russian corpus of texts, build a model for abbreviating the text in Russian. The solution of the problem is evaluated on a small set of data for testing the model in Russian, the quality of the solution to the problem is determined by the ratio of the values of the ROUGE criteria in English and Russian sets. | ||

+ | * Data: Data for training the model in English (SummaRuNNer2016), OPUS parallel corpus, data for verification in Russian. | ||

+ | * '''References:''' The article (SummaRuNNer2016) describes the basic text reduction algorithm, the work Neural machine translation by jointly learning to align and translate.(NMT2016) describes the translation model. The idea of sharing models is presented in Cross-Language Document Summarization Based on Machine Translation Quality Prediction (CrossSum2010). | ||

+ | * Basic algorithm: One idea of the basic algorithm is presented in (CrossSum2010), a translation model is implemented (OpenNMT), an implementation of a text reduction model is provided (SummaRuNNer2016). | ||

+ | * '''Solution:''' It is suggested to explore the solution idea proposed in the article (CrossSum2010) and options for combining reduction and translation models. Basic models and dataset preprocessing implemented (OpenNMT), PyTorch and Tensorflow libraries. Analysis of text reduction errors is performed as described in (SummaRuNNer2016), analysis of the quality of model training by standard library tools, . | ||

+ | * '''Novelty:''' For the base model, the applicability was investigated on a couple of datasets, confirming the possibility of transferring training to a dataset in another language and specifying the conditions for this transfer will expand the scope of the model and indicate the necessary new refinements of the model or data preprocessing. | ||

+ | * '''Authors:''' Alexey Romanov (consultant), Anton Khritankov (Expert). | ||

+ | |||

+ | ===30. 2018=== | ||

+ | * Title: Method for constructing an HG-LBP descriptor based on gradient histograms for pedestrian detection. | ||

+ | * '''Problem description:''' It is proposed to develop a new descriptor that generalizes the LBP descriptor based on histograms of gradient modules, having HOG-LBP composition properties for The problem of detecting pedestrians in an image. As an analysis of the quality of a new descriptor, it is proposed to use FAR/FRR detection error plots based on INRIA. | ||

+ | * Data: INRIA pedestrian database: http://pascal.inrialpes.fr/data/human/ | ||

+ | * '''References:''' | ||

+ | *# T. Ojala and M. Pietikainen. Multiresolution Gray-Scale and Rotation Invariant Texture Classification with Local Binary Patterns, IEEE Trans on Pattern Analysis and Machine Intelligence, Vol. 24. No. 7, July, 2002. | ||

+ | *# T. Bouwmans, C. Silva, C. Marghes, M. Zitouni, H. Bhaskar, C. Frelicot, "On the Role and the Importance of Features for Background Modeling and Foreground Detection", https:// arxiv.org/pdf/1611.09099v1.pdf | ||

+ | *# N. Dalal and B. Triggs, Histograms of Oriented Gradients for Human Detection, Proc. IEEE Conference on Computer Vision and Pattern Recognition, 2005, pp.886-893 | ||

+ | *# T. Ahonen, A. Hadid, M. Pietikainen Face Description with Local Binary Patterns: Application to Face Recognition \\ IEEE Transactions on Pattern Analysis and Machine Intelligence, Volume:28 , Issue: 121. | ||

+ | *# http://www.magicandlove.com/blog/2011/08/26/people-detection-in-opencv-again/ | ||

+ | *# http://www.cse.oulu.fi/CMV/Downloads/LBPMatlab2 | ||

+ | *# http://www.mathworks.com/help/vision/ref/extractlbpfeatures.html3. | ||

+ | *# http://www.codeproject.com/Articles/741559/Uniform-LBP-Features-and-Spatial-Histogram-Computa4. | ||

+ | *# http://www.cse.oulu.fi/CMV/Research | ||

+ | * Basic algorithm: Xiaoyu Wang, Tony X. Han, Shuicheng Yan. An HOG-LBP Human Detector with Partial Occlusion Handling \\ ICCV 2009 | ||

+ | * '''Solution:''' One of the options for generalizing LBP can be to use instead of histograms of distribution of points by LBP code, histograms of distribution of modules of point gradients in a block by LBP code (HG-LBP). It is proposed to use the OpenCV library for the basis of experiments, in which the HOG and LBP algorithms are implemented. It is necessary to modify the source code of the LBP implementation and insert the calculation of the modules of the gradient and the accumulation of the corresponding histogram over the LBP. It is necessary to write a program for reading the INRIA base, learning the linear SVM method on the original and modified descriptors, collecting detection statistics and plotting FAR/FRR DET plots. | ||

+ | * '''Novelty:''' The development of computationally simple methods for extracting the most informative features in recognition The problems is relevant in the field of creating embedded systems with low computing resources. Replacing the composition of descriptors with one that is more informative than each individually can simplify the solution of the problem. The use of gradient values in LPB descriptor histograms is new. | ||

+ | * '''Authors:''' Gneushev Alexander Nikolaevich | ||

+ | |||

+ | ===31. 2018=== | ||

+ | * Name: Using the HOG descriptor to train a neural network in a pedestrian detection The problem | ||

+ | * '''Problem description:''' It is proposed to replace the linear SVM classifier in the classical HOG algorithm with a simple convolutional neural network of small depth, while the HOG descriptor should be represented by a three-dimensional tensor that preserves the spatial structure of local blocks. As an analysis of the quality of a new descriptor, it is proposed to use FAR/FRR detection error plots based on INRIA. | ||

+ | * Data: INRIA pedestrian database: http://pascal.inrialpes.fr/data/human/ | ||

+ | * '''References:''' | ||

+ | *# 1. N. Dalal and B. Triggs, Histograms of Oriented Gradients for Human Detection, Proc. IEEE Conference on Computer Vision and Pattern Recognition, 2005, pp.886-893 | ||

+ | *# 3. Q. Zhu, S. Avidan, M.-C. Yeh, and K.-T. Cheng. Fast human detection using a cascade of histograms of oriented gradients. In CVPR, pages 1491-1498, 2006 O. Tuzel, F. Porikli, and P. Meer. Human detection via classification on riemannian manifolds. In CVPR, 2007 | ||

+ | *# 4. P. Dollar, C. Wojek, B. Schiele and P. Perona Pedestrian Detection: An Evaluation of the State of the Art / IEEE Transactions on Pattern Analysis and Machine Intelligence (PAMI), Vol 34. Issue 4, pp . 743-761 | ||

+ | *# 5. Xiaoyu Wang, Tony X. Han, Shuicheng Yan, An HOG-LBP Human Detector with Partial Occlusion Handling, ICCV 2009 http://www.xiaoyumu.com/s/PDF/Wang_HOG_LBP.pdf | ||

+ | *# 6. https://en.wikipedia.org/wiki/Pedestrian_detection | ||

+ | *# 7. HOG person detector tutorial https://chrisjmccormick.wordpress.com/2013/05/09/hog-person-detector-tutorial/ | ||

+ | *# 8. NavneetDalalThesis.pdf Navneet Dalal. Finding People in Images and Videos. PhD Thesis. Institut National Polytechnique de Grenoble / INRIA Rhone-Alpes, Grenoble, July 2006) | ||

+ | *# 9. People Detection in OpenCV http://www.magicandlove.com/blog/2011/08/26/people-detection-in-opencv-again/ | ||

+ | *# 10. Andrew G. Howard, Menglong Zhu, Bo Chen, Dmitry Kalenichenko, Weijun Wang, Tobias Weyand, Marco Andreetto, Hartwig Adam. MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications | ||

+ | *Basic algorithm: | ||

+ | *# 1. N. Dalal and B. Triggs, Histograms of Oriented Gradients for Human Detection, Proc. IEEE Conference on Computer Vision and Pattern Recognition, 2005, pp.886-893 | ||

+ | *# 2. Xiaoyu Wang, Tony X. Han, Shuicheng Yan, An HOG-LBP Human Detector with Partial Occlusion Handling, ICCV 2009 | ||

+ | * '''Solution:''' One of the options for generalizing the HOG algorithm can be to use another classifier instead of the linear SVM algorithm, for example, some kind of neural network. It is proposed to use the OpenCV library for the basis of experiments, which implements the HOG algorithm and the SVM classifier. It is necessary to analyze the source code of the HOG implementation, formalize the internal structure of the descriptor HOG vector in the form of a three-dimensional tensor — two spatial and one spectral dimensions. It is necessary to write a program for reading the INRIA base, learning the linear SVM method on HOG descriptors from it, collecting detection statistics and plotting FAR/FRR DET plots. Based on some neural network training system (for example, mxnet), it is necessary to assemble a shallow (no more than 2-3 convolutional layers) convolutional neural network of known architecture, train it on the basis of INRIA and on HOG tensor descriptors, build the corresponding FAR / FRR graphs. | ||

+ | * '''Novelty:''' The development of computationally simple methods for extracting the most informative features in recognition The problems is relevant in the field of creating embedded systems with low computing resources. Using a small number of the most informative descriptors can reduce computational complexity compared to using a large composition of simple features, such as in a deep convolutional neural network. Typically, classifiers use the HOG descriptor as a vector as a whole, however, information about the local spatial structure and feature spectrum is lost. The novelty lies in the use of the block locality property in the HOG descriptor and the representation of the HOG as a 3D tensor. The use of this information makes it possible to achieve detection resistance to pedestrian overlap. | ||

+ | * '''Authors:''' Gneushev Alexander Nikolaevich | ||

+ | |||

+ | ==2017== | ||

+ | |||

+ | {|class="wikitable" | ||

+ | |- | ||

+ | ! Author | ||

+ | ! Topic | ||

+ | ! Links | ||

+ | ! Consultant | ||

+ | ! Reviewer | ||

+ | ! Report | ||

+ | ! Letters | ||

+ | ! <tex>\Sigma=3+13</tex> | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Goncharovalex Goncharov Alexey] | ||

+ | |Metric classification of time series | ||

+ | |[http://svn.code.sf.net/p/mlalgorithms/code/Group274/Goncharov2015MetricClassification/code code], | ||

+ | [http://svn.code.sf.net/p/mlalgorithms/code/Group274/Goncharov2015MetricClassification/doc/Goncharov2015MetricClassification.pdf paper], | ||

+ | [http://svn.code.sf.net/p/mlalgorithms/code/Group274/Goncharov2015MetricClassification/doc/GoncharovAlexey2015PresentationMetricClassification.pdf slides] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Mpopova Maria Popova] | ||

+ | |Zadayanchuk Andrey | ||

+ | |BMF | ||

+ | |AILSBRCVTDSWH> | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:AstakhovAnton Astakhov Anton] | ||

+ | | Restoring the structure of a predictive model from a probabilistic representation | ||

+ | |[https://svn.code.sf.net/p/mlalgorithms/code/Group574/Astakhov2018RestorePrognosticStructure/ folder] | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group574/Astakhov2018RestorePrognosticStructure/code/ code] | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group574/Astakhov2018RestorePrognosticStructure/doc/paper/Astakhov2018RestorePrognosticStructure.pdf paper] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Katrutsa Alexander Katrutsa] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:KislinskiVadim Kislinsky Vadim] | ||

+ | | BHF | ||

+ | |A-I-L0S0B0R0C0V0T0 [A-I-L-S-B0R0C0V0T0E0D0W0S] + [AILSBRCBTEDWS] | ||

+ | |2+4 | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:GavYur Gavrilov Yuri] | ||

+ | | Choice of Interpreted Multimodels in Credit Scoring The problems | ||

+ | |[https://svn.code.sf.net/p/mlalgorithms/code/Group574/Gavrilov2018CreditScoringMultimodels/ folder] | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group574/Gavrilov2018CreditScoringMultimodels/code/ code] | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group574/Gavrilov2018CreditScoringMultimodels/doc/paper/Gavrilov574CreditScoringMultimodels.pdf paper] | ||

+ | [https://youtu.be/ZOzprVyK8bc video] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Goncharovalex Goncharov Alexey] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Twelveth Ostroukhov Petr] | ||

+ | | BF | ||

+ | |A+IL-S0B-R0 [A+ILSBRC-VT0E0D0W0S] + (W) | ||

+ | | 2+9+1 | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Tamaz Gadaev Tamaz] | ||

+ | |Estimating the optimal sample size | ||

+ | |[https://svn.code.sf.net/p/mlalgorithms/code/Group574/Gadaev2018OptimalSampleSIze/ folder] | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group574/Gadaev2018OptimalSampleSIze/code/ code] | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group574/Gadaev2018OptimalSampleSIze/doc/paper/Gadaev2018OptimalSampleSize.pdf paper] | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group574/Gadaev2018OptimalSampleSIze/slides/Gadaev2018OptimalSample.pdf slides] | ||

+ | [https://youtu.be/N7UnR1cRTOI video] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Katrutsa Alexander Katrutsa] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:ShulginEgor Shulgin Egor] | ||

+ | |BHF | ||

+ | |A-IL>SB-R-C0V0T0 [AILSBR0CVT0E-D0W0S] | ||

+ | | 2+9 | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Egorgladin Gladin Egor] | ||

+ | |Accelerometer Battery Savings Based on Time Series Forecasting | ||

+ | |[https://svn.code.sf.net/p/mlalgorithms/code/Group574/Gladin2018AccelerometerChargeSaving/ folder] | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group574/Gladin2018AccelerometerChargeSaving/code code] | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group574/Gladin2018AccelerometerChargeSaving/doc/paper/Gladin2018AccelerometerChargeSaving.pdf paper] | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group574/Gladin2018AccelerometerChargeSaving/doc/slides slides] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Mvladimirova Maria Vladimirova] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:KozlinskyEvg Kozlinsky Evgeny] | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group574/Kozlinsky2018WNTMvsTM/review_on_Gladin.docx review] | ||

+ | |.F | ||

+ | |AILS [A-I-L-SB0R0C000V0T0E0D0W0S] | ||

+ | |1+4 | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Andriygav Grabovoi Andrey] | ||

+ | |Automatic determination of the relevance of neural network parameters. | ||

+ | |[http://svn.code.sf.net/p/mlalgorithms/code/Group574/Grabovoy2018OptimalBrainDamage/ folder] | ||

+ | [http://svn.code.sf.net/p/mlalgorithms/code/Group574/Grabovoy2018OptimalBrainDamage/code/ code] | ||

+ | [http://svn.code.sf.net/p/mlalgorithms/code/Group574/Grabovoy2018OptimalBrainDamage/doc/paper/Grabovoy2018OptimalBrainDamage.pdf paper] | ||

+ | [http://svn.code.sf.net/p/mlalgorithms/code/Group574/Grabovoy2018OptimalBrainDamage/doc/slides/Grabovoy2018OptimalBrainDamage.pdf slides] | ||

+ | [https://www.youtube.com/watch?v=OnW3t5jk-r0&feature=youtu.be video] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Oleg_Bakhteev Oleg Bakhteev] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Oleksandr_Kulkov Kulkov Alexander] | ||

+ | |BHMF | ||

+ | | A+ILS+BRC+VTE>D> [AILSBRCVTEDWS] [<tex>\emptyset</tex>] | ||

+ | |3+13 | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Nurlanov_zh Nurlanov Zhakshylyk] | ||

+ | | Deep Learning for reliable detection of tandem repeats in 3D protein structures | ||

+ | |[http://svn.code.sf.net/p/mlalgorithms/code/Group574/Nurlanov2018DeepSymmetry/ folder] | ||

+ | [http://svn.code.sf.net/p/mlalgorithms/code/Group574/Nurlanov2018DeepSymmetry/code/ code] | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group574/Nurlanov2018DeepSymmetry/doc/paper/Nurlanov2018DeepSymmetry.pdf paper] | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group574/Nurlanov2018DeepSymmetry/doc/slides/Nurlanov2018DeepSymmetry.pdf slides] | ||

+ | [https://youtu.be/y_HKeBlj45s video] | ||

+ | |[https://team.inria.fr/nano-d/team-members/sergei-grudinin/ S. V. Grudinin], Guillaume Pages | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Nikita_Pletnev Pletnev Nikita] | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group574/Nurlanov2018DeepSymmetry/feedback/Pletnev2018Recension.pdf Review] | ||

+ | |BHF | ||

+ | |AILB [A-I-LS-BRC0V0T-E0D0W0S] | ||

+ | |2+7 | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:AnnRogozina Rogozina Anna] | ||

+ | | Deep learning for RNA secondary structure prediction | ||

+ | |[http://svn.code.sf.net/p/mlalgorithms/code/Group574/Rogozina2018StructurePredictionRNA/ folder] | ||

+ | [http://svn.code.sf.net/p/mlalgorithms/code/Group574/Rogozina2018StructurePredictionRNA/code/ code] | ||

+ | [http://svn.code.sf.net/p/mlalgorithms/code/Group574/Rogozina2018StructurePredictionRNA/doc/paper/Rogozina2018StructurePredictionRNA.pdf paper] | ||

+ | [http://svn.code.sf.net/p/mlalgorithms/code/Group574/Rogozina2018StructurePredictionRNA/doc/slides/Rogozina2018RNAPredictionsSlides.pdf slides] | ||

+ | [https://youtu.be/r6S5_5b24hg video] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Mpopova Maria Popova] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Tamaz Gadaev Tamaz] | ||

+ | |BHMF | ||

+ | |AILSBR> [AILSBRC0V0T0E0D0W0S]+CW | ||

+ | |3+9 | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Ol_terekhov Terekhov Oleg] | ||

+ | |Generation of features using locally approximating models | ||

+ | |[http://svn.code.sf.net/p/mlalgorithms/code/Group574/Terekhov2018LocallyApproxModels/ folder] | ||

+ | [http://svn.code.sf.net/p/mlalgorithms/code/Group574/Terekhov2018LocallyApproxModels/code/ code] | ||

+ | [http://svn.code.sf.net/p/mlalgorithms/code/Group574/Terekhov2018LocallyApproxModels/doc/Terekhov2018LocalApproxModels.pdf paper] | ||

+ | [http://svn.code.sf.net/p/mlalgorithms/code/Group574/Terekhov2018LocallyApproxModels/slides/Terekhov2018LAM_Presentation.pdf slides] | ||

+ | |S.D. Ivanychev, [http://www.machinelearning.ru/wiki/index.php?title=Участник:Neychev R.G. Neichev] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Egorgladin Gladin Egor] | ||

+ | [http://svn.code.sf.net/p/mlalgorithms/code/Group574/Terekhov2018LocallyApproxModels/doc/Gladin2018LAM_Review.pdf review] | ||

+ | |BHM | ||

+ | |AILSBRCVTDSW [AIL0SB0R0C0V0TE0D0W0S] | ||

+ | |2+12 | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:ShulginEgor Shulgin Egor] | ||

+ | | Generation of features that are invariant to changes in the frequency of the time series | ||

+ | |[http://svn.code.sf.net/p/mlalgorithms/code/Group574/Shulgin2018InvariantFeatureGeneration/ folder] | ||

+ | [http://svn.code.sf.net/p/mlalgorithms/code/Group574/Shulgin2018InvariantFeatureGeneration/code/ code] | ||

+ | [http://svn.code.sf.net/p/mlalgorithms/code/Group574/Shulgin2018InvariantFeatureGeneration/doc/paper/ paper] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Neychev R.G. Neichev] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Ol_terekhov Terekhov Oleg] | ||

+ | | BHM | ||

+ | |AIL [AI-LS-BR0CV0T0E0D0W0S] | ||

+ | | 2+5 | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Gmalinovsky Malinovsky Grigory] | ||

+ | |Graph Structure Prediction of a Neural Network Model | ||

+ | |[http://svn.code.sf.net/p/mlalgorithms/code/Group574/Malinovskyi2018StructureCNN/ folder] | ||

+ | [http://svn.code.sf.net/p/mlalgorithms/code/Group574/Malinovskyi2018StructureCNN/code/ code] | ||

+ | [https://gmalinovskyi@svn.code.sf.net/p/mlalgorithms/code/Group574/Malinovskyi2018StructureCNN/paper/Malinovskyi2018GraphStructure.pdf paper] | ||

+ | [http://svn.code.sf.net/p/mlalgorithms/code/Group574/Malinovskyi2018StructureCNN/paper/Malinovskyi2018NeuralStructureF_talk.pdf slides] | ||

+ | [https://youtu.be/GjsJxE6Msbg video] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Oleg_Bakhteev Oleg Bakhteev] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Andriygav Grabovoi Andrey] | ||

+ | [http://svn.code.sf.net/p/mlalgorithms/code/Group574/Malinovskyi2018StructureCNN/paper/Grabovoy2018GraphStructure_Review.pdf review] | ||

+ | | BHMF | ||

+ | | A+I+L+SBR>C>V>T>E>D> [AILSBRC0VTED0WS]+(C) | ||

+ | | 3+11 | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Oleksandr_Kulkov Kulkov Alexander] | ||

+ | |Brain signal decoding and intention prediction | ||

+ | |[http://svn.code.sf.net/p/mlalgorithms/code/Group574/Kulkov2018PartialLeastSquares/ folder] | ||

+ | [http://svn.code.sf.net/p/mlalgorithms/code/Group574/Kulkov2018PartialLeastSquares/code/ code] | ||

+ | [http://svn.code.sf.net/p/mlalgorithms/code/Group574/Kulkov2018PartialLeastSquares/doc/kulkov2018_pls.pdf paper] | ||

+ | [http://svn.code.sf.net/p/mlalgorithms/code/Group574/Kulkov2018PartialLeastSquares/doc/slides/kulkov2018_pls.pdf slides] | ||

+ | [https://youtu.be/7TLzV-oK7mk video] | ||

+ | |[[http://www.machinelearning.ru/wiki/index.php?title=Участник:Isachenkoroma R.V. Isachenko] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Gmalinovsky Malinovsky Grigory] | ||

+ | [https://sourceforge.net/p/mlalgorithms/code/13746/#diff-1 review] | ||

+ | | BHMF | ||

+ | | AILSBR [AILSBRCVTED0W0S] | ||

+ | | 3+11 | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Nikita_Pletnev Pletnev Nikita] | ||

+ | |Approximation of the boundaries of the iris | ||

+ | |[http://svn.code.sf.net/p/mlalgorithms/code/Group574/Pletnev2018IrisApproximation/paper/Pletnev2018IrisApproximation.pdf paper] | ||

+ | [http://svn.code.sf.net/p/mlalgorithms/code/Group574/Pletnev2018IrisApproximation/slides/Pletnev2018IrisApproximationSlides.pdf slides] | ||

+ | [ video] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Aduenko Alexander Aduenko] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Nurlanov_zh Nurlanov Zhakshylyk] | ||

+ | |BF | ||

+ | |AILSB>R> [AILSTWS] | ||

+ | | 2+7 | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Twelveth Ostroukhov Petr] | ||

+ | |Selection of models superposition for identification of a person on the basis of a ballistocardiogram | ||

+ | |[https://sourceforge.net/p/mlalgorithms/code/HEAD/tree/Group374/Ostroukhov2018BCGIdentification/ folder] | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group374/Ostroukhov2018BCGIdentification/doc/Ostroukhov2018BCGIdentification.pdf paper] | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group374/Ostroukhov2018BCGIdentification/code/ code] | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group374/Ostroukhov2018BCGIdentification/slides/Ostroukhov2018BCGIdentification_slides.pdf slides] | ||

+ | |Alexander Prozorov | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:GavYur Gavrilov Yuri] | ||

+ | [http://svn.code.sf.net/p/mlalgorithms/code/Group574/Gavrilov2018CreditScoringMultimodels/ReviewOnOstroukhov.pdf review] | ||

+ | |BhF | ||

+ | |AIL>S?B?R? [AILSBRCVT-E0D0W0S] | ||

+ | | 2+10 | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:KislinskiVadim Kislinsky Vadim] | ||

+ | |Predicting user music playlists in a recommender system. | ||

+ | |[https://svn.code.sf.net/p/mlalgorithms/code/Group574/Kislinskiy2018APContinuation/ folder] | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group574/Kislinskiy2018APContinuation/code code] | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group574/Kislinskiy2018APContinuation/doc/slides/Kislinskiy2018APContinuation.pdf slides] | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group574/Kislinskiy2018APContinuation/doc/paper/Kislinskiy2018APcontinution.pdf paper] | ||

+ | [https://youtu.be/YTqe9dkVgyw video] | ||

+ | | Evgeny Frolov | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:AstakhovAnton Astakhov Anton] | ||

+ | | .F | ||

+ | | (AIL)------(SB)---(RCVT)-- [AILS-BRCVTED0W0S] | ||

+ | | 1+11 | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:KozlinskyEvg Kozlinsky Evgeny] | ||

+ | | Analysis of banking transactional data of individuals to identify customer consumption patterns. | ||

+ | |[https://svn.code.sf.net/p/mlalgorithms/code/Group574/Kozlinsky2018WNTMvsTM/ folder] | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group574/Kozlinsky2018WNTMvsTM/code/ code] | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group574/Kozlinsky2018WNTMvsTM/doc/paper/kozlinsky18wntm-individuals.pdf paper] | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group574/Kozlinsky2018WNTMvsTM/doc/slides/analiz-tranzaktsii-slash.pdf slides] | ||

+ | [https://youtu.be/0WCyndULNIM video] | ||

+ | | Rosa Aisina | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:AnnRogozina Rogozina Anna] | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group574/Kozlinsky2018WNTMvsTM/doc/paper/Kozlinsky18wntm-individuals_Review.pdf review] | ||

+ | | BHMF | ||

+ | | AILSBR>CV> [AILSBR0C0V0TE0D0WS]+(С) | ||

+ | | 3+8+1 | ||

+ | |- | ||

+ | |} | ||

+ | |||

+ | |||

+ | ===1 === | ||

+ | * '''Title:''' Approximation of the boundaries of the iris | ||

+ | * '''Problem:''' Based on the image of the human eye, determine the circles approximating the inner and outer border of the iris. | ||

+ | * '''Data:''' Bitmap monochrome images, typical size 640*480 pixels (however other sizes are possible)[http://www.bath.ac.uk/elec-eng/research/sipg/irisweb/ ], [http://www.cb-sr.ia.ac.cn/IrisDatabase.htm]. | ||

+ | * '''References:''' | ||

+ | *# Aduenko A.A. Selection of multi-models in The problems classification (supervisor Strijov V.V.). Moscow Institute of Physics and Technology, 2017. [http://www.frccsc.ru/sites/default/files/docs/ds/002-073-05/diss/11-aduenko/11-Aduenko_main.pdf?626] | ||

+ | *# K.A. Gankin, A.N. Gneushev, I.A. Matveev Segmentation of the iris image based on approximate methods with subsequent refinements // Izvestiya RAN. Theory and control systems, 2014, no. 2, p. 78–92. | ||

+ | *# Duda, R. O. Use of the Hough transformation to detect lines and curves in pictures / R. O. Duda, P. E. Hart // Communications of the ACM. 1972 Vol. 15, no. 1.Pp. | ||

+ | * '''Basic algorithm''': Efimov Yury. Search for the outer and inner boundaries of the iris in the eye image using the paired gradient method, 2015. | ||

+ | * '''Solution:''' See [[Media:Iris_circle_problem.pdf | iris_circle_problem.pdf]] | ||

+ | * '''Novelty:''' A fast non-enumerative algorithm for approximating boundaries using linear multimodels is proposed. | ||

+ | * '''consultant''': Alexander Aduenko (by Strijov V.V., Expert Matveev I.A.) | ||

+ | |||

+ | ===2 === | ||

+ | * '''Title:''' Estimated optimal sample size | ||

+ | * '''Problem:''' In conditions of an insufficient number of expensive measurements, it is required to predict the optimal size of the replenished sample. | ||

+ | * '''Data:''' Samples of measurements in medical diagnostics, in particular, a sample of immunological markers. | ||

+ | * '''References:''' | ||

+ | *# Motrenko A.P. Materials on algorithms for estimating the optimal sample size in the MLAlgorithms repository [http://svn.code.sf.net/p/mlalgorithms/code/PhDThesis/Motrenko/doc/], [http://svn.code.sf.net/ p/mlalgorithms/code/Group874/Motrenko2014KL/]. | ||

+ | * '''Basic algorithm''': Sample size estimation algorithms for . | ||

+ | * '''Solution:''' Investigation of the properties of the parameter space when replenishing the sample. | ||

+ | * '''Novelty:''' A new methodology for sample size forecasting is proposed, justified in terms of classical and Bayesian statistics. | ||

+ | * '''Authors:''' A.M. Katrutsa, Strijov V.V., Expert A.P. Motrenko | ||

+ | |||

+ | ===3 === | ||

+ | * '''Title:''' Restoring the structure of the prognostic model from a probabilistic representation | ||

+ | * '''Problem:''' It is required to reconstruct the superposition tree from the generated connection probability graph. | ||

+ | * '''Data:''' Segments of time series, spatio-temporal series (and text collections). | ||

+ | * '''References:''' | ||

+ | *# Works by Tommy Yakkola and others at LinkReview [https://docs.google.com/document/d/1j-1eZ4Az05yBR3GvgZusqFVIZeE_HcZDawZDzz41zS4/edit?usp=sharing]. | ||

+ | * '''Basic algorithm''': Branch and bound method, dynamic programming when building a fully connected graph. | ||

+ | * '''Solution:''' Building a model in the form of GAN, VAE generates a weighted graph, NN approximates a tree structure. | ||

+ | * '''Novelty:''' Suggested a way to penalize a graph for not being a tree. A method for predicting the structures of prognostic models is proposed. | ||

+ | * '''Authors:''' A.M. Katrutsa, Strijov V.V. | ||

+ | |||

+ | ===4 === | ||

+ | * '''Title:''' Text recognition based on skeletal representation of thick lines and convolutional networks | ||

+ | * '''Problem:''' It is required to build two CNNs, one recognizes a bitmap representation of an image, the other a vector one. | ||

+ | * '''Data:''' Bitmap fonts. | ||

+ | * '''References:''' List of works [http://www.machinelearning.ru/wiki/images/a/a2/Morozov2017Synthesis_of_medicines.pdf], in particular arXiv:1611.03199 and | ||

+ | * '''Basic algorithm''': Convolution network for bitmap. | ||

+ | * '''Solution:''' It is required to propose a method for collapsing graph structures, which allows generating an informative description of the skeleton of a thick line. | ||

+ | * '''Novelty:''' A way to improve the quality of recognition of thick lines due to a new way of generating their descriptions is proposed. | ||

+ | * '''Authors:''' L.M. Mestetsky, I.A. Reyer, Strijov V.V. | ||

+ | |||

+ | ===5 === | ||

+ | * '''Title:''' Generation of features using locally approximating models | ||

+ | * '''Problem:''' It is required to test the feasibility of the hypothesis of simplicity of sampling for the generated features. Features are the optimal parameters of approximating models. Moreover, the entire sample is not simple and requires a mixture of models to approximate it. Explore the information content of the generated features - the parameters of the approximating models trained on the segments of the original time series. | ||

+ | * '''Data:''' | ||

+ | *# WISDM (Kwapisz, J.R., G.M. Weiss, and S.A. Moore. 2011. Activity recognition using cell phone accelerometers. ACM SigKDD Explorations Newsletter. 12(2):74–82.), USC-HAD or higher. Accelerometer data (Human activity recognition using smart phone embedded sensors: A Linear Dynamical Systems method, W Wang, H Liu, L Yu, F Sun - Neural Networks (IJCNN), 2014) | ||

+ | *# ([[Time series (examples library)]], Accelerometry section). | ||

+ | * '''References:''' | ||

+ | *# Kuznetsov M.P., Ivkin N.P. Algorithm for Classifying Accelerometer Time Series by Combined Feature Description // Machine Learning and Data Analysis. 2015. V. 1, No. 11. C. 1471-1483. [http://jmlda.org/papers/doc/2015/no11/Ivkin2015TSclassification.pdf] | ||

+ | *# Karasikov M.E., Strijov V.V. Classification of time series in the space of parameters of generating models // Informatics and its applications, 2016.[http://strijov.com/papers/Karasikov2016TSC.pdf URL] | ||

+ | *# Kuznetsov M.P., Ivkin N.P. Algorithm for Classifying Accelerometer Time Series by Combined Feature Description // Machine Learning and Data Analysis. 2015. V. 1, No. 11. C. 1471 - 1483. [http://jmlda.org/papers/doc/2015/no11/Ivkin2015TSclassification.pdf URL] | ||

+ | *# Isachenko R.V., Strijov V.V. Metric learning in The problemx multiclass classification of time series // Informatics and its applications, 2016, 10(2) : 48-57. [http://strijov.com/papers/Isachenko2016MetricsLearning.pdf URL] | ||

+ | *# Zadayanchuk A.I., Popova M.S., Strijov V.V. Choosing the optimal model for classifying physical activity based on accelerometer measurements // Information technologies, 2016. [http://strijov.com/papers/Zadayanchuk2015OptimalNN4.pdf URL] | ||

+ | *# Motrenko A.P., Strijov V.V. Extracting fundamental periods to segment human motion time series // Journal of Biomedical and Health Informatics, 2016, Vol. 20, no. 6, 1466 - 1476. | ||

+ | *# Ignatov A., Strijov V. Human activity recognition using quasiperiodic time series collected from a single triaxial accelerometer // Multimedia Tools and Applications, 2015, 17.05.2015 : 1-14. [http://strijov.com/papers/Ignatov2015HumanActivity.pdf URL] | ||

+ | * '''Basic algorithm''': Described by Kuznetsov, Ivkin. | ||

+ | * '''Solution:''' It is required to build a set of locally approximating models and choose the most adequate ones. | ||

+ | * '''Novelty:''' A standard for building locally approximating models has been created. | ||

+ | * '''Authors:''' S.D. Ivanychev, R.G. Neichev, Strijov V.V. | ||

+ | |||

+ | ===6 === | ||

+ | * '''Title:''' Brain signal decoding and intention prediction | ||

+ | * '''Problem:''' It is required to build a model that restores the movement of the limbs from the corticogram. | ||

+ | * '''Data:''' neurotycho.org [http://neurotycho.org/] | ||

+ | * '''References:''' | ||

+ | *# Neichev R.G., Katrutsa A.M., Strijov V.V. Selection of the optimal set of features from a multicorrelated set in the forecasting problem. Zavodskaya Lab. Diagnostics of materials, 2016, 82(3) : 68-74. [http://strijov.com/papers/Neychev2015FeatureSelection.pdf] | ||

+ | *# MLAlgorithms: Motrenko, Isachenko (submitted) | ||

+ | * '''Basic algorithm''': Partial Least Squares[https://en.wikipedia.org/wiki/Partial_least_squares_regression] | ||

+ | * '''Solution:''' Create a feature selection algorithm alternative to PLS and taking into account the non-orthogonal structure of feature interdependence. | ||

+ | * '''Novelty:''' A feature selection method is proposed that takes into account the regularities of both the and independent variable and the dependent variable. | ||

+ | * '''Authors:''' R.V. Isachenko, Strijov V.V. | ||

+ | |||

+ | ===7 === | ||

+ | * '''Title:''' Automatic determination of the relevance of neural network parameters. | ||

+ | * '''Problem:''' The problem of finding a stable (and not redundant in terms of parameters) neural network structure is considered. To cut off redundant parameters, it is proposed to introduce a priori probabilistic assumptions about the distribution of parameters and remove non-informative parameters from the neural network using the Belsley method. To adjust the prior distribution, it is proposed to use gradient methods. | ||

+ | * '''Data:''' A selection of handwritten MNIST digits | ||

+ | * '''Basic algorithm''': Optimal Brain Damage, decimation based on variance inference. The structure of the final model is proposed to be compared with the model obtained by the AdaNet algorithm. | ||

+ | * '''References:''' | ||

+ | *# [https://arxiv.org/pdf/1502.03492.pdf] Gradient hyperparameter optimization methods. | ||

+ | *# [http://proceedings.mlr.press/v48/luketina16.pdf] Gradient hyperparameter optimization methods. | ||

+ | *# [http://yann.lecun.com/exdb/publis/pdf/lecun-90b.pdf] Optimal Brain Damage. | ||

+ | *# [https://arxiv.org/abs/1607.01097] AdaNet | ||

+ | *# [http://strijov.com/papers/SanduleanuStrijov2011FeatureSelection_Preprint.pdf] Belsley Method | ||

+ | * '''Authors:''' Oleg Bakhteev, Strijov V.V. | ||

+ | |||

+ | ===8 === | ||

+ | * '''Title:''' Prediction of the graph structure of the neural network model. | ||

+ | * '''Problem:''' The problem is considered to find a stable (and non-redundant in terms of parameters) structure of a convolutional neural network. It is proposed to predict the structure of a neural network using doubly-recurrent neural networks. As a training sample, it is proposed to use the structures of models that have shown good quality on subsamples of small power. | ||

+ | * '''Data:''' Samples MNIST, CIFAR-10 | ||

+ | * '''Basic algorithm''': random search. Comparison with work on reinforcement learning is possible. | ||

+ | * '''References:''' | ||

+ | *# [https://pdfs.semanticscholar.org/e7bd/0e7a7ee6b0904d5de6e76e095a6a3b88dd12.pdf] doubly-recurrent neural networks. | ||

+ | *# [https://arxiv.org/pdf/1707.07012] Similar approach using reinforcement learning. | ||

+ | * '''Authors:''' Oleg Bakhteev, Strijov V.V. | ||

+ | |||

+ | ===9=== | ||

+ | * '''Title:''' Deep Learning for reliable detection of tandem repeats in 3D protein structures [[Media:Strijov_3D_CNN.pdf|more in PDF]] | ||

+ | * '''Problem:''' Deep learning algorithms pushed computer vision to a level of accuracy comparable or higher than a human vision. Similarly, we believe that it is possible to recognize the symmetry of a 3D object with a very high reliability, when the object is represented as a density map. The optimization problem includes i) multiclass classification of 3D data. The output is the order of symmetry. The number of classes is ~10-20 ii) multioutput regression of 3D data. The output is the symmetry axis (a 3-vector). The input data are typically 24x24x24 meshes. The total amount of these meshes is of order a million. Biological motivation : Symmetry is an important feature of protein tertiary and quaternary structures that has been associated with protein folding, function, evolution, and stability. Its emergence and ensuing prevalence has been attributed to gene duplications, fusion events, and subsequent evolutionary drift in sequence. Methods to detect these symmetries exist, either based on the structure or the sequence of the proteins, however, we believe that they can be vastly improved. | ||

+ | * '''Data:''' Synthetic data are obtained by ‘symmetrizing’ folds from top8000 library (http://kinemage.biochem.duke.edu/databases/top8000.php). | ||

+ | * '''References:''' Our previous 3D CNN: [https://arxiv.org/abs/1801.06252] Invariance of CNNs (and references therein): [https://hal.inria.fr/hal- 01630265/document], [https://arxiv.org/pdf/1706.03078.pdf] | ||

+ | * '''Base algorithm:''' A prototype has already been created using the Tensorflow framework [4], which is capable of detecting the order of cyclic structures with about 93% accuracy. The main goal of this internship is to optimize the topology of the current neural network prototype and make it rotational and translational invariant with respect to input data. [4] [https://www.tensorflow.org/] | ||

+ | * '''Solution:''' The network architecture needs to be modified according to the invariance properties (most importantly, rotational invariance). Please see the links below [https://hal.inria.fr/hal-01630265/document], | ||

+ | [https://arxiv.org/pdf/1706.03078.pdf] The code is written using the Tensorflow library, and the current model is trained on a single GPU (Nvidia Quadro 4000)of a desktop machine. | ||

+ | * '''Novelty:''' Applications of convolutional networks to 3D data are still very challenging due to large amount of data and specific requirements to the network architecture. More specifically, the models need to be rotationally and transnationally invariant, which makes classical 2D augmentation tricks loosely applicable here. Thus, new models need to be developed for 3D data. | ||

+ | * '''Authors:''' Expert Sergei Grudinin, consultants Guillaume Pages, Strijov V.V. | ||

+ | |||

+ | ===10=== | ||

+ | * '''Title:''' Semi-supervised representation learning with attention | ||

+ | * '''Problem:''' training of vector representations using the attention mechanism, thanks to which the quality of machine translation has increased significantly. It is proposed to use it in the encoder-decoder architecture network to obtain vectors of text fragments of arbitrary length. | ||

+ | * '''Data:''' It is proposed to consider two samples: Microsoft Paraphrase Corpus (a small set of proposals, https://www.microsoft.com/en-us/download/details.aspx?id=52398) and PPDB (a set of short segments, not always correct markup. http://sitem.herts.ac.uk/aeru/ppdb/en/) | ||

+ | * '''References:''' | ||

+ | *# Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, Illia Polosukhin. Attention Is All You Need (https://arxiv.org/abs/1706.03762). | ||

+ | *# John Wieting, Mohit Bansal, Kevin Gimpel, Karen Livescu. Towards Universal Paraphrastic Sentence Embeddings (https://arxiv.org/abs/1511.08198). | ||

+ | *# Ryan Kiros, Yukun Zhu, Ruslan Salakhutdinov, Richard S. Zemel, Antonio Torralba, Raquel Urtasun, Sanja Fidler. Skip Thought Vectors (https://arxiv.org/abs/1506.06726). | ||

+ | *# Keras seq2seq (https://github.com/farizrahman4u/seq2seq). | ||

+ | * '''Basic algorithm''': solution [3] or vector representations obtained using seq2seq[]. | ||

+ | * '''Solution:''' in The problem it is proposed to train vector representations for phrases using the attention and partial learning mechanism. As an internal quality functional, it is proposed to use the improved error function from [2]. As an applied problem, we can consider the problem of detecting paraphrases and sentiment analysis. Moreover, based on the results obtained in [1], it can be assumed that the attention mechanism has a greater influence on obtaining universal vectors for phrases than the network architecture. It is proposed to test this hypothesis using two different architectures - a standard recurrent and feed-forward network. | ||

+ | * '''Novelty:''' new method. | ||

+ | * '''Authors:''' Rita Kuznetsova, consultant | ||

+ | |||

+ | ===11 === | ||

+ | * '''Title:''' Selection of Interpreted Multi-Models in Credit Scoring The problems | ||

+ | * '''Problem:''' The problem of credit scoring is to determine the level of creditworthiness of the borrower. For this, a borrower's questionnaire is used, containing both numerical (age, income) and categorical features (gender, profession). It is required, having historical information about the repayment of loans by other borrowers, to determine whether the borrower will return the loan. The data can be heterogeneous (example, if there are different income regions in a country), and several models will be needed to adequately classify. It is necessary to determine the optimal number of models. Based on the set of model parameters, it is necessary to draw up a portrait of the borrower. | ||

+ | * '''Data:''' It is proposed to consider five samples from the UCI and Kaggle repositories, with a capacity of 50,000 objects or more. | ||

+ | * '''References:''' A.A. Aduenko \MLAlgorithms\PhDThesis; C. Bishop, Pattern recognition and machine learning, final chapter; 20 years of Mixture experts. | ||

+ | * '''Base algorithm:''' Clustering and building independent logistic regression models, Adaboost, Decision Forest (with restrictions on complexity), Blend of Experts. | ||

+ | * '''Solution:''' An algorithm is proposed for selecting a multi-model (a mixture of models or a mixture of Experts) and determining the optimal number of models. | ||

+ | * '''Novelty:''' Proposed function of distance between models in which parameter distributions are given on different media. | ||

+ | * '''Authors:''' Goncharov Alexey, Strijov V.V. | ||

+ | |||

+ | ===12 === | ||

+ | * '''Title:''' Generation of features that are invariant to changes in the frequency of the time series. | ||

+ | * '''Problem:''' Informally: there is a set of time series of a certain frequency (s1), and the information we are interested in is distinguishable and at a lower sampling rate (in the example, the samples occur every millisecond, and the events of interest to us occur at an interval of 0.1 s). These series are integrated reducing the frequency by a factor of 10 (i.e. every 10 values are simply summed) and a set of time series s2 is obtained. be described in the same way. Formally: Given a set of time series s1, .., sNS with a high sampling rate 1. Target information (example, hand movement/daily price fluctuation/…) is distinguishable and at a lower sampling rate 2 < 1. It is necessary to find such a mapping f: S G, - the frequency of the series, that it will generate similar feature descriptions for series of different frequencies. Those. | ||

+ | f* = argminf E(f1(s1) -f2(s2)) , where E is some error function. | ||

+ | * '''Data:''' Sets of time series of people's physical activity from accelerometers; human EEG time series; time series of energy consumption of cities/industrial facilities. Sample link: UCI repository, our EEG and accelerometer samples. | ||

+ | * '''References:''' See above for Accelerometers | ||

+ | * '''Base algorithm:''' Fourier transform. | ||

+ | * '''Solution:''' Building an autoencoder with a partially fixed internal representation as the same time series with a lower frequency. | ||

+ | * '''Novelty:''' For time series, there is no “common approach” to analysis, in contrast, in the example, to image analysis. If you look at the problem abstractly, now the cat is defined as well as and the cat, which takes up half the space in the image. An analogy with time series suggests itself. Moreover, the nature of data in pictures and in time series is similar: in pictures there is a hierarchy between values along two axes (x and y), and in time series - one at a time - along the time axis. The hypothesis is that methods similar to image analysis will provide qualitative results. The resulting feature representation can be further used for classification and prediction of time series. | ||

+ | * '''Authors:''' R. G. Neichev, Strijov V.V. | ||

+ | |||

+ | ===18 === | ||

+ | * '''Title:''' Comparison of neural network and continuous morphological methods in the Text Detection The problem. | ||

+ | * '''Problem:''' Automatically Detect Text in Natural Images. | ||

+ | * '''Data:''' synthetic generated data + trained photo sample + [https://vision.cornell.edu/se3/coco-text-2/ COCO-Text dataset] + [http://www.machinelearning .ru/ Avito Competition 2014]. | ||

+ | * '''References:''' [https://vision.cornell.edu/se3/wp-content/uploads/2016/01/1601.07140v1.pdf COCO benchmark], [https://vision.cornell. edu/se3/wp-content/uploads/2016/01/1601.07140v1.pdf One of a state-of-the-art architecture] | ||

+ | * '''Base algorithm:''' [https://github.com/eragonruan/text-detection-ctpn code] + morphological methods, [http://www.machinelearning.ru/wiki/images/f/f1 /Avito.ru-2014_Ulyanov_presentation.pdf Avito 2014 winner's solution]. | ||

+ | * '''Solution:''' It is proposed to compare the performance of several state-of-the-art algorithms that need a large training set with morphological methods that require a small amount of data. It is proposed to determine the limits of applicability of certain methods. | ||

+ | * '''Novelty:''' propose an algorithm based on the use of both neural network and morphological methods (solution of the word detection problem). | ||

+ | * '''Authors:''' I.N. Zharikov. | ||

+ | * '''Expert''': L.M. Mestetsky (morphological methods). | ||

+ | |||

+ | ==2017 Group 2== | ||

+ | {|class="wikitable" | ||

+ | |- | ||

+ | ! Author | ||

+ | ! Topic | ||

+ | ! Link | ||

+ | ! Consultant | ||

+ | ! Reviewer | ||

+ | ! Report | ||

+ | ! Letters | ||

+ | !<tex>\Sigma=3+13</tex> | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Goncharovalex Goncharov Alexey] | ||

+ | |Metric classification of time series | ||

+ | |[http://svn.code.sf.net/p/mlalgorithms/code/Group274/Goncharov2015MetricClassification/code code], | ||

+ | [http://svn.code.sf.net/p/mlalgorithms/code/Group274/Goncharov2015MetricClassification/doc/Goncharov2015MetricClassification.pdf paper], | ||

+ | [http://svn.code.sf.net/p/mlalgorithms/code/Group274/Goncharov2015MetricClassification/doc/GoncharovAlexey2015PresentationMetricClassification.pdf slides] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Mpopova Maria Popova] | ||

+ | |Zadayanchuk Andrey | ||

+ | |BMF | ||

+ | |AILSBRCVTDSWH> | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:white2302 Belykh Evgeny] [http://www.machinelearning.ru/wiki/index.php?title=Участник:Alladdin Proskurin Alexander] | ||

+ | |Classification of superpositions of movements of physical activity | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/Group594/raw/master/ProskurinBelykh2018ClassificationOfPhysicalActivitySuperposition/ClassificationOfPhysicalActivitySuperposition.pdf paper] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/Group594/raw/master/ProskurinBelykh2018ClassificationOfPhysicalActivitySuperposition/ProskurinBelykh2018Presentation.pdf slides] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/Group594/tree/master/ProskurinBelykh2018ClassificationOfPhysicalActivitySuperposition/code code] | ||

+ | |Maria Vladimirova, Alexandra Malkova | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:IlyaSM Romanenko Ilya], [http://www.machinelearning.ru/wiki/index.php?title=Участник:popovkin Popovkin Andrey], [https://github.com/Intelligent-Systems-Phystech/Group594/raw/master/ProskurinBelykh2018ClassificationOfPhysicalActivitySuperposition/RomanenkoPopovkin2018ClassificationOfPhysicalActivitySuperposition_Review.pdf review] | ||

+ | [https://www.youtube.com/watch?v=QnjOlVVVu2k video] | ||

+ | |MF | ||

+ | |AILSBRC>V> [AILSBRC0VT0E0D0WS] CTD | ||

+ | |2+9 | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:zueva.nn Zueva Nadezhda] | ||

+ | |Style Change Detection | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/Group594/raw/master/Zueva2018TextStyleTransfer/StyleChangeDetection%20(10).pdf paper] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/Group594/blob/master/Zueva2018TextStyleTransfer/Zueva_Presentation_Plagiarism%20(2).pdf slides] | ||

+ | [https://www.youtube.com/watch?v=1-GWn5uYvsc video] | ||

+ | |Rita Kuznetsova | ||

+ | |Igashov Ilya, [https://drive.google.com/file/d/1I-IWRxh39VhZuU2FPzbJAwkqfdYRcqRV/view?usp=sharing review] | ||

+ | |BHMF | ||

+ | |AIL-S-B-R- [AILSBRCV0TE0D0WS] | ||

+ | |3+10 | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Igashov Igashov Ilya] | ||

+ | |Formulation and solution of an optimization problem combining classification and regression to estimate the binding energy of a protein and small molecules. | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/Group594/raw/master/Igashov2018ProteinLigandComplexes/Igashov2018ProteinLigandComplexes.pdf paper] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/Group594/raw/master/Igashov2018ProteinLigandComplexes/presentation/presentation.pdf slides] | ||

+ | [https://www.youtube.com/watch?v=U0rDFG0-lzE video] | ||

+ | |Sergei Grudinin, Maria Kadukova | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:vanderwardan Manucharyan Vardan], [https://github.com/Intelligent-Systems-Phystech/Group594/raw/master/Igashov2018ProteinLigandComplexes/Igashov2018ProteinLigandComplexes_Review.pdf review], [https://github.com/Intelligent-Systems-Phystech/Group594/raw/master/Igashov2018ProteinLigandComplexes/Igashov2018ProteinLigandComplexes_Correction.pdf correction] | ||

+ | |BHMF | ||

+ | |AILBS+BRHC>V> [AILSBRCVTE0D0WS] | ||

+ | |3+11 | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:kalugin_di Kalugin Dmitry] | ||

+ | |Graph Structure Prediction of a Neural Network Model | ||

+ | |[https://drive.google.com/file/d/1ZTP7Uhi622cj5BnItDmlz0k988Twd9UZ/view?usp=sharing paper] | ||

+ | [https://drive.google.com/file/d/1iErLatXyIoqjH9yDXBbATc9vuA_8dmgZ/view?usp=sharing slides] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Oleg_Bakhteev Oleg Bakhteev] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:zueva.nn Zueva Nadezhda] [https://drive.google.com/drive/u/1/folders/1SV29oCjnqnrmjZ_pb1iNGgukodwLk-Bf review] | ||

+ | |BHM | ||

+ | |AI-L-S--B0R0C0V0 [A-ILSBR0CVT0ED0WS] | ||

+ | |2+11 | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:vanderwardan Manucharyan Vardan] | ||

+ | |Prediction of properties and types of atoms in molecular graphs using convolutional networks | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/Group594/raw/master/Manucharyan2018AtomicTypePredictionInUsingCNN/doc/Manucharyan2018AtomicTypePredictionInUsingCNN.pdf paper], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/Group594/raw/master/Manucharyan2018AtomicTypePredictionInUsingCNN/slides/Manucharyan2018AtomicTypePredictionInUsingCNNPresentation.pdf slides], | ||

+ | [https://github.com/Intelligent-Systems-Phystech/Group594/blob/master/Manucharyan2018AtomicTypePredictionInUsingCNN/code/Manucharyan2018AtomicTypePredictionInUsingCNN.ipynb code] | ||

+ | [https://www.youtube.com/watch?v=sShO-zIbidE video] | ||

+ | |Sergei Grudinin, [http://www.machinelearning.ru/wiki/index.php?title=Участник:Kadukovam Maria Kadukova] | ||

+ | |Fattakhov Artur [https://github.com/Intelligent-Systems-Phystech/Group594/raw/master/Manucharyan2018AtomicTypePredictionInUsingCNN/rev.pdf review] | ||

+ | |BMF | ||

+ | |AILS>B> [AILSB0R0CV0TE0D0WS] VED | ||

+ | |3+7 | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:kirill_mouraviev Muraviev Kirill] | ||

+ | |Determination of neural network parameters to be optimized. | ||

+ | |[https://github.com/KirillMouraviev/science_publication/blob/master/doc/Muravyev2018ParameterOptimization.pdf paper], | ||

+ | [https://github.com/KirillMouraviev/science_publication/raw/master/doc/Muravyev2018FinalTalk.pdf slides], | ||

+ | [https://github.com/KirillMouraviev/science_publication/tree/master/code code] | ||

+ | [https://www.youtube.com/watch?v=1KkQnx249rU video] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Oleg_Bakhteev Oleg Bakhteev] | ||

+ | |Kalugin Dmitry [https://github.com/Intelligent-Systems-Phystech/Group594/blob/master/Muravyev2018ParameterOptimization/Muravyev2018ParameterOptimization_Review.pdf review] | ||

+ | |BHMF | ||

+ | |A+IL-S-B-RCVTED [AILSBRCV0TE0DWS] | ||

+ | |3+12 | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:diraria Murzin Dmitry], [http://www.machinelearning.ru/wiki/index.php?title=Участник:andnlv Danilov Andrey] | ||

+ | |Text recognition based on skeletal representation of thick lines and convolutional networks | ||

+ | |[https://rawgit.com/Intelligent-Systems-Phystech/Group594/master/DanilovMurzin2018TextRecognitionUsingSkeletonRepresentationAndCNN/doc/DanilovMurzin2018TextRecognitionUsingSkeletonRepresentationAndCNN.pdf paper], [https://rawgit.com/Intelligent-Systems-Phystech/Group594/master/DanilovMurzin2018TextRecognitionUsingSkeletonRepresentationAndCNN/slides/DanilovMurzin2018TextRecognitionUsingSkeletonRepresentationAndCNN.pdf slides], [https://github.com/Intelligent-Systems-Phystech/Group594/tree/master/DanilovMurzin2018TextRecognitionUsingSkeletonRepresentationAndCNN/code code] | ||

+ | [video] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Mest L. M. Mestetsky], [http://www.machinelearning.ru/wiki/index.php?title=Участник:Ivan_Reyer Ivan Reyer], Zharikov I. N. | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:kirill_mouraviev Muraviev Kirill] [https://github.com/Intelligent-Systems-Phystech/Group594/blob/master/DanilovMurzin2018TextRecognitionUsingSkeletonRepresentationAndCNN/%D0%A0%D0%B5%D1%86%D0%B5%D0%BD%D0%B7%D0%B8%D1%8F.docx?raw=true review] | ||

+ | |BHMF | ||

+ | |A+IL> [AILSB0R0CV0TE0D0WS] | ||

+ | |3+8 | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:popovkin Popovkin Andrey] [http://www.machinelearning.ru/wiki/index.php?title=Участник:IlyaSM Romanenko Ilya] | ||

+ | |Creation of ranking models for information retrieval systems. Algorithm for Predicting the Structure of Locally Optimal Models | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/Group594/raw/master/PopovkinRomanenko2018PredictionStructureOfIRFunctions/PredictionStructureOfIRFunctions.pdf paper] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/Group594/raw/master/PopovkinRomanenko2018PredictionStructureOfIRFunctions/RomanenkoPopovkin2018Presentation.pdf slides] | ||

+ | [https://github.com/IlRomanenko/Information-retrieval code] | ||

+ | [https://www.youtube.com/watch?v=wBUt1SIWDBA video] | ||

+ | |Kulunchakov Andrey, Strijov V.V. | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Alladdin Proskurin Alexander], [http://www.machinelearning.ru/wiki/index.php?title=Участник:white2302 Belykh Evgeny], [https://github.com/Intelligent-Systems-Phystech/Group594/raw/master/PopovkinRomanenko2018PredictionStructureOfIRFunctions/ProskurinBelykh2018PredictionStructureOfIRFunctions_Review.doc review] | ||

+ | |BHMF | ||

+ | |AILS0BC>V> [AILSBRC0VTED0WS] | ||

+ | |3+11 | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:fartuk Fattakhov Artur] | ||

+ | |Style Change Detection | ||

+ | |[https://github.com/Intelligent-Systems-Phystech/Group594/raw/master/Fattakhov2018TextStyleTransfer/Fattakhov2018.pdf paper] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/Group594/raw/master/Fattakhov2018TextStyleTransfer/final_slides_Fattakhov_ansamble.pdf slides] | ||

+ | [https://github.com/Intelligent-Systems-Phystech/Group594/tree/master/Fattakhov2018TextStyleTransfer/code code] | ||

+ | [https://www.youtube.com/watch?v=PM5CmOmlAlw video] | ||

+ | |Rita Kuznetsova | ||

+ | |Danilov Andrey, Murzin Dmitry, [https://rawgit.com/Intelligent-Systems-Phystech/Group594/master/Fattakhov2018TextStyleTransfer/review/Fattakhov2018_Review.pdf review] | ||

+ | |BMF | ||

+ | |AIL-S-B-R-CVTDSWH [AILSBRCVTE0D0WS] | ||

+ | |3+11 | ||

+ | |- | ||

+ | |} | ||

+ | |||

+ | |||

+ | ===1 (1-2) === | ||

+ | * '''Title:''' Classification of superpositions of movements of physical activity | ||

+ | * '''Problem:''' Human behavior analysis by mobile phone sensor measurements: detect human movements from accelerometer data. The accelerometer data is a signal without precise periodicity, which contains an unknown superposition of physical models. We will consider the superposition of models: body + arm/bag/backpack. | ||

+ | Classification of human activities according to measurements of fitness bracelets. According to the measurements of the accelerometer and gyroscope, it is required to determine the type of activity of the worker. It is assumed that the time series of measurements contain elementary movements that form clusters in the space of time series descriptions. (Development: The characteristic duration of movement is seconds. Time series are marked with activity type marks: work, rest. The characteristic duration of activity is minutes. It is required to restore the type of activity by the description of the time series and cluster.) | ||

+ | * '''Data:''' | ||

+ | *# Self assembled | ||

+ | *# Builders data | ||

+ | *# WISDM accelerometer time series ([[Time series (examples library)]], Accelerometry section). | ||

+ | * '''References:''' | ||

+ | *# Karasikov M. E., Strijov V. V. Classification of time series in the space of parameters of generating models // Informatics and its applications, 2016. [[http://strijov.com/papers/Karasikov2016TSC.pdf URL]] | ||

+ | *# Kuznetsov M.P., Ivkin N.P. Algorithm for classification of accelerometer time series by combined feature description // Machine Learning and Data Analysis. 2015. T. 1, No. 11. C. 1471-1483. [[http://jmlda.org/papers/doc/2015/no11/Ivkin2015TSclassification.pdf URL]] | ||

+ | *# Isachenko R. V., Strijov V. V. Metric learning in The problems of multiclass classification of time series // Informatics and its applications, 2016, 10(2): 48-57. [[http://strijov.com/papers/Isachenko2016MetricsLearning.pdf URL]] | ||

+ | *# Zadayanchuk A.I., Popova M.S., Strijov V.V. Choice of the optimal model for classifying physical activity based on accelerometer measurements // Information technologies, 2016. [[http://strijov.com/papers/Zadayanchuk2015OptimalNN4.pdf URL]] | ||

+ | *# Motrenko A.P., Strijov V.V. Extracting fundamental periods to segment human motion time series // Journal of Biomedical and Health Informatics, 2016, Vol. 20, no. 6, 1466-1476. [[http://sourceforge.net/p/mlalgorithms/code/HEAD/tree/Group874/Motrenko2014TSsegmentation/JBHI/MotrenkoStrijov2014RV2.pdf?format=raw URL]] | ||

+ | *# Ignatov A., Strijov V. Human activity recognition using quasiperiodic time series collected from a single triaxial accelerometer // Multimedia Tools and Applications, 2015, 17.05.2015 : 1-14. [[http://strijov.com/papers/Ignatov2015HumanActivity.pdf URL]] | ||

+ | * '''Base algorithm:''' Basic algorithm is described in [Karasikov, Strijov: 2016] and [Kuznetsov, Ivkin: 2014]. | ||

+ | * '''Solution:''' Find the optimal segmentation method and optimal description of the time series. Construct a metric space of descriptions of elementary motions. | ||

+ | * '''Novelty:''' A method for classifying and analyzing complex movements is proposed (Development: Connection of two characteristic times of a description of a person's life, combined problem statement.) | ||

+ | * '''Authors:''' Alexandra Malkova, Maria Vladimirova, R. G. Neichev, Strijov V.V. | ||

+ | |||

+ | ===2 (1) === | ||

+ | * '''Title:''' Comparison of neural network and continuous morphological methods in the Text Detection The problem. | ||

+ | * '''Problem:''' Automatically Detect Text in Natural Images. | ||

+ | * '''Data:''' synthetic generated data + trained photo sample + [https://vision.cornell.edu/se3/coco-text-2/ COCO-Text dataset] + [http://www.machinelearning .ru/wiki/index.php?title=%D0%9A%D0%BE%D0%BD%D0%BA%D1%83%D1%80%D1%81_Avito.ru-2014:_%D1%80% D0%B0%D1%81%D0%BF%D0%BE%D0%B7%D0%BD%D0%B0%D0%B2%D0%B0%D0%BD%D0%B8%D0%B5_%D0% BA%D0%BE%D0%BD%D1%82%D0%B0%D0%BA%D1%82%D0%BD%D0%BE%D0%B9_%D0%B8%D0%BD%D1%84% D0%BE%D1%80%D0%BC%D0%B0%D1%86%D0%B8%D0%B8_%D0%BD%D0%B0_%D0%B8%D0%B7%D0%BE%D0% B1%D1%80%D0%B0%D0%B6%D0%B5%D0%BD%D0%B8%D1%8F%D1%85 Avito Competition 2014]. | ||

+ | * '''References:''' [https://vision.cornell.edu/se3/wp-content/uploads/2016/01/1601.07140v1.pdf COCO benchmark], [https://vision.cornell. edu/se3/wp-content/uploads/2016/01/1601.07140v1.pdf One of a state-of-the-art architecture] | ||

+ | * '''Base algorithm:''' [https://github.com/eragonruan/text-detection-ctpn code] + morphological methods, [http://www.machinelearning.ru/wiki/images/f/f1 /Avito.ru-2014_Ulyanov_presentation.pdf Avito 2014 winner's solution]. | ||

+ | * '''Solution:''' It is proposed to compare the performance of several state-of-the-art algorithms that need a large training set with morphological methods that require a small amount of data. It is proposed to determine the limits of applicability of certain methods. | ||

+ | * '''Novelty:''' propose an algorithm based on the use of both neural network and morphological methods (solution of the word detection problem). | ||

+ | * '''Authors:''' I. N. Zharikov. | ||

+ | * '''Expert''': L. M. Mestetsky (morphological methods). | ||

+ | |||

+ | ===3 (1-2) === | ||

+ | * '''Title:''' Text recognition based on skeletal representation of thick lines and convolutional networks | ||

+ | * '''Problem:''' It is required to build two CNNs, one recognizes a bitmap representation of an image, the other a vector one. (Development: generation of thick lines by neural networks) | ||

+ | * '''Data:''' Bitmap fonts. | ||

+ | * '''References:''' List of works [http://www.machinelearning.ru/wiki/images/a/a2/Morozov2017Synthesis_of_medicines.pdf], in particular arXiv:1611.03199 and | ||

+ | * '''Basic algorithm''': Convolution network for bitmap. | ||

+ | * '''Solution:''' It is required to propose a method for collapsing graph structures, which allows generating an informative description of the skeleton of a thick line. | ||

+ | * '''Novelty:''' A way to improve the quality of recognition of thick lines due to a new way of generating their descriptions is proposed. | ||

+ | * '''Authors:''' L. M. Mestetsky, I. A. Reyer, Strijov V.V. | ||

+ | |||

+ | ===4 (1-2) === | ||

+ | * '''Title:''' Creation of ranking models for information retrieval systems. Algorithm for Predicting the Structure of Locally Optimal Models | ||

+ | * '''Problem:''' It is required to predict a time series using some parametric superposition of algebraic functions. It is proposed not to cost the prognostic model, but to predict it, that is, to predict the structure of the approximating superposition. A class of considered superpositions is introduced, and on the set of such structural descriptions, a search is made for a locally optimal model for the problem under consideration. The problem consists in 1) searching for a suitable structural description of the model 2) describing the search algorithm for the structure that will correspond to the optimal model 3) describing the algorithm for inverse construction of the model according to its structural description. For an already existing example of the answer to questions 1-3, see the works of A. A. Varfolomeeva. | ||

+ | * '''Data:''' | ||

+ | *# Collection of text documents TREC (!) | ||

+ | *# A set of time series, which implies the restoration of functional dependencies. It is proposed to first use synthetic data or immediately apply the algorithm to forecasting time series 1) electricity consumption 2) physical activity with subsequent analysis of the resulting structures. | ||

+ | * '''References:''' | ||

+ | *# (!) Kulunchakov A.S., Strijov V.V. Generation of simple structured Information Retrieval functions by genetic algorithm without stagnation // [http://strijov.com/papers/Kulunchakov2014RankingBySimpleFun.pdf Expert Systems with Applications, 2017, 85: 221–230.] | ||

+ | *# A. A. Varfolomeeva Selection of features when marking up bibliographic lists using structural learning methods, 2013, [http://www.machinelearning.ru/wiki/images/f/f2/Varfolomeeva2013Diploma.pdf?format=raw] | ||

+ | *# Bin Cao, Ying Li and Jianwei Yin Measuring Similarity between Graphs Based on the Levenshtein Distance, 2012, [http://naturalspublishing.com/files/published/92cn7jm44d8wt1.pdf?format=raw] | ||

+ | * '''Base algorithm:''' Specifically, there is no basic algorithm for the proposed problem. It is proposed to try to repeat the experiment of A.A. Varfolomeeva for a different structural description in order to understand what is happening. | ||

+ | * '''Solution:''' The superposition of algebraic functions defines an ortree, on the vertices of which the labels of the corresponding algebraic functions or variables are given. Therefore, the structural description of such a superposition can be its DFS-code. This is a string consisting of vertex labels, written in the order in which the tree is traversed by depth-first search. Knowing the arities of the corresponding algebraic functions, we can restore any such DFS-code in O(n) and get back the superposition of functions. On the set of similar string descriptions, it is proposed to search for the string description that will correspond to the optimal model. | ||

+ | * '''Authors:''' Kulunchakov Andrey, Strijov V.V. | ||

+ | |||

+ | ===5 (1) === | ||

+ | * '''Title:''' Definition of neural network parameters to be optimized. | ||

+ | * '''Problem:''' The problem of neural network optimization is considered. It is required to divide the model parameters into two groups: | ||

+ | *# a) Model parameters to be optimized | ||

+ | *# b) Model parameters whose optimization has been completed. Further optimization of these parameters will not improve the quality of the model. | ||

+ | It is proposed to consider the optimization of parameters as a stochastic process. Based on the history of the process, we find those parameters whose optimization is no longer required. | ||

+ | * '''Data:''' A selection of handwritten MNIST digits | ||

+ | * '''Basic algorithm''': Random choice of parameters. | ||

+ | * '''References:''' | ||

+ | *# [https://arxiv.org/pdf/1704.04289.pdf] SGD as a stochastic process. | ||

+ | *# [http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.704.7138&rep=rep1&type=pdf] Variational inference in neural networks. | ||

+ | * '''Novelty:''' The resulting algorithm will significantly reduce the computational cost of optimizing neural networks. A possible further development of the method is to obtain estimates for the parameters of the network obtained from the original operations of expansion, compression, adding and removing layers. | ||

+ | * '''Authors:''' Oleg Bakhteev, Strijov V.V. | ||

+ | |||

+ | ===6 (1) === | ||

+ | * '''Title:''' Prediction of the graph structure of the neural network model. | ||

+ | * '''Problem:''' The problem is considered to find a stable (and non-redundant in terms of parameters) structure of a convolutional neural network. It is proposed to predict the structure of a neural network using doubly-recurrent neural networks. As a training sample, it is proposed to use the structures of models that have shown good quality on subsamples of small power. | ||

+ | * '''Data:''' Samples MNIST, CIFAR-10 | ||

+ | * '''Basic algorithm''': random search. Comparison with work on reinforcement learning is possible. | ||

+ | * '''References:''' | ||

+ | *# [https://pdfs.semanticscholar.org/e7bd/0e7a7ee6b0904d5de6e76e095a6a3b88dd12.pdf] doubly-recurrent neural networks. | ||

+ | *# [https://arxiv.org/pdf/1707.07012] Similar approach using reinforcement learning. | ||

+ | * '''Authors:''' Oleg Bakhteev, Strijov V.V. | ||

+ | |||

+ | ===7 (1) === | ||

+ | * '''Title:''' Style Change Detection. | ||

+ | * '''Problem:''' Given a collection of documents, it is required to determine if each document is written by one author or by several (http://pan.webis.de/clef18/pan18-web/author-identification.html). | ||

+ | * '''Data:''' PAN 2018 (http://pan.webis.de/clef18/pan18-web/author-identification.html) | ||

+ | PAN 2017 (http://pan.webis.de/clef17/pan17-web/author-identification.html) | ||

+ | PAN 2016 (http://pan.webis.de/clef16/pan16-web/author-identification.html) | ||

+ | * '''References:''' | ||

+ | *# Ian Goodfellow. NIPS 2016 Tutorial: Generative Adversarial Networks (https://arxiv.org/pdf/1701.06547.pdf) | ||

+ | *# Jiwei Li, Will Monroe, Tianlin Shi, Sebastien Jean, Alan Ritter and Dan Jurafsky. Adversarial Learning for Neural Dialogue Generation(https://arxiv.org/pdf/1701.06547.pdf) | ||

+ | *# M. Kuznetsov, A. Motrenko, R. Kuznetsova, V. Strijov. Methods for Intrinsic Plagiarism Detection and Author Diarization | ||

+ | *# K. Safin, R. Kuznetsova. Style Breach Detection with Neural Sentence Embeddings (https://pdfs.semanticscholar.org/c70e/7f8fbc561520accda7eea2f9bbf254edb255.pdf) | ||

+ | * '''Basic algorithm''': solution described in [3, 4]. | ||

+ | * '''Solution:''' is proposed to solve the problem using generative adversarial networks — the generative model generates texts in the same author's style, the discriminative model — a binary classifier. | ||

+ | * '''Novelty:''' it is assumed that the solution of this problem by the proposed method can give an increase in quality compared to typical methods for solving this problem, as well as related clustering problems of the authors. | ||

+ | * '''Authors:''' Rita Kuznetsova (consultant), Strijov V.V. | ||

+ | |||

+ | ===8 (1) === | ||

+ | * '''Title:''' Obtaining likelihood estimates using autoencoders | ||

+ | * '''Problem:''' it is assumed that the objects under consideration obey the manifold hypothesis (manifold learning) - high-dimensional vectors are concentrated around some subspace of lower dimension. Works [1, 2] show that some modifications of autoencoders are looking for a k-dimensional manifold in the object space, which most fully conveys the data structure. In [2], an estimate of the probability density of data is derived using an autoencoder. It is required to obtain this estimate for the plausibility of the model. | ||

+ | * '''Data:''' it is proposed to experiment on short text fragments of Google ngrams (http://storage.googleapis.com/books/ngrams/books/datasetsv2.html) | ||

+ | * '''References:''' | ||

+ | *# Pascal Vincent, Hugo Larochelle, Isabelle Lajoie, Yoshua Bengio, Pierre-Antoine Manzagol. Stacked Denoising Autoencoders: Learning Useful Representations in a Deep Network with a Local Denoising Criterion (http://www.jmlr.org/papers/volume11/vincent10a/vincent10a.pdf). | ||

+ | *# Guillaume Alain, Yoshua Bengio. What Regularized Auto-Encoders Learn from the Data Generating Distribution (https://arxiv.org/pdf/1211.4246.pdf) | ||

+ | *# Hanna Kamyshanska, Roland Memisevic. The Potential Energy of an Autoencoder (https://www.iro.umontreal.ca/~memisevr/pubs/AEenergy.pdf) | ||

+ | * '''Basic algorithm''': | ||

+ | * '''Solution:''' in the problem it is proposed to train vector representations for phrases (n-grams) using an autoencoder, using Theorem 2 in [2] to obtain an estimate for the likelihood of the sample and, using this estimate, derive the likelihood of the model . Using the estimates obtained, one can also consider the sampling process. | ||

+ | * '''Novelty:''' obtaining data and model likelihood estimates, generating texts using the resulting estimates. | ||

+ | * '''Authors:''' Rita Kuznetsova (consultant). | ||

+ | |||

+ | ===9 (1) === | ||

+ | * '''Title:''' Predict properties and types of atoms in molecular graphs using convolutional networks. | ||

+ | * '''Problem:''' Multilabel classification using convolutional neural networks (CNN) on graphs. | ||

+ | To predict the interaction of molecules with each other, it is often necessary to correctly describe their constituent atoms by assigning certain types to them. For small molecules, not many descriptors are available: the coordinates and chemical elements of atoms, the lengths of bonds and the magnitude of the angles between them. Using these features, we successfully predict atomic hybridizations and bond types. In this approach, each atom is considered "individually", the information about neighboring atoms necessary to determine the type of an atom is practically not used, and the types of atoms are determined by checking a large number of conditions. At the same time, molecules are represented as 3D molecular graphs, and it would be interesting to use this to predict their types with machine learning methods, for example, using CNNs. | ||

+ | It is necessary to predict the types of vertices and edges of molecular graphs: | ||

+ | *# atom type (graph vertex type, about 150 classes), | ||

+ | *# atom hybridization (auxiliary feature, vertex type, 4 classes), | ||

+ | *# connection type (auxiliary feature, edge type, 5 classes). | ||

+ | |||

+ | The type of an atom (graph vertex) is based on information about its hybridization and the properties of neighboring atoms. Therefore, in the case of a successful solution of the classification problem, clustering can be carried out to find other ways to determine the types of atoms. | ||

+ | |||

+ | * '''Data:''' About 15 thousand molecules represented as molecular graphs. For each vertex (atom), 3D coordinates and a chemical element are known. Additionally, bond lengths, angles and dihedral angles between atoms (3D graph coordinates), binary signs reflecting whether an atom is included in the cycle and whether it is terminal are calculated. The sample is labeled, but the labeled data may contain ~5% errors. | ||

+ | If there is not enough data, it is possible to increase the sample (up to 200 thousand molecules), associated with an increase in inaccuracies in labeling. | ||

+ | |||

+ | * '''References:''' | ||

+ | *# [http://proceedings.mlr.press/v48/niepert16.pdf] | ||

+ | *# [https://arxiv.org/pdf/1603.00856.pdf] | ||

+ | *# [https://arxiv.org/pdf/1204.4539.pdf] | ||

+ | * '''Base algorithm:''' Prediction of hybridizations and link orders using a multiclass non-linear SVM with a small number of descriptors. https://hal.inria.fr/hal-01381010/document | ||

+ | * '''Solution:''' Proposed solution to the problem and ways of conducting research. | ||

+ | Methods for presenting and visualizing data and conducting error analysis, analyzing the quality of the algorithm. | ||

+ | At the first stage, it will be necessary to determine the operations on the graphs necessary to build the network architecture. Next, you will need to train the network for multi-class classification of the types of vertices (and edges) of the input graph. | ||

+ | To assess the quality of the algorithm, it is supposed to evaluate the accuracy using cross-validation. For the final publication (in a specialized journal), it will be necessary to make a specific test for the quality of predictions: based on the predicted bond types, the molecule is written as a string (in SMILES format) and compared with a sample. In this case, for each molecule, the prediction will be considered correct only if the types of all bonds in it were predicted without errors. | ||

+ | * '''Novelty:''' The proposed molecular graphs have a 3D structure and internal hierarchy, making them an ideal CNN application. | ||

+ | * '''Authors:''' Sergei Grudinin, Maria Kadukova, Strijov V.V. | ||

+ | |||

+ | ===10 (1) === | ||

+ | * '''Title:''' Formulation and solution of an optimization problem combining classification and regression to estimate the binding energy of a protein and small molecules. The problem description [https://www.overleaf.com/read/rjdnyyxpdkyj] | ||

+ | * '''Problem:''' | ||

+ | From the point of view of bioinformatics, The problem is to estimate the free energy of protein binding to a small molecule (ligand): the best ligand in its best position has the \textbf{lowest free energy} of interaction with the protein. (Following a large text, see the file at the link above.) | ||

+ | * '''Data:''' | ||

+ | *# Data for binary classification. | ||

+ | Approximately 12,000 protein-ligand complexes: for each of them there is 1 native position and 18 non-native ones. The main descriptors are histograms of distributions of distances between different atoms of the protein and ligand, the dimension of the vector of descriptors is ~ 20,000. In the case of continued research and publication in a specialized journal, the set of descriptors can be expanded. | ||

+ | The data will be provided as binary files with a python script to read. | ||

+ | *# Data for regression. | ||

+ | For each of the presented complexes, the value of the quantity is known, which can be interpreted as the binding energy. | ||

+ | * '''References:''' | ||

+ | *# SVM [http://cs229.stanford.edu/notes/cs229-notes3.pdf] | ||

+ | *# Ridge Regression [http://scikit-learn.org/stable/modules/linear_model.html#ridge-regression] | ||

+ | *# [https://alex.smola.org/papers/2003/SmoSch03b.pdf] (section 1) | ||

+ | * '''Base algorithm:''' [https://hal.inria.fr/hal-01591154/] | ||

+ | In the classification problem, we used an algorithm similar to linear SVM, whose relationship with the energy estimate, which is outside the scope of the classification problem, is described in the above article. Various loss functions can be used in a regression problem. | ||

+ | * '''Solution:''' It is necessary to connect the previously used optimization problem with the regression problem and solve it using standard methods. Cross-validation will be used to check the operation of the algorithm. | ||

+ | There is a separate test set consisting of (1) 195 complexes of proteins and ligands, for which it is necessary to find the best ligand pose (the algorithm for obtaining ligand positions differs from that used in training), (2) complexes of proteins and ligands, for which native poses it is necessary to predict the energy binding, and (3) 65 proteins for which the most strongly binding ligand is to be found. | ||

+ | * '''Novelty:''' First of all, the interest is ''combining classification and regression problems'''. | ||

+ | The correct assessment of the quality of protein and ligand binding is used in drug development to search for molecules that interact most strongly with the protein under study. Using the classification problem described above to predict the binding energy results in an insufficiently high correlation of predictions with experimental values, while using the regression problem alone leads to overfitting. | ||

+ | * '''Authors''' Sergei Grudinin, Maria Kadukova, Strijov V.V. | ||

+ | |||

+ | ==2017== | ||

+ | {|class="wikitable" | ||

+ | |- | ||

+ | ! Author | ||

+ | ! Topic | ||

+ | ! Link | ||

+ | ! Consultant | ||

+ | ! Reviewer | ||

+ | ! Report | ||

+ | ! Letters | ||

+ | |- | ||

+ | |Goncharov Alexey (example) | ||

+ | |Metric classification of time series | ||

+ | |[http://svn.code.sf.net/p/mlalgorithms/code/Group274/Goncharov2015MetricClassification/code code], | ||

+ | [http://svn.code.sf.net/p/mlalgorithms/code/Group274/Goncharov2015MetricClassification/doc/Goncharov2015MetricClassification.pdf paper], | ||

+ | [http://svn.code.sf.net/p/mlalgorithms/code/Group274/Goncharov2015MetricClassification/doc/GoncharovAlexey2015PresentationMetricClassification.pdf slides] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Mpopova Maria Popova] | ||

+ | |Zadayanchuk Andrey | ||

+ | |BMF | ||

+ | |AILSBRCVTDSWH> | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Alvant Alekseev Vasily] | ||

+ | |Intratext coherence as a measure of interpretability of thematic models of text collections | ||

+ | |[http://svn.code.sf.net/p/mlalgorithms/code/Group474/Alekseev2017IntraTextCoherence/code code] | ||

+ | [http://svn.code.sf.net/p/mlalgorithms/code/Group474/Alekseev2017IntraTextCoherence/data/postnauka_original_reduced/postnauka_clean data] | ||

+ | [http://svn.code.sf.net/p/mlalgorithms/code/Group474/Alekseev2017IntraTextCoherence/doc/Alekseev2017IntraTextCoherence.pdf paper] | ||

+ | [http://svn.code.sf.net/p/mlalgorithms/code/Group474/Alekseev2017IntraTextCoherence/doc/Alekseev2017Presentation.pdf slides] | ||

+ | [https://www.youtube.com/watch?v=6v2dNMJG4iA video] | ||

+ | |Viktor Bulatov | ||

+ | |Zakharenkov Anton | ||

+ | |BMF | ||

+ | |AILSB+RC+V+TDHW | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Dmitriy_Anikeyev Anikeev Dmitry] | ||

+ | |Local approximation of time series for building predictive metamodels | ||

+ | |[https://sourceforge.net/p/mlalgorithms/code/HEAD/tree/Group474/Anikeyev_Penkin2017ClassifyingMetamodels/code/ code] | ||

+ | [https://sourceforge.net/p/mlalgorithms/code/HEAD/tree/Group474/Anikeyev_Penkin2017ClassifyingMetamodels/paper/AnikeyevPenkin2017Splines.pdf paper] | ||

+ | [https://sourceforge.net/p/mlalgorithms/code/HEAD/tree/Group474/Anikeyev_Penkin2017ClassifyingMetamodels/paper/Anikeev%20F-talk.pdf slides] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:strijov Strijov V.V.] | ||

+ | |[https://svn.code.sf.net/p/mlalgorithms/code/Group474/Anikeyev2017ClassifyingMetamodels/paper/Review.pdf Smerdov Anton] | ||

+ | |BMF | ||

+ | |AILS>B0R0C0V0T0D0H0W0 | ||

+ | |- | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Gasanov_Elnur Gasanov Elnur] | ||

+ | |Construction of an approximating description of a scalogram in the problem of predicting movements using an electrocorticogram | ||

+ | |[https://svn.code.sf.net/p/mlalgorithms/code/Group474/Gasanov2017ECoGAnalysis/Code code] [https://svn.code.sf.net/p/mlalgorithms/code/Group474/Gasanov2017ECoGAnalysis/Paper/Gasanov2017ECoGAnalysis.pdf paper] | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group474/Gasanov2017ECoGAnalysis/Paper/FTalk.pdf slides] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Anastasiya Anastasia Motrenko] | ||

+ | |Kovalev Dmitry | ||

+ | |BMF | ||

+ | |AILSBRCVTDH0W0 | ||

+ | |- | ||

+ | |Zakharenkov Anton | ||

+ | |Massively multiThe problem deep learning for drug discovery | ||

+ | |[https://svn.code.sf.net/p/mlalgorithms/code/Group474/Zakharenkov2017MassivelyMultiThe problemNetworks/code/ code] | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group474/Zakharenkov2017MassivelyMultiThe problemNetworks/doc/Zakharenkov2017MassivelyMultiThe problemNetworks.pdf paper] | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group474/Zakharenkov2017MassivelyMultiThe problemNetworks/doc/Zakharenkov2016Presentation.pdf slides] | ||

+ | [https://youtu.be/l6M-CfpkZKQ video] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Mpopova Maria Popova] | ||

+ | |Alekseev Vasily | ||

+ | |BMF | ||

+ | |AILSBRCVT>D>H0W0 | ||

+ | |- | ||

+ | |Kovalev Dmitry | ||

+ | |Unsupervised representation for molecules | ||

+ | |[https://svn.code.sf.net/p/mlalgorithms/code/Group474/Kovalev2017MoleculesRepresentation/code/ code] | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group474/Kovalev2017MoleculesRepresentation/doc/paper/Kovalev2017MoleculesRepresentation.pdf paper] | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group474/Kovalev2017MoleculesRepresentation/doc/slides/Kovalev2017MoleculesRepresentation.pdf slides] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Mpopova Maria Popova] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Gasanov_Elnur Gasanov Elnur] | ||

+ | |BMF | ||

+ | |AILSBRCVT>D>H0W0 | ||

+ | |- | ||

+ | |Novitsky Vasily | ||

+ | |Feature Selection in Problems of Autoregressive Prediction of Biomedical Signals | ||

+ | |[https://svn.code.sf.net/p/mlalgorithms/code/Group474/Novitskiy2017Biosignal/doc/novitskiy.pdf paper] | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group474/Novitskiy2017Biosignal/code/ code] | ||

+ | [http://svn.code.sf.net/p/mlalgorithms/code/Group474/Novitskiy2017Biosignal/slides/presentation.pdf slides] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Katrutsa Alexander Katrutsa] | ||

| | | | ||

+ | |B - F | ||

+ | |AILS>B0R0C0V0T0D0H0W0 | ||

+ | |- | ||

+ | |Selezneva Maria | ||

+ | |Aggregation of heterogeneous text collections in a hierarchical thematic model of Russian-language popular science content | ||

+ | |[http://svn.code.sf.net/p/mlalgorithms/code/Group474/Seleznova2017AggregationARTM/paper/Seleznova2017AggregationARTM.pdf paper] | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group474/Seleznova2017AggregationARTM/code/ code] | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group474/Seleznova2017AggregationARTM/slides/FinalTalk.pdf slides] | ||

+ | [https://www.youtube.com/watch?v=eKUJtfGGlTY video] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Iefimova Irina Efimova] | ||

+ | |[https://sourceforge.net/p/mlalgorithms/code/HEAD/tree/Group474/Seleznova2017AggregationARTM/feedback/Selezniova2017_Sholokhov-Feedback.rtf Sholokhov Alexey] | ||

+ | |BMF | ||

+ | |A+IL+SBRCVTDHW | ||

+ | |- | ||

+ | |Smerdov Anton | ||

+ | |Choosing the optimal recurrent network model in the Paraphrase Search The problems | ||

+ | |[http://svn.code.sf.net/p/mlalgorithms/code/Group474/Smerdov2017Paraphrase/doc/Smerdov2017Paraphrase.pdf paper] | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group474/Smerdov2017Paraphrase/code/ code] | ||

+ | [http://svn.code.sf.net/p/mlalgorithms/code/Group474/Smerdov2017Paraphrase/doc/Smerdov2017ParaphrasePresentation.pdf slides] | ||

+ | [https://www.youtube.com/watch?v=dW_xv2IlhC4 video] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Oleg_Bakhteev Oleg Bakhteev] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Dmitriy_Anikeyev Dmitry Anikeev] | ||

+ | |BMF | ||

+ | |AIL+SB+RC>V+M-T>D0H0W0 | ||

+ | |- | ||

+ | |Uvarov Nikita | ||

+ | |Optimal Algorithm for Reconstruction of Dynamic Models | ||

+ | |[http://svn.code.sf.net/p/mlalgorithms/code/Group474/Uvarov2017DynamicGraphicalModels/doc/Uvarov2017DynamicGraphicalModels.pdf paper] | ||

+ | [http://svn.code.sf.net/p/mlalgorithms/code/Group474/Uvarov2017DynamicGraphicalModels/slides/Uvarov2017DynamicGraphicalModels.pdf slides] | ||

+ | [http://svn.code.sf.net/p/mlalgorithms/code/Group474/Uvarov2017DynamicGraphicalModels/code/ code] | ||

+ | [https://www.youtube.com/watch?v=79t61GB40nU video] | ||

+ | |Yuri Maksimov | ||

| | | | ||

+ | |BMF | ||

+ | |AILS0B0R0C0V0T0D0H0W0 | ||

|- | |- | ||

- | | | + | |Usmanova Karina |

+ | |Multiple Manifold Learning (Joint diagonalization for 3D shapes - AJD on Hessian matrices) | ||

+ | |[http://svn.code.sf.net/p/mlalgorithms/code/Group474/Usmanova2017MultipleManifoldLearning/doc/Usmanova2017MultipleManifoldLearning.pdf paper] | ||

+ | [http://svn.code.sf.net/p/mlalgorithms/code/Group474/Usmanova2017MultipleManifoldLearning/slides/Usmanova2017PresentationAJD.pdf slides] | ||

+ | [http://svn.code.sf.net/p/mlalgorithms/code/Group474/Usmanova2017MultipleManifoldLearning/code/ code] | ||

+ | [https://www.youtube.com/watch?v=sqHLmSU-2iM video] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Mkarasikov Mikhail Karasikov] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:IShibaev Innokenty Shibaev] | ||

+ | |BMF | ||

+ | |AILSBRC+VT+EDH>W | ||

+ | |- | ||

+ | |Innokenty Shibaev | ||

+ | |Convex relaxations for multiple structure alignment (synchronization problem for SO(3)) | ||

+ | |[http://svn.code.sf.net/p/mlalgorithms/code/Group474/Shibaev2017MultipleStructureAlignment/doc/Shibaev2017MultipleStructureAlignment.pdf paper] | ||

+ | [http://svn.code.sf.net/p/mlalgorithms/code/Group474/Shibaev2017MultipleStructureAlignment/doc/Shibaev2017MultipleStructureAlignment_Final.pdf slides] | ||

+ | [https://nbviewer.jupyter.org/urls/svn.code.sf.net/p/mlalgorithms/code/Group474/Shibaev2017MultipleStructureAlignment/code/Shibaev2017MultipleStructureAlignment_different_algs.ipynb code] | ||

+ | [https://youtu.be/qs1Rchb02C0 video] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Mkarasikov Mikhail Karasikov] | ||

+ | |Usmanova Karina | ||

+ | |BMF | ||

+ | |AILS-BRCVT>D>H>W | ||

+ | |- | ||

+ | |Sholokhov Alexey | ||

+ | |Noise immunity of methods for informational analysis of ECG signals | ||

| | | | ||

+ | [https://sourceforge.net/p/mlalgorithms/code/HEAD/tree/Group474/Sholokhov2017NoiseSustainability/doc/Sholokhov2017NoiseSustainability.pdf paper] | ||

+ | [https://sourceforge.net/p/mlalgorithms/code/HEAD/tree/Group474/Sholokhov2017NoiseSustainability/code/stage2_statistics_calculation.ipynb code] | ||

+ | [https://sourceforge.net/p/mlalgorithms/code/HEAD/tree/Group474/Sholokhov2017NoiseSustainability/slides/Sholokhov2017NiseSustainability_MidTalk.pdf slides] | ||

+ | [https://www.youtube.com/watch?v=5BHIpUiY9VU video] | ||

+ | |Vlada Bunakova | ||

+ | |[https://sourceforge.net/p/mlalgorithms/code/HEAD/tree/Group474/Sholokhov2017NoiseSustainability/feedback/Sholokhov2017NoiseSustainability_SelezniovaFeedback.rtf Selezneva Maria] | ||

+ | |BMF | ||

+ | |AILSBRCVTDHW | ||

+ | |- | ||

+ | |} | ||

+ | |||

+ | |||

+ | Risky works | ||

+ | {|class="wikitable" | ||

+ | |- | ||

+ | ! Author | ||

+ | ! Topic | ||

+ | ! Link | ||

+ | ! Consultant | ||

+ | ! Reviewer | ||

+ | ! Report | ||

+ | ! Letters | ||

+ | |- | ||

+ | |Kaloshin Pavel | ||

+ | |Using deep learning networks to transfer classification models in case of insufficient data. | ||

| | | | ||

+ | [https://sourceforge.net/p/mlalgorithms/code/HEAD/tree/Group474/KaloshinBolotin2017TransferLearning/paper/main.pdf paper] | ||

+ | [https://sourceforge.net/p/mlalgorithms/code/HEAD/tree/Group474/KaloshinBolotin2017TransferLearning/code code] | ||

+ | [https://sourceforge.net/p/mlalgorithms/code/HEAD/tree/Group474/KaloshinBolotin2017TransferLearning/data data] | ||

+ | |[[Участник:khritankov|Anton Khritankov]] | ||

| | | | ||

+ | | - MF | ||

+ | |AIL-SBRC-VT+D>H>W0 | ||

|- | |- | ||

- | | | + | |Malinovsky Grigory |

+ | |Choice of Interpreted Multimodels in Credit Scoring The problems | ||

+ | |[https://svn.code.sf.net/p/mlalgorithms/code/Group474/Malinovskiy2017CreditScoring/doc/paper.pdf paper] | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group474/Malinovskiy2017CreditScoring/code/ code] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Aduenko Alexander Aduenko] | ||

+ | | | ||

+ | |out B - - | ||

+ | |AILS-B>R>C>V>T0D0H0W0 | ||

+ | |- | ||

+ | |Pletnev Nikita | ||

+ | |Internal plagiarism detection | ||

+ | |[https://svn.code.sf.net/p/mlalgorithms/code/Group474/Pletnev2017PlagiarismDetecting/Pletnev2017PlagiarismDetecting.pdf paper] | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Rita_Kuznetsova Rita Kuznetsova] | ||

+ | | | ||

+ | |out - - - | ||

+ | |A-I-L-S>B0R0C0V0T0D0H0W0 | ||

+ | |- | ||

+ | |Grevtsev Alexander | ||

+ | |Parallel Algorithms for Parametric Identification of the Tersoff Potential for AlN | ||

+ | | | ||

+ | [https://sourceforge.net/p/mlalgorithms/code/HEAD/tree/Group474/Grevtsev2017Problem3/doc/Article.pdf paper] | ||

+ | |Karine Abgaryan | ||

| | | | ||

| | | | ||

| | | | ||

|- | |- | ||

- | | | + | |Zaitsev Nikita |

+ | |Automatic classification of scientific articles on crystallography | ||

+ | | | ||

+ | [https://sourceforge.net/p/mlalgorithms/code/HEAD/tree/Group474/Zaytsev2017ArticlesClassification/report/report.pdf paper] | ||

+ | [https://sourceforge.net/p/mlalgorithms/code/HEAD/tree/Group474/Zaytsev2017ArticlesClassification/README.txt readme] | ||

+ | |Evgeny Gavrilov | ||

| | | | ||

| | | | ||

| | | | ||

|- | |- | ||

- | | | + | |Diligul Alexander |

+ | |Determination of the optimal potential parameters for the Rosato-Guillope-Legrand (RGL) model from experimental data and the results of quantum mechanical calculations | ||

+ | | | ||

+ | [https://sourceforge.net/p/mlalgorithms/code/HEAD/tree/Group474/Diligul2017Problem4/Doc/Article.pdf paper] | ||

+ | |Karine Abgaryan | ||

| | | | ||

| | | | ||

| | | | ||

|- | |- | ||

- | | | + | |Daria Fokina |

+ | |Selection of Candidates in the Problem of Finding Text Borrowings with Paraphrasing Based on the Vectorization of Text Fragments | ||

| | | | ||

+ | |[http://www.machinelearning.ru/wiki/index.php?title=Участник:Fess10 Alexey Romanov] | ||

| | | | ||

| | | | ||

+ | |AILSB0R0C0V0T0D0H0W0 | ||

|- | |- | ||

- | | | + | |} |

+ | |||

+ | ===1. 2017=== | ||

+ | * '''Title:''' Classification of human activities according to fitness bracelet measurements. | ||

+ | * '''Problem:''' According to the accelerometer and gyroscope measurements, it is required to determine the type of worker's activity. It is assumed that the time series of measurements contain elementary movements that form clusters in the space of time series descriptions. The characteristic duration of the movement is seconds. Time series are labeled with activity type labels: work, leisure. The typical duration of activity is minutes. It is required to restore the type of activity according to the description of the time series and cluster. | ||

+ | * '''Data:''' WISDM accelerometer time series ([[Time series (examples library)]], Accelerometry section). | ||

+ | * '''References:''' | ||

+ | *# Karasikov M.E., Strijov V.V. Classification of time series in the space of parameters of generating models // Informatics and its applications, 2016. [[http://strijov.com/papers/Karasikov2016TSC.pdf URL]] | ||

+ | *# Kuznetsov M.P., Ivkin N.P. Algorithm for Classifying Accelerometer Time Series by Combined Feature Description // Machine Learning and Data Analysis. 2015. V. 1, No. 11. C. 1471 - 1483. [[http://jmlda.org/papers/doc/2015/no11/Ivkin2015TSclassification.pdf URL]] | ||

+ | *# Isachenko R.V., Strijov V.V. Metric learning in The problemx multiclass classification of time series // Informatics and its applications, 2016, 10(2) : 48-57. [[http://strijov.com/papers/Isachenko2016MetricsLearning.pdf URL]] | ||

+ | *# Zadayanchuk A.I., Popova M.S., Strijov V.V. Choosing the optimal model for classifying physical activity based on accelerometer measurements // Information technologies, 2016. [[http://strijov.com/papers/Zadayanchuk2015OptimalNN4.pdf URL]] | ||

+ | *# Motrenko A.P., Strijov V.V. Extracting fundamental periods to segment human motion time series // Journal of Biomedical and Health Informatics, 2016, Vol. 20, no. 6, 1466 - 1476. | ||

+ | *# Ignatov A., Strijov V. Human activity recognition using quasiperiodic time series collected from a single triaxial accelerometer // Multimedia Tools and Applications, 2015, 17.05.2015 : 1-14. [[http://strijov.com/papers/Ignatov2015HumanActivity.pdf URL]] | ||

+ | * '''Base algorithm:''' Basic algorithm is described in [Karasikov, Strijov: 2016] and [Kuznetsov, Ivkin: 2014]. | ||

+ | * '''Solution:''' Find the optimal segmentation method and optimal description of the time series. Construct a metric space of descriptions of elementary motions. | ||

+ | * '''Novelty:''': Connection of two characteristic times of the description of a person's life, combined statement of the problem. | ||

+ | * '''Authors:''' Strijov V.V., M.P. Kuznetsov, P.V. Levdik. | ||

+ | |||

+ | ===2. 2017=== | ||

+ | * '''Title:''' Construction of an approximating description of a scalogram in the problem of predicting movements using an electrocorticogram. | ||

+ | * '''Problem:''' As part of solving the problem of decoding ECoG signals, The problem of classifying movements by time series of electrode readings is solved. The tools for extracting features from ECoG time series are the coefficients of the wavelet transform of the signal under study [Makarchuk 2016], on the basis of which a scalogram is built for each electrode - a two-dimensional array of features in frequency-time space. Combining scalograms for each electrode gives signs of a time series in the spatio-frequency-time domain. The feature description constructed in this way obviously contains multicorrelated features and is redundant. It is required to propose a method for reducing the dimension of the feature space. | ||

+ | * '''Data:''' Measurements of the positions of the fingers when performing simple gestures. [https://purl.stanford.edu/zk881ps0522 Description of experiments] [https://stacks.stanford.edu/file/druid:zk881ps0522/gestures.zip data]. | ||

+ | * '''References:''' | ||

+ | *# Makarchuk G.I., Zadayanchuk A.I. Strijov V.V. 2016. Using partial least squares to decode hand movement using ECoG cues in monkeys. [http://svn.code.sf.net/p/mlalgorithms/code/Group374/Makarchuk2016ECoGSignals/doc/Makarchuk2016ECoGSignals.pdf pdf] | ||

+ | *# Karasikov M.E., Strijov V.V. Classification of time series in the space of parameters of generating models // Informatics and its applications, 2016. [[http://strijov.com/papers/Karasikov2016TSC.pdf URL]] | ||

+ | *# Kuznetsov M.P., Ivkin N.P. Algorithm for Classifying Accelerometer Time Series by Combined Feature Description // Machine Learning and Data Analysis. 2015. T. 1, No. 11. C. 1471 - 1483. | ||

+ | * '''Base algorithm:''' PLS | ||

+ | Chen C, Shin D, Watanabe H, Nakanishi Y, Kambara H, et al. (2013) Prediction of Hand Trajectory from Electrocorticography Signals in Primary Motor Cortex. PLoS ONE 8(12): e83534. | ||

+ | * '''Solution:''' To reduce the dimension, it is proposed to use the local approximation method proposed in [Kuznetsov 2015] used to classify accelerometric time series [Karasikov 2016]. | ||

+ | * '''Novelty:''' A new method of movement recovery based on electrocorticograms is proposed. | ||

+ | * '''Authors:''' Strijov V.V., A.P. Motrenko | ||

+ | |||

+ | ===3. 2017=== | ||

+ | * '''Title:''' Multiple Manifold Learning (Joint diagonalization for 3D shapes - AJD on Hessian matrices). | ||

+ | * '''Problem:''' Building an optimal algorithm for the Multiple Manifold Learning The problem. Two protein conformations (two tertiary structures) are given. In the vicinity of each state, a model of an elastic body is specified (oscillations of the structure in the vicinity of these states). The problem is to build a general model of an elastic body to find intermediate states with the maximum match with these models in the vicinity of given conformations. The space of motion of an elastic body is given by the Hessian eigenvectors. It is required to find a common low-rank approximation of the space of motions of two elastic bodies. | ||

+ | * '''Data:''' Protein structures in double conformations from PDB, about 100 sets from the article https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4677049/ | ||

+ | * '''References:''' A list of scientific papers, supplemented by 1) the statement of the problem being solved, 2) links to new results (a recent article that is close in results), 3) basic information about the problem under study. | ||

+ | Tirion, M. M. (1996). Large amplitude elastic motions in proteins from a single-parameter, atomic analysis. Physical Review Letters, 77(9), 1905. | ||

+ | Moal, I. H., & Bates, P. A. (2010). {SwarmDock} and the Use of Normal Modes in Protein-Protein Docking. IJMS, 11(10), 3623–3648. https://doi.org/10.3390/ijms11103623 | ||

+ | * '''Base algorithm:''' AJD algorithm: http://perso.telecom-paristech.fr/~cardoso/jointdiag.html, AJD algorithms implemented as part of Shogun ML toolbox http://shogun-toolbox.org , http://shogun-toolbox.org/api/latest/classshogun_1_1CApproxJointDiagonalizer.html. | ||

+ | * '''Solution:''' Computing Hessians (C++ code from Sergey), learning and running standard joint diagonalization algorithms for the first n non-trivial eigenvectors, analyzing loss functions, adapting the standard algorithm to solve the original problem. | ||

+ | * '''Novelty:''' Using simple elasticity models with one or more free parameters, thermal fluctuations in proteins can be described. However, such models do not describe transitions between several stable conformations in proteins. The purpose of this work is to refine the elastic model so that it also describes the space of conformational changes. | ||

+ | * '''Authors:''' Sergey Grudinin, consultant: Mikhail Karasikov / Yury Maksimov. | ||

+ | |||

+ | ===4. 2017=== | ||

+ | * '''Title:''' Convex relaxations for multiple structure alignment (synchronization problem for SO(3)). | ||

+ | * '''Problem:''' Find transformations to align protein tertiary structures simultaneously (in simple words: find orthogonal transformations that align data in R^3 molecules that have the same chemical formula). If the structures are the same (the RMSD is equal to zero after alignment, the structures are aligned exactly), then you can align in pairs. However, if this is not the case, then the Basic algorithm, generally speaking, does not find the optimum of the original problem with a loss function for simultaneous equalization. | ||

+ | * '''Data:''' Protein structures in PDB format in various states and coordinate systems. | ||

+ | * '''References:''' | ||

+ | *# Multiple structural alignment: | ||

+ | *# Kearsley.S.K. (1990)7. Comput. Chem., 11, 1187-1192. | ||

+ | *# Shapiro., BothaJ.D., PastorA and Lesk.A.M. (1992) Acta Crystallogr., A48, 11-14. | ||

+ | *# Diamond,R. (1992) Protein Sci., 1, 1279-1287. | ||

+ | *# May AC, Johnson MS, Improved genetic algorithm-based protein structure comparisons: pairwise and multiple superpositions. ProteinEng. 1995 Sep;8(9):873-82. | ||

+ | *# Synchronization problem: | ||

+ | *# O. Özyeşil, N. Sharon, A. Singer, ``Synchronization over Cartan motion groups via contraction”, Available at arXiv. | ||

+ | *# L. Wang, A. Singer, `ʻExact and Stable Recovery of Rotations for Robust Synchronization”, Information and Inference: A Journal of the IMA, 2(2), pp. 145--193 (2013). | ||

+ | *# Semidefinite relaxations for optimization problems over rotation matrices J Saunderson, PA Parrilo… - Decision and Control ( …, 2014 - ieeexplore.ieee.org | ||

+ | *# Spectral synchronization of multiple views in SE (3) F Arrigoni, B Rossi, A Fusiello - SIAM Journal on Imaging Sciences, 2016 - SIAM | ||

+ | *# Robust Rotation Synchronization via Low-rank and Sparse Matrix Decomposition, F Arrigoni, A Fusiello, B Rossi, P Fragneto - arXiv preprint arXiv: …, 2015 - arxiv.org | ||

+ | *# Spectral relaxation for SO(2) | ||

+ | *# A. Singer, Angular synchronization by eigenvectors and semidefinite programming, Applied and Computational Harmonic Analysis 30 (1) (2011) 20 – 36. | ||

+ | *# Spectral relaxation for SO(3) | ||

+ | *# M.Arie-Nachimson,S.Z.Kovalsky,I.Kemelmacher-Shlizerman,A.Singer,R.Basri,Global motion estimation from point matches, in: International Conference on 3D Imaging, Modeling, Processing, Visualization and Transmission, 2012 , pp. 81–88. | ||

+ | *# A. Singer, Y. Shkolnisky, Three-dimensional structure determination from common lines in cryo-em by eigenvectors and semidefinite programming, SIAM Journal on Imaging Sciences 4 (2) (2011) 543–572. | ||

+ | * '''Base algorithm:''' Local (pairwise) alignment algorithm. Kearsley S.K. (1989) Acta Crystallogr., A45, 208-210; Rapid determination of RMSDs corresponding to macromolecular rigid body motions | ||

+ | Petr Popov, Sergei Grudinin, Journal of Computational Chemistry, Wiley, 2014, 35(12), pp.950-956. <10.1002/jcc.23569> | ||

+ | DOI: 10.1002/jcc.23569 | ||

+ | * '''Solution:''' Two options for setting optimization problems (through rotation matrices and through quaternions). Relaxation of the obtained problems by convex ones, comparison of the solutions of the problem by the basic algorithm and relaxations (spectral relaxation, SDP). | ||

+ | * '''Novelty:''' A method that flattens structures by minimizing the loss function, taking into account all pairwise losses. | ||

+ | * '''Authors:''' Sergey Grudinin, consultant: Mikhail Karasikov. | ||

+ | |||

+ | ===5. 2017=== | ||

+ | * '''Title:''' Local approximation of time series for building predictive metamodels. | ||

+ | * '''Problem:''' The physical activity of a person is investigated by time series - accelerometer measurements. The aim of the project is to create a tool for analyzing the problem of creating models for predicting models - metamodels. The segment of the time series is investigated. It is required to predict the class of the segment. (Option: predict the end of the segment, the next segment, its class. In this case, the class of the next segment may differ from the class of the previous one). | ||

+ | * '''Data:''' Based on a Santa Fe or WISDM sample (samples consist of segments with many elementary movements and class labels corresponding to the segments), a variant of the OPPORTUNITY Activity Recognition Challenge. | ||

+ | * '''References:''' | ||

+ | *# Karasikov M.E., Strijov V.V. Classification of time series in the space of parameters of generating models // Informatics and its applications, 2016. [[http://strijov.com/papers/Karasikov2016TSC.pdf URL]] | ||

+ | *# Kuznetsov M.P., Ivkin N.P. Algorithm for Classifying Accelerometer Time Series by Combined Feature Description // Machine Learning and Data Analysis. 2015. V. 1, No. 11. C. 1471 - 1483. [[http://jmlda.org/papers/doc/2015/no11/Ivkin2015TSclassification.pdf URL]] | ||

+ | * '''Base algorithm:''' [Karasikov 2016] | ||

+ | * '''Solution:''' See [[Media:Local_appr.pdf|The problem description]]. | ||

+ | * '''Novelty:''' When creating meta-prognostic models (predictive models of predictive models), the problem of using the values of parameters of local models when creating meta-models remains open. The purpose of the project below is to create a tool to analyze this problem. | ||

+ | * '''Authors:''' Strijov V.V. | ||

+ | |||

+ | ===6. 2017=== | ||

+ | * '''Title:''' Choosing the optimal recurrent network model in the Paraphrase Search The problems | ||

+ | * '''Problem:''' Given a selection of pairs of sentences labeled <<similar>> and <<dissimilar>>. It is required to build a recurrent network of low complexity (that is, with a small number of parameters) that delivers a minimum error in the classification of pairs of sentences. | ||

+ | * '''Data:''' It is proposed to consider two samples: [https://www.microsoft.com/en-us/download/details.aspx?id=52398 Microsoft Paraphrase Corpus] (a small set of sentences) and [http ://sitem.herts.ac.uk/aeru/ppdb/en/ PPDB] (set of short segments, markup not always correct) | ||

+ | * '''References:''' | ||

+ | *# [http://deeplearning.net/tutorial/lstm.html [1]] Step by step description of the implementation of the LSTM recurrent network | ||

+ | *# [http://www.cs.toronto.edu/~graves/nips_2011.pdf [2]] Thinning algorithm based on building a network with a minimum description length | ||

+ | *# [http://papers.nips.cc/paper/250-optimal-brain-damage.pdf Optimal Brain Damage] [3] | ||

+ | * '''Basic algorithm''': The basic algorithm can be: | ||

+ | *# Solution without thinning | ||

+ | *# Solution described in [3] | ||

+ | *# Optimal Brain Damage | ||

+ | * '''Solution:''' It is proposed to consider the thinning method described in [3] with a block covariance matrix: either neurons or parameters grouped by input features act as blocks. | ||

+ | * '''Novelty:''' The proposed method will effectively reduce the complexity of the recurrent network, taking into account the relationship between neurons or input features. | ||

+ | * '''Authors:''' Oleg Bakhteev, consultant | ||

+ | |||

+ | ===7. 2017=== | ||

+ | * '''Title:''' Internal plagiarism detection | ||

+ | * '''Problem:''' Solved by The problem to identify internal borrowings in text. It is required to test the hypothesis that the given text was written by a single author, and if it is not fulfilled, highlight the borrowed parts of the text. A borrowing is a part of the text, presumably written by another author and containing characteristic differences from the style of the main author. It is required to develop such a style function that allows to distinguish with a high degree of certainty the style of the main author of the text from borrowings. | ||

+ | * '''Data:''' It is proposed to consider the corpus PAN-2011, PAN-2016 | ||

+ | * '''References:''' | ||

+ | *# [http://deeplearning.net/tutorial/lstm.html [1]] Step by step description of the implementation of the LSTM recurrent network | ||

+ | *# [https://arxiv.org/pdf/1608.04485.pdf [2]] Author clustering algorithm | ||

+ | *# [http://www.fit.vutbr.cz/imikolov/rnnlm/thesis.pdf [3]] Statistical Language Models Based on Neural Networks | ||

+ | *# [https://pdfs.semanticscholar.org/1011/6d82a8438c78877a8a142be47c4ee8662138.pdf [4]] Methods for intrinsic plagiarism detection and author diarization | ||

+ | * '''Basic algorithm''': The solution described in [4] can be used as the Basic algorithm | ||

+ | * '''Solution:''' It is proposed to consider the method described in [2] and build a style function based on the neural network outputs. | ||

+ | * '''Novelty:''' It is assumed that the construction of a style function by the proposed method can give an increase in quality compared to typical solutions to this problem. | ||

+ | * '''Authors:''' Rita Kuznetsova, consultant | ||

+ | |||

+ | ===8. 2017=== | ||

+ | * '''Title:''' Adaptive relaxations of NP hard problems through machine learning | ||

+ | * '''Problem:''' Modern problems of optimizing power flows in power networks lead to non-convex optimization The problems with a large number of restrictions. Statements similar in structure also arise in a number of other engineering problems and in classical The problems of combinatorial optimization. The traditional approach to solving such NP hard problems is to write their convex relaxations (semidefinite/SDP, second order conic/SOCP, etc), which usually have a much larger set of feasible solutions than in the original problem. and by the subsequent projection of the obtained solution into the region where the constraints of the original problem are satisfied. In many practical cases, the quality of the solution obtained in this way is not high. Alternative approaches, for example MILP (mixed integer linear programming) relaxation, are substantially more time consuming but result in a more accurate answer. | ||

+ | The main problem is the impossibility of using known methods for solving large-scale problems (networks of 1000 nodes and more). One of the key obstacles is not so much the dimension of the problem as a large number of restrictions. At the same time, in real The problems it is possible to single out a small set of restrictions such that the sets of admissible points in the selected set and in the original one are very close. This will allow us to replace The problem with another one with fewer restrictions, which will increase the speed of the algorithms used. | ||

+ | It is proposed to use machine learning methods to build the indicated set of the most important constraints. | ||

+ | * '''References:''' Sampling/machine learning methods: | ||

+ | *# Beygelzimer, A., Dasgupta, S., & Langford, J. (2009, June). Importance weighted active learning. In Proceedings of the 26th annual international conference on machine learning (pp. 49-56). ACM. | ||

+ | *# Tong, S., & Koller, D. (2001). Support vector machine active learning with applications to text classification. Journal of machine learning research, 2(Nov), 45-66. | ||

+ | *# Owen, A., & Zhou, Y. (2000). Safe and effective importance sampling. Journal of the American Statistical Association, 95(449), 135-143. | ||

+ | Relaxations: Nagarajan, H., Lu, M., Yamangil, E., & Bent, R. (2016). Tightening McCormick Relaxations for Nonlinear Programs via Dynamic Multivariate Partitioning. arXiv preprint arXiv:1606.05806. | ||

+ | * '''Data:''' ieee + matpower data containing descriptions of energy networks and their modes of operation. | ||

+ | * '''Novelty:''' This approach seems to be the first application of applied statistics/machine learning methods to solve difficult optimization problems. We expect substantial gains in labor-intensive style methods | ||

+ | * '''Author''': consultant: Yuri Maksimov, Expert: Mikhail Chertkov | ||

+ | |||

+ | ===9. 2017=== | ||

+ | * '''Title:''' Optimal Algorithm for Reconstruction of Dynamic Models. | ||

+ | * '''Problem:''' A standard machine learning problem statement in the context of unsupervised learning assumes that the examples are independent and come from the same probability distribution. However, often observed data are of dynamic origin and are correlated. The problem is to develop an efficient method for restoring a dynamic graphical model (graph and model parameters) from observed correlated dynamic configurations. This The problem is theoretically important and has many applications. The basis of the algorithm will be the adaptation of a new optimal method of screening interactions (interaction screening), developed for the Ising model. The solution process will combine familiarity with computer science/machine learning theoretical methods and numerical experiments. | ||

+ | * '''Data:''' Simulated dynamic configurations of spins in the kinetic Ising model. | ||

+ | * '''References:''' | ||

+ | *# Lokhov et al., "Optimal structure and parameter learning of Ising models", arXiv:1612.05024 (2016) {https://arxiv.org/abs/1612.05024} | ||

+ | *# Vuffray et al., "Interaction screening: efficient and sample-optimal learning of Ising models", NIPS 2016 {https://arxiv.org/abs/1605.07252} | ||

+ | *# Decelle and Zhang, "Inference of the sparse kinetic Ising model using the decimation method", Phys. Rev. E 2016 {https://arxiv.org/abs/1502.01660} | ||

+ | *# Bresler et al., "Learning graphical models from the Glauber dynamics", Allerton 2014 {https://arxiv.org/abs/1410.7659} | ||

+ | *# Zeng et al., "Maximum likelihood reconstruction for Ising models with asynchronous updates", Phys. Rev. Lett. 2013 | ||

+ | * '''Base algorithm:''' Dynamic method for shielding interactions. Comparison with the maximum likelihood method. | ||

+ | * '''Novelty:''' Currently, the optimal (ie using the minimum possible number of examples) algorithm for this problem is unknown. The dynamic method of interaction screening has a good chance of finally "closing" this The problem, because is optimal for a static problem. | ||

+ | * '''Author''': consultants Andrey Lokhov, Yuri Maksimov. Expert Mikhail Chertkov | ||

+ | |||

+ | ===10. 2017=== | ||

+ | * '''Title:''' Choice of Interpreted Multimodels in Credit Scoring The problems | ||

+ | * '''Problem:''' The problem of credit scoring is to determine the level of creditworthiness of the borrower. For this, a borrower's questionnaire is used, containing both numerical (age, income) and categorical features (gender, profession). It is required, having historical information about the repayment of loans by other borrowers, to determine whether the borrower will return the loan. The data can be heterogeneous (example, if there are different income regions in a country), and several models will be needed to adequately classify. It is necessary to determine the optimal number of models. Based on the set of model parameters, it is necessary to draw up a portrait of the borrower. | ||

+ | * '''Data:''' It is proposed to consider five samples from the UCI and Kaggle repositories, with a capacity of 50,000 objects or more. | ||

+ | * '''References:''' A.A. Aduenko \MLAlgorithms\PhDThesis; C. Bishop, Pattern recognition and machine learning, final chapter; 20 years of Mixture experts. | ||

+ | * '''Base algorithm:''' Clustering and building independent logistic regression models, Adaboost, Decision Forest (with restrictions on complexity), Blend of Experts. | ||

+ | * '''Solution:''' An algorithm is proposed for selecting a multi-model (a mixture of models or a mixture of Experts) and determining the optimal number of models. | ||

+ | * '''Novelty:''' Proposed function of distance between models in which parameter distributions are given on different media. | ||

+ | * '''Authors:''' A.A. Aduenko, Strijov V.V. | ||

+ | |||

+ | ===11. 2017=== | ||

+ | * '''Title:''' Feature Selection in Problems of Autoregressive Prediction of Biomedical Signals. | ||

+ | * '''Problem:''' The problem of predicting biomedical signals and IoT signals is being solved. It is required to predict the vector - the next few signal samples. It is assumed that the proper dimension of the space of both the predicted variable and the independent variable can be significantly reduced, thereby increasing the stability of the forecast without significant loss of accuracy. For this, the Partial Least Squares approach in autoregressive forecasting is used. | ||

+ | * '''Data:''' SantaFe biomedical time series sample, IoT signal sample. | ||

+ | * '''References:''' Katrutsa A.M., Strijov V.V. Stresstest procedure for feature selection algorithms // Chemometrics and Intelligent Laboratory Systems, 2015, 142 : 172-183; : Katrutsa A.M., Strijov V.V. Comprehensive study of feature selection methods to solve multicollinearity problem according to evaluation criteria // Expert Systems with applications, 2017; Kee Siong Ng A Simple Explanation of Partial Least Squares keesiong.ng@gopivotal.com Draft, April 27, 2013, http://users.cecs.anu.edu.au/~kee/pls.pdf | ||

+ | * '''Base algorithm:''' PLS, quadratic optimization algorithm for feature selection. | ||

+ | * '''Solution:''' build a design matrix with a suboptimal set of objects and features, propose a quadratic optimization error function (if possible, develop it for the case of a tensor representation of the design matrix). | ||

+ | * '''Novelty:''' Generalized feature selection algorithm (published two weeks ago) for the PLS case. | ||

+ | * '''Authors:''' A.M. Katrutsa, Strijov V.V. | ||

+ | |||

+ | ===12. 2017=== | ||

+ | * '''Title:''' Massively multiThe problem deep learning for drug discovery | ||

+ | * '''Problem:''' Develop a multi-The problem recurrent neural network to predict biological activity. For each molecule-protein pair, it is required to predict the binary value 0/1, which means that the molecule binds/does not bind to the protein. | ||

+ | * '''Data:''' sparse biological activity data for ~100K molecules versus ~1000 proteins. Molecules are represented as SMILES strings (sequence of characters encoding a molecule) | ||

+ | * '''References:''' https://arxiv.org/pdf/1502.02072 | ||

+ | * '''Base algorithm:''' multi-The problem neural network that predicts activity by numerical features, single-The problem recurrent neural network | ||

+ | * '''Solution:''' MultiThe probleming means that you need to build a model that is obtained for the input of a molecule and predicts its biological activity against all proteins in the sample. | ||

+ | * '''Novelty:''' Existing methods did not show a significant improvement in the quality of the DL model compared to standard ML models | ||

+ | * '''Authors:''' Expert -- Alexander Isaev, consultant -- Maria Popova | ||

+ | |||

+ | ===13. 2017=== | ||

+ | * '''Title:''' Unsupervised representation for molecules | ||

+ | * '''Problem:''' Develop an unsupervised method for representing molecules | ||

+ | * '''Data:''' ~1.5M molecules in SMILES string format (character sequence encoding the molecule) | ||

+ | * '''References:''' https://www.cs.toronto.edu/~hinton/science.pdf | ||

+ | * '''Base algorithm:''' currently hand-selected numerical features are used as such representation. The quality of the resulting representations can be compared with the tox21 dataset (10K molecules versus 12 proteins) | ||

+ | * '''Solution:''' use convolutional or recurrent networks to build an autoencoder. | ||

+ | * '''Novelty:''' building an end-to-end model to get informative features | ||

+ | * '''Authors:''' Expert -- Alexander Isaev, consultant -- Maria Popova | ||

+ | |||

+ | ===14. 2017=== | ||

+ | * '''Title:''' Intratext coherence as a measure of interpretability of thematic models of text collections. | ||

+ | * '''Problem:''' Interpretability is a subjective measure of the quality of topic models, as measured by Expert Scores. Coherence is a measure of the occurrence of thematic words, calculated automatically from the text and correlates well with interpretability, as shown in the Newman and Mimno series. The first The problem is to evaluate the representativeness of the sequence of words in the text, according to which the coherence is estimated. The second The problem is to compare several new methods for measuring interpretability and coherence based on the selection of the most representative sequence of words in the source text. | ||

+ | * '''Data:''' A collection of popular science content PostNauka, a collection of news content. | ||

+ | * '''References:''' | ||

+ | *# Vorontsov K. V. [[Media:voron17survey-artm.pdf|Review of probabilistic thematic models]], 2017. | ||

+ | *# N.Aletras, M.Stevenson. Evaluating Topic Coherence Using Distributional Semantics, 2013. | ||

+ | *# D. Newman et al. Automatic evaluation of topic coherence, 2010 | ||

+ | *# D.Mimno et al. Optimizing semantic coherence in topic models, 2011 | ||

+ | *# http://palmetto.aksw.org/palmetto-webapp/ | ||

+ | * '''Base algorithm:''' Standard methods for estimating the interpretability and coherence of topics in topic models. | ||

+ | * '''Solution:''' A new method for measuring interpretability and coherence, experiments to find the most correlated measures of interpretability and coherence, similar to [D.Newman, 2010]. | ||

+ | * '''Novelty:''' inline measures of interpretability and coherence were not previously proposed. | ||

+ | * '''Authors:''' Vorontsov K. V.. consultants: Viktor Bulatov, Anna Potapenko, Artyom Popov. | ||

+ | |||

+ | ===15. 2017=== | ||

+ | * '''Title:''' Aggregation of heterogeneous text collections in a hierarchical thematic model of Russian-language popular science content. | ||

+ | * '''Problem:''' Implement and compare multiple ways of combining text collections from different sources into one hierarchical topic model. Build a classifier that determines the presence of a topic in the source. | ||

+ | * '''Data:''' Collection of popular science content PostNauka, Wikipedia collection. | ||

+ | * '''References:''' | ||

+ | *# Vorontsov K. V. [[Media:voron17survey-artm.pdf|Review of probabilistic thematic models]], 2017. | ||

+ | *# Chirkova N. A, Vorontsov K. V. [http://jmlda.org/papers/doc/2016/no2/Chirkova2016hARTM.pdf Additive regularization of multimodal hierarchical topic models] // Machine Learning and Data Analysis, 2016. T. 2. No. 2. | ||

+ | * '''Base algorithm:''' An algorithm for constructing a thematic hierarchy in BigARTM, implemented by Nadezhda Chirkova. Marking tool | ||

+ | * '''Solution:''' Build a topic model with source modalities and highlight topics specific to only one of the sources. Prepare a sample for training a classifier that determines the presence of a topic in the source. | ||

+ | * '''Novelty:''' Additive regularization of topic models has not been applied to this problem before. | ||

+ | * '''Authors:''' Vorontsov K. V.. consultants: Alexander Romanenko, Irina Efimova, Nadezhda Chirkova. | ||

+ | |||

+ | ===16. 2017=== | ||

+ | * '''Title:''' Application of the methods of symbolic dynamics in the technology of informational analysis of electrocardiosignals. | ||

+ | * '''Problem:''' The technology of informational analysis of electrocardiosignals, proposed by V.M.Uspensky, involves converting a raw signal into a character sequence and searching for disease patterns in this sequence. So far, symbolic n-grams have been predominantly used to search for patterns. In the framework of this work, it is proposed to expand the class of templates in which the search for diagnostic signs of diseases is performed. Quality criterion -- AUC and MAP ranking of diagnoses. | ||

+ | * '''Data:''' A selection of electrocardiograms with known diagnoses. | ||

+ | * '''References:''' | ||

+ | *# Uspensky V.M. Informational function of the heart. Theory and practice of diagnosing diseases of internal organs by the method of information analysis of electrocardiosignals. - M .: "Economics and Information", 2008. - 116s | ||

+ | *# Technology of information analysis of electrocardiosignals. | ||

+ | * '''Base algorithm:''' Classification methods . | ||

+ | * '''Solution:''' Search for logical patterns in character strings, methods of character dynamics, comparison of algorithms according to the quality criteria AUC and MAP (diagnosis ranking). | ||

+ | * '''Novelty:''' So far, character n-grams have been used predominantly to search for patterns. | ||

+ | * '''Authors:''' Vorontsov K. V.. consultants: Vlada Tselykh. | ||

+ | |||

+ | === Vorontsov The problems +=== | ||

+ | * '''Title''': Dynamic hierarchical thematic model of the news flow. | ||

+ | * '''Problem:''' Develop an algorithm for classifying topics in news flows into new and ongoing ones. Apply the obtained criteria for creating new topics at all levels of the topic model hierarchy when adding the next piece of data to the text collection (for example, all news for one day). | ||

+ | * '''Data:''' Collection of news in Russian. A subsample of news classified into two classes: new and ongoing topics. | ||

+ | * '''Literature''': | ||

+ | *#''Vorontsov K.V.'' [[Media:voron17survey-artm.pdf|Review of probabilistic thematic models]], 2017. | ||

+ | *#''Chirkova N. A, Vorontsov K. V.'' [http://jmlda.org/papers/doc/2016/no2/Chirkova2016hARTM.pdf Additive regularization of multimodal hierarchical topic models] // Machine Learning and Data Analysis , 2016 T. 2. No. 2. | ||

+ | * '''Basic Algorithm''': An algorithm for constructing a thematic hierarchy in BigARTM, implemented by Nadezhda Chirkova. Known Topic Detection & Tracking algorithms. | ||

+ | * '''Solution''': Using BigARTM, selecting regularizers and their parameters, using the topic selection regularizer. Building an algorithm for classifying topics into new and ongoing. | ||

+ | * '''Novelty''': Additive regularization of topic models has not been applied to this problem before. | ||

+ | * '''Authors''': KV Vorontsov. Consultants: Alexander Romanenko, Artyom Popov. | ||

+ | |||

+ | ===Antiplagiarism + === | ||

+ | * '''Title:''' Selection of Candidates in the Problem of Finding Text Borrowings with Paraphrasing Based on the Vectorization of Text Fragments. | ||

+ | * '''Problem:''' Searching for text borrowings in a collection of documents involves selecting a small set of candidates for subsequent detailed analysis. The Candidate Selection The problem is formulated as finding the optimal ranking of documents in a collection for a query with respect to some function that is an estimate for the total length of borrows from a collection document to a query document. | ||

+ | * '''Data:''' [http://pan.webis.de/clef11/pan11-web/plagiarism-detection.html PAN] | ||

+ | * '''References:''' | ||

+ | *# Romanov A.V., Khritankov A.S. Selection of candidates when searching for borrowings in a collection of documents in a foreign language [http://www.machinelearning.ru/wiki/images/c/c4/6.Romanov .pdf] | ||

+ | * '''Basic algorithm''': shingles method with reverse index construction. | ||

+ | * '''Solution:''' Vectorization of text fragments (word embeddings + convolutional / recurrent neural networks) and subsequent search for nearest objects in a multidimensional metric space. | ||

+ | * '''Novelty:''' a new approach to solving the problem. | ||

+ | * '''Authors:''' Alexey Romanov (consultant) | ||

+ | |||

+ | Additional projects | ||

+ | === Vorontsov+=== | ||

+ | * '''Title:''' Thematic modeling of an economic sector based on bank transaction data. | ||

+ | * '''Problem:''' Test the hypothesis that a large sample of transactions between firms is adequately described by a relatively small set of economic activities (aka topics). The problem is reduced to decomposing the matrix of transactional data "buyers × sellers" into the product of three non-negative matrices "buyers × topics", "topics × topics", "topics × sellers", while the middle matrix describes a directed graph of financial flows in the industry. It is required to compare several methods for constructing such expansions and find the number of topics for which the observed set of transactions is modeled with sufficient accuracy. | ||

+ | * '''Data:''' selection of transactions between firms, such as "buyer, seller, volume". | ||

+ | * '''References:''' | ||

+ | *# Vorontsov K. V. [[Media:voron17survey-artm.pdf|Review of probabilistic thematic models]], 2017. | ||

+ | * '''Base algorithm:''' Standard methods for non-negative matrix expansions. | ||

+ | * '''Solution:''' Regularized EM-algorithm for sparse non-negative matrix expansions. Visualization of the graph of financial flows. Testing the algorithm on synthetic data, testing the hypothesis about the stability of sparse solutions. | ||

+ | * '''Novelty:''' Thematic modeling has not previously been applied to the analysis of financial transactional data. | ||

+ | * '''Authors:''' Vorontsov K. V.. consultants: Viktor Safronov, Rosa Aisina. | ||

+ | |||

+ | ===scoring+=== | ||

+ | * '''Title:''' Generating and selecting features when building a credit scoring model. | ||

+ | * '''Problem:''' Credit scoring models are built step by step. In particular, a number of independent transformations of individual features are performed, and new features are generated. Each step uses its own quality criterion. It is required to build a scoring model that adequately describes the sample. Maximizing the quality of the model at each step does not guarantee the maximum quality of the resulting model. It is proposed to abandon the step-by-step construction of the scoring model. To do this, the quality criterion must include all the optimized parameters of the model. | ||

+ | * '''Data:''' The computational experiment will be performed on 5-7 samples to be found. It is desirable that the samples be of the same nature, for example, the samples of consumer credit questionnaires. | ||

+ | * '''References:''' Siddique N. Constructing scoring models, SAS. Hosmer D., Lemeshow S., Applied logistic regression, Wiley. Katrutsa A.M., Strijov V.V. Comprehensive study of feature selection methods to solve multicollinearity problem according to evaluation criteria // Expert Systems with applications, 2017. | ||

+ | * '''Base algorithm:''' The scoring model construction algorithm recommended by SAS. | ||

+ | * '''Solution:''' Each step of the procedure is represented as an optimization problem. The parameters to be optimized are combined, and the Feature Selection The problem is included as a Mixed Optimization The problem. | ||

+ | * '''Novelty:''' An error function is proposed, when using which the generation and selection of features, as well as the optimization of model parameters, are performed together. | ||

+ | * '''Authors:''' T.V. Voznesenskaya, Strijov V.V. | ||

+ | |||

+ | ===Popova+=== | ||

+ | * '''Title:''' Representation of molecules in 3D | ||

+ | * '''Problem:''' Develop representations of the 3D structure of molecules that would have the property of rotational and translational invariance. | ||

+ | * '''Data:''' Millions of molecules given by 3D coordinates | ||

+ | * '''References:''' https://arxiv.org/abs/1610.08935, http://journals.aps.org/prl/abstract/10.1103/PhysRevLett.98.146401 | ||

+ | * '''Base algorithm:''' low rank matrix/tensor factorization | ||

+ | * '''Solution:''' Molecules have a different number of atoms, and therefore the matrix of their 3D coordinates is Nx3. We need to find a mathematical transformation that would be independent of N (N is the number of atoms). | ||

+ | * '''Novelty:''' existing algorithms depend on the number of atoms in the molecule | ||

+ | * '''Authors:''' Expert -- Alexander Isaev, consultant -- Maria Popova | ||

+ | |||

+ | ===Maksimov+=== | ||

+ | * '''Title:''' Optimal algorithm for recovering block Hamiltonians (XY and Heisenberg models). | ||

+ | * '''Problem:''' The problem is to reconstruct block Hamiltonians with continuous spins (a generalization of the Ising model to two- and three-dimensional spins) from the observed data. This setting is a special case of a field of machine learning known as unsupervised learning. Reconstruction of a graphical spin model from observational data is an important problem in physics. The basis of the algorithm will be the adaptation of a new optimal method of screening interactions (interaction screening), developed for the Ising model. The solution process will combine familiarity with computer science/machine learning theoretical methods and numerical experiments. | ||

+ | * '''Data:''' Simulated block spin model configurations. | ||

+ | * '''References:''' | ||

+ | *# Lokhov et al., "Optimal structure and parameter learning of Ising models", arXiv:1612.05024 (2016) {https://arxiv.org/abs/1612.05024} | ||

+ | *# Vuffray et al., "Interaction screening: efficient and sample-optimal learning of Ising models", NIPS 2016 {https://arxiv.org/abs/1605.07252} | ||

+ | *# Tyagi et al., "Regularization and decimation pseudolikelihood approaches to statistical inference in XY spin models", Phys. Rev. B 2016 {https://arxiv.org/abs/1603.05101} | ||

+ | * '''Base algorithm:''' Dynamic method for shielding interactions. Comparison with the method of maximum pseudo-likelihood (pseudolikelihood). | ||

+ | * '''Novelty:''' An algorithm based on the dynamic interaction shielding method has a good chance of being optimal for this problem, because the corresponding method is optimal for the inverse Ising problem. | ||

+ | * '''Author''': consultants Andrey Lokhov, Yuri Maksimov. Expert Mikhail Chertkov | ||

+ | |||

+ | ===Khritankova (Transfer Learning) === | ||

+ | * '''Title:''' Using deep learning networks to transfer classification models in case of insufficient data. | ||

+ | * '''Problem description:''' | ||

+ | *# Develop an algorithm for calculating a set of latent features in the symmetric homogeneous transfer learning problem, the solution of the classification problem in which does not depend on the original area, and which is no worse than when solving for each area separately (transfer error) for the case of small sample sizes with errors in markup | ||

+ | *# Develop an algorithm for transitioning to a hidden set of features without using markup (unsupervised domain adaptation) | ||

+ | * '''Data:''' teraPromise-CK (33 datasets with the same features but different distributions). | ||

+ | * '''References:''' Base article: Xavier Glorot , Antoine Bordes , Yoshua Bengio. (2011) Domain Adaptation for Large-Scale sentiment classification: A Deep Learning approach / In Proceedings of the Twenty-eight International Conference on Machine Learning, ICML. | ||

+ | Articles with ideas for improving the algorithm will be handed out (several). | ||

+ | * '''Base algorithm:''' SDA (Stacked Denoising Autoencoder) – described in the Glorot et al. | ||

+ | * '''Solution:''' Take the Basic algorithm, a) try to improve it for application to small datasets of 100-1000 objects (when transfer learning is applied) by applying regularizers, adjusting the architecture of the autoencoder, adjusting the learning algorithm (for example, bootstrapping) b ) investigate the model for resistance to markup errors (label corruption / noisy labels) and propose improvements to increase stability (robustness). | ||

+ | * '''Novelty:''' Obtaining a stable algorithm for transferring classification models on small amounts of data with markup errors. | ||

+ | * '''Authors:''' Khritankov | ||

+ | |||

+ | ===INRIA=== | ||

+ | * '''Title:''' Estimated binding energy of protein and small molecules. | ||

+ | * '''Problem:''' Modeling the binding of a protein and a small molecule (hereinafter referred to as a ligand) is based on the fact that the best ligand in its best position has the lowest free energy of interaction with the protein. It is necessary to estimate the free energy of protein and ligand binding. Complexes of proteins with ligands can be used for training, and for each protein there are several positions of the ligand: 1 correct, "native", for which the energy is minimal, and several generated incorrect ones. For a third of the data set, values are known that are proportional to the desired binding energy of ligands in native positions with the protein. There is a separate test set consisting of 1) complexes of proteins and ligands, for which it is necessary to find the best ligand position (the algorithm for obtaining ligand positions differs from that used in training), 2) complexes of proteins and ligands, for whose native positions it is necessary to predict the binding energy, and 3) proteins for which it is necessary to find the most strongly binding ligand. | ||

+ | * '''Data:''' About 10000 complexes: for each of them there is 1 native pose and 18 (more can be generated) non-native ones. The main descriptors are histograms of distributions of distances between different atoms of the protein and ligand, the dimension of the vector of descriptors is ~ 20,000. The set of descriptors can be extended (you can generate poses with different deviations and use it as a descriptor, you can add the properties of small molecules: the number of bonds around which rotation is possible in a molecule, its surface area, its surface division by a Voronoi diagram. The data will be provided in the form of binary files with a python script to read. | ||

+ | * '''References:''' PEPSI-Dock: a detailed data-driven protein–protein interaction potential accelerated by polar Fourier correlation Predicting Binding Poses and Affinities in the CSAR 2013―2014 Docking Exercises Using the Knowledge-Based Convex-PL Potential | ||

+ | * '''Base algorithm:''' We used a linear SVM (these are just lecture notes, I see no reason to give Vapnik here, especially since all this, including these lecture notes, is googled), the connection of which with an energy estimate that goes beyond scope of the classification The problem is described in the articles listed above. To take into account experimentally known values proportional to energy, it is proposed to use linear regression SVR . | ||

+ | * '''Solution:''' It is necessary to reduce the previously used SVM problem to a regression problem and solve it using standard methods. To check the operation of the algorithm, both the test described above and several other test sets with similar The problems but different data will be used. | ||

+ | * '''Novelty:''' Proper assessment of the quality of protein and ligand binding is used in drug development to find molecules that interact most strongly with the protein under study. | ||

+ | Of particular importance is the assessment of the values of the binding energy of the protein with the ligand: the coefficient of correlation (Pearson) of the energy with its experimental values determined by different groups on the proposed test does not exceed 0.7. Prediction of the most strongly binding ligand from a large number of non-protein-binding molecules is also difficult. The aim of this work is to obtain a method that allows a fairly accurate assessment of protein binding to ligands. From the point of view of machine learning and optimization, it is of interest to combine classification and regression problems. | ||

+ | * '''Appendix''' Given several data sets describing an atom in a molecule or a bond between atoms, with a small feature vector (usually 3-10 descriptors) and several classes corresponding to the atom's hybridization or bond order. The data itself can be from ~100 to 20,000 vectors depending on the type of atom. You need to test some kind of multiclass machine learning on this (random forests, neural network, something else), you can do anything with descriptors. We are currently using SVM. Not only the accuracy is important, but also the computational complexity of the prediction. | ||

+ | * '''Authors:''' Sergei Grudinin, Maria Kadukova | ||

+ | |||

+ | ===Strijov and Kulunchakov+=== | ||

+ | * '''Title:''' Creation of delay-operators for multiscale forecasting by means of symbolic regression | ||

+ | * '''Problem:''' Suppose that one needs to build a forecasting machine for a response variable. Given a large set of time series, one can advance a hypothesis that they are related to this variable. Relying upon this hypothesis, we can use given time series as features for the forecasting machine. However, the values of time series could be produced with different frequencies. Therefore, we should take into account not only the values, but the delays as well. The simplest model for forecast is a linear one. In the presence of large set of features this model can approximate the response quite well. To avoid the problem of multiscaling, we introduce a definition of delay-operators. Each delay-operator corresponds to one time series and represents continuous correlation function. This correlation function shows a dependence between the response variable and corresponding time series. Therefore, each delay-operator put weights on the values of corresponding time series depending on the greatness of the delay. Having these delay-operators, we avoid the problem of multiscaling. To find them, we use genetic programming and symbolic regression. If the resulted weighted linear regression model would produce poor approximation, we can use a nonlinear one instead. To find good nonlinear function, we would use symbolic regression as well. | ||

+ | * '''Data:''' Any data from the domain of multiscalse forecating of time series. See the [[Media:Kulunchakov2016MultiscaleForecast.pdf|full version]] of this introduction. | ||

+ | * '''References:''' to be handed by V.V.Strijov | ||

+ | * '''Base algorithm:''' to be handed by V.V.Strijov | ||

+ | * '''Solution:''' Use genetic algorithms applied to symbolic regression to create and test delay-operators in multiscale forecasting. | ||

+ | * '''Novelty:''' to be handed by V.V.Strijov | ||

+ | * '''Authors:''' supervisor: V.V.Strijov, consultant: A.S. Kulunchakov | ||

+ | |||

+ | ==2016== | ||

+ | {|class="wikitable" | ||

+ | |- | ||

+ | ! Author | ||

+ | ! Topic | ||

+ | ! Link | ||

+ | ! Consultant | ||

+ | ! Reviewer | ||

+ | ! Report | ||

+ | ! Letters | ||

+ | ! Grade | ||

+ | ! Journal | ||

+ | |- | ||

+ | |Bayandina Anastasia | ||

+ | |Thematic models of distributive semantics for highlighting ethno-relevant topics in social networks | ||

+ | |[https://svn.code.sf.net/p/mlalgorithms/code/Group374/Bayandina2016TopicModeling/doc/Bayandina2016TopicModeling.pdf paper] | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group374/Bayandina2016TopicModeling/doc/Bayandina2016TopicModelingPresentation.pdf slides] | ||

+ | [https://www.youtube.com/watch?v=7IbYWWO_evY video] | ||

+ | |Anna Potapenko | ||

+ | |Oleg Gorodnitsky | ||

+ | |BF | ||

+ | |AILSB++RCVTDEWHS | ||

+ | |10 | ||

| | | | ||

+ | |- | ||

+ | |Belozerova Anastasia | ||

+ | |Coordination of logical and linear classification models in the information analysis of electrocardiosignals | ||

+ | |[https://svn.code.sf.net/p/mlalgorithms/code/Group374/Belozerova2016LogicLinearClassificator/code code] | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group374/Belozerova2016LogicLinearClassificator/doc/Belozerova2016LogicLinearClassificator.pdf paper] | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group374/Belozerova2016LogicLinearClassificator/doc/Belozerova2016Presentation.pdf slides] | ||

+ | [https://www.youtube.com/watch?v=3XhaIN1bgDI video] | ||

+ | |Vlada Tselykh | ||

+ | |Malygin Vitaly | ||

+ | |BF | ||

+ | |AILSB+RC+VTD>E0WH>S | ||

+ | |10 | ||

| | | | ||

+ | |- | ||

+ | |Maria Vladimirova | ||

+ | |Bagging of neural networks in the problem of predicting the biological activity of cell receptors | ||

+ | |[https://svn.code.sf.net/p/mlalgorithms/code/Group374/Vladimirova2016BaggingNN/code code] | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group374/Vladimirova2016BaggingNN/doc/Vladimirova2016BaggingNN.pdf paper] | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group374/Vladimirova2016BaggingNN/doc/Vladimirova2016Presentation.pdf slides] | ||

+ | [https://www.youtube.com/watch?v=pPumIZ81KU4 vido] | ||

+ | |Maria Popova | ||

+ | |Volodin Sergey | ||

+ | |BMF | ||

+ | |AILSBRCVTD>E>WHS | ||

+ | |10 | ||

| | | | ||

|- | |- | ||

- | | | + | |Volodin Sergey |

+ | |A probabilistic approach to the problem of predicting the biological activity of nuclear receptors | ||

+ | |[https://svn.code.sf.net/p/mlalgorithms/code/Group374/Volodin2016ProbabilisticReceptorPrediction/code code] [https://svn.code.sf.net/p/mlalgorithms/code/Group374/Volodin2016ProbabilisticReceptorPrediction/doc/Volodin2016ProbabilisticReceptorPrediction.pdf paper] [https://svn.code.sf.net/p/mlalgorithms/code/Group374/Volodin2016ProbabilisticReceptorPrediction/doc/Volodin2016ProbabilisticReceptorPredictionSlides.pdf slides] | ||

+ | [https://www.youtube.com/watch?v=TsQ8v778d0s video], [http://itas2016.iitp.ru/pdf/1570303389.pdf itis] | ||

+ | |Maria Popova | ||

+ | |Maria Vladimirova | ||

+ | |BMF | ||

+ | |AILSBRCVTDEWHS | ||

+ | |10 | ||

| | | | ||

+ | |- | ||

+ | |Gorodnitsky Oleg | ||

+ | |An Adaptive Nonlinear Method for Recovering a Matrix from Partial Observations | ||

+ | |[http://svn.code.sf.net/p/mlalgorithms/code/Group374/Gorodnitskii2016AdaptiveApproximation/code code] | ||

+ | [http://svn.code.sf.net/p/mlalgorithms/code/Group374/Gorodnitskii2016AdaptiveApproximation/doc/Gorodnitskii2016AdaptiveApproximation2.pdf paper] | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group374/Gorodnitskii2016AdaptiveApproximation/doc/Gorodnitskii2016NNMF.pdf slides], [http://itas2016.iitp.ru/pdf/1570303466.pdf itis] | ||

+ | |Mikhail Trofimov | ||

+ | |Bayandina Anastasia | ||

+ | |M | ||

+ | |A++I++L++S+B+R+C++VTDE+WH | ||

+ | |10 | ||

| | | | ||

+ | |- | ||

+ | |Ivanychev Sergey | ||

+ | |Synergy of classification algorithms (SVM Multimodelling) | ||

+ | |[https://svn.code.sf.net/p/mlalgorithms/code/Group374/Ivanychev2016SVM_Multimodelling/code/ code] | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group374/Ivanychev2016SVM_Multimodelling/doc/Ivanychev2016SVM_Multimodelling.pdf paper] | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group374/Ivanychev2016SVM_Multimodelling/doc/Ivanychev2016SVM_Slides.pdf slides] | ||

+ | |Alexander Aduenko | ||

+ | | | ||

+ | |BM | ||

+ | |A+I+L++S+BRCVTDEW+H | ||

+ | |10 | ||

| | | | ||

|- | |- | ||

- | | | + | |Kovaleva Valeria |

+ | |Regular structure of rare macromolecular clusters | ||

+ | |[https://svn.code.sf.net/p/mlalgorithms/code/Group374/Kovaleva2016Spectra/code/ code] | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group374/Kovaleva2016Spectra/doc/Kovaleva2016Spectra.pdf paper] | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group374/Kovaleva2016Spectra/doc/Kovaleva2016Spectra_slides.pdf slides] | ||

+ | [https://www.youtube.com/watch?v=JaeyrqJr1KU video], [http://itas2016.iitp.ru/pdf/1570303499.pdf itis] | ||

+ | |Olga Valba, Yuri Maksimov | ||

+ | |Dmitry Fedoryaka | ||

+ | |BM | ||

+ | |A+IL+SBRCVTD0E0WH | ||

+ | |10 | ||

| | | | ||

+ | |- | ||

+ | |Makarchuk Gleb | ||

+ | |Time series transformations for hand motion decoding using ECoG signals (electrocorticographic signals) of monkeys | ||

+ | |[http://svn.code.sf.net/p/mlalgorithms/code/Group374/Makarchuk2016ECoGSignals/code code], | ||

+ | [http://svn.code.sf.net/p/mlalgorithms/code/Group374/Makarchuk2016ECoGSignals/doc/Makarchuk2016ECoGSignals.pdf paper] | ||

+ | [http://svn.code.sf.net/p/mlalgorithms/code/Group374/Makarchuk2016ECoGSignals/doc/Makarchuk2016ECoGSignalsPresentation.pdf slides] | ||

+ | [https://www.youtube.com/watch?v=xQvszwD9JAE video] | ||

+ | |Andrey Zadayanchuk | ||

| | | | ||

+ | |BF | ||

+ | |AI+L+S+BRС>V>T+D>E0WH>S | ||

+ | |10 | ||

| | | | ||

|- | |- | ||

- | | | + | |Malygin Vitaly |

+ | |Application of combinatorial estimates of retraining of threshold decision rules for feature selection in the problem of medical diagnostics by the method of V. M. Uspensky | ||

+ | |[http://svn.code.sf.net/p/mlalgorithms/code/Group374/Malygin2016FeatureSelection/code code], | ||

+ | [http://svn.code.sf.net/p/mlalgorithms/code/Group374/Malygin2016FeatureSelection/doc/Malygin2016FeatureSelection.pdf paper], | ||

+ | [http://svn.code.sf.net/p/mlalgorithms/code/Group374/Malygin2016FeatureSelection/doc/Malygin2016FSPresentation.pdf slides] | ||

+ | |Shaura Ishkina | ||

+ | |Belozerova Anastasia | ||

+ | |B | ||

+ | |AILSBRCVTDEWH | ||

+ | |10 | ||

| | | | ||

+ | |- | ||

+ | |Molibog Igor | ||

+ | |Using Dimension Reduction Methods When Building a Feature Space in the Problem of Internal Plagiarism Detection | ||

| | | | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group374/Molybog2016DimReduction/doc/MolybogMotrenkoStrijov2017DimRed.pdf paper], | ||

+ | [http://svn.code.sf.net/p/mlalgorithms/code/Group374/Molybog2016DimReduction/doc doc], | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group374/Molybog2016DimReduction/doc/Molybog2016DimReduction_Presentation.pdf slides], [http://itas2016.iitp.ru/pdf/1570303407.pdf itis] | ||

+ | |Anastasia Motrenko | ||

+ | |Safin Kamil | ||

+ | |BMF | ||

+ | |AILSBRCVTDEWHS | ||

+ | |10 | ||

| | | | ||

|- | |- | ||

- | | | + | |Pogodin Roman |

+ | |Determining the position of proteins using an electronic map | ||

+ | |[https://svn.code.sf.net/p/mlalgorithms/code/Group374/Pogodin2016ProteinsFitting/code code], [http://svn.code.sf.net/p/mlalgorithms/code/Group374/Pogodin2016ProteinsFitting/doc/Pogodin2016ProteinsFitting.pdf paper], [http://svn.code.sf.net/p/mlalgorithms/code/Group374/Pogodin2016ProteinsFitting/doc/Pogodin2016ProteinsFittingPresentation.pdf slides] | ||

+ | [https://www.youtube.com/watch?v=0DskvHR4waE video], [http://itas2016.iitp.ru/pdf/1570303519.pdf itis] | ||

+ | |Alexander Katrutsa | ||

+ | |Andrey Ryazanov | ||

+ | |BMF | ||

+ | |AILSBRСVTDEWHS | ||

+ | |10 | ||

| | | | ||

+ | |- | ||

+ | |Andrey Ryazanov | ||

+ | |Restoration of the primary structure of a protein according to the geometry of its main chain | ||

+ | |[https://svn.code.sf.net/p/mlalgorithms/code/Group374/Ryazanov2016InverseFolding/ folder] | ||

+ | [http://svn.code.sf.net/p/mlalgorithms/code/Group374/Ryazanov2016InverseFolding/doc/Ryazanov2016InverseFolding.pdf paper] | ||

+ | [http://svn.code.sf.net/p/mlalgorithms/code/Group374/Ryazanov2016InverseFolding/doc/Ryazanov2016InverseFoldingPresentation.pdf slides] | ||

+ | [https://www.youtube.com/watch?v=ZGx14xat2Jg video], [http://itas2016.iitp.ru/pdf/1570303468.pdf itis] | ||

+ | |Mikhail Karasikov | ||

+ | |Roman Pogodin | ||

+ | |BMF | ||

+ | |AIL+SBRC++VTD+EWHS | ||

+ | |10 | ||

| | | | ||

+ | |- | ||

+ | |Safin Kamil | ||

+ | |Definition of borrowings in the text without indicating the source | ||

+ | |[https://svn.code.sf.net/p/mlalgorithms/code/Group374/Safin2016IntrinsicPlagiarism/code code], [https://svn.code.sf.net/p/mlalgorithms/code/Group374/Safin2016IntrinsicPlagiarism/doc/Safin2016IntrinsicPlagiarism.pdf paper] | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group374/Safin2016IntrinsicPlagiarism/doc/Safin2016Presentation1.pdf slides] | ||

+ | [https://www.youtube.com/watch?v=lHYH1f5kYXU video] | ||

+ | |Mikhail Kuznetsov | ||

+ | |Molibog Igor | ||

+ | |BMF | ||

+ | |AIL+SBRC>V>T>D>E0WHS | ||

+ | |10 | ||

| | | | ||

|- | |- | ||

- | | | + | |Dmitry Fedoryaka |

+ | |Mixtures of vector autoregression models in the problem of time series forecasting | ||

+ | |[http://svn.code.sf.net/p/mlalgorithms/code/Group374/Fedoriaka2016TimeSeriesPrediction/code code], | ||

+ | [http://svn.code.sf.net/p/mlalgorithms/code/Group374/Fedoriaka2016TimeSeriesPrediction/doc/Fedoriaka2016TSPPresentation.pdf slides], | ||

+ | [http://svn.code.sf.net/p/mlalgorithms/code/Group374/Fedoriaka2016TimeSeriesPrediction/doc/Fedoriaka2016TimeSeriesPrediction.pdf paper] | ||

+ | |Radoslav Neichev | ||

+ | |Kovaleva Valeria | ||

+ | |BM | ||

+ | |AILSBRCV-T>D0E0WH> | ||

+ | |10 | ||

| | | | ||

+ | |- | ||

+ | |Tsvetkova Olga | ||

+ | |Building scoring models in the SAS system | ||

+ | |[https://svn.code.sf.net/p/mlalgorithms/code/Group374/Tsvetkova2016ScoringCards/code code], | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group374/Tsvetkova2016ScoringCards/doc/ScoringCards.pdf paper] | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group374/Tsvetkova2016ScoringCards/doc/presentation.pdf slides] | ||

+ | |Raisa Jamtyrova | ||

+ | |Chygrynskiy Viktor | ||

+ | |BF | ||

+ | |A+I+L+S+B+R+C+V0T0D0E0WH>S | ||

+ | |10 | ||

+ | | | ||

+ | |- | ||

+ | |Chygrynskiy Viktor | ||

+ | |Approximation of the boundaries of the iris | ||

+ | |[https://svn.code.sf.net/p/mlalgorithms/code/Group374/Chigrinskiy2016ApproximationOfIrisBoundaries/code code] [https://svn.code.sf.net/p/mlalgorithms/code/Group374/Chigrinskiy2016ApproximationOfIrisBoundaries/doc/Chigrinskiy2016ApproximationOfIrisBoundaries.pdf paper] | ||

+ | [https://svn.code.sf.net/p/mlalgorithms/code/Group374/Chigrinskiy2016ApproximationOfIrisBoundaries/doc/Chigrinskiy2016ApproximationOfIrisBoundariesSlides.pdf slides] | ||

+ | [https://www.youtube.com/watch?v=3kuNMYhVBw4 video] | ||

+ | |Yuri Efimov | ||

| | | | ||

+ | |B | ||

+ | |AI+L+SBRCV+TDEHFS | ||

+ | |10 | ||

| | | | ||

|- | |- | ||

- | |11 | + | |} |

+ | ===1. 2016=== | ||

+ | * '''Data:''' Synergy of classification algorithms. Data from the UCI repository so that it can be compared directly with other works, in particular the work of Vapnik. | ||

+ | * '''References:''' There are different approaches to combining SVMs: on example, bagging (http://www.ecse.rpiscrews.us/~cvrl/FaceProject/Homepage/Publication/ICPR04_final_cameraready_v4.pdf), also try and boosting (http://www.researchgate.net/profile/Hong-Mo_Je/publication/3974309_Pattern_classification_using_support_vector_machine_ensemble/links/09e415091bdc559051000000.pdf). | ||

+ | * '''Base algorithm:''' Described in the problem statement | ||

+ | * '''Solution:''' a modification of the basic algorithm, or simply the Basic algorithm itself. The main thing is to compare with other methods and draw conclusions, in particular, about the relationship between the presence of an improvement in the quality and diversity of sets of reference objects built by different SVMs. | ||

+ | * '''Novelty:''' It is known (for example, from Konstantin Vyacheslavovich's lectures) that it is not possible to build short compositions from strong classifiers (for example, SVM) using boosting (although they still try (see literature)). Therefore, it is proposed to build a nonlinear combination instead of a linear one. It is assumed that such a composition can give an increase in quality compared to a single SVM. | ||

+ | * '''consultant''': Alexander Aduenko | ||

+ | |||

+ | ===2. 2016=== | ||

+ | * '''Title:''' Temporal theme model of the press release collection. | ||

+ | * '''Problem:''' Development of methods for analyzing the thematic structure of a large text collection and its dynamics over time. The problem is the assessment of the quality of the constructed structure. It is required to implement the criteria of stability and completeness of the temporal thematic model using manual selection of the found topics according to their interpretability, difference and eventfulness. | ||

+ | * '''Data:''' A collection of press releases from the foreign ministries of a number of countries over 10 years, in English. | ||

+ | * '''References:''' | ||

+ | *# Doikov N.V. [[Media:2015_417_DoykovNV.pdf|Adaptive regularization of probabilistic topic models]]. VKR bachelor, VMK MSU. 2015. | ||

+ | * '''Base algorithm:''' Blay's classic LDA with post-hoc time analysis. | ||

+ | * '''Solution:''' Implementation of an additively regularized topic model using the [[BigARTM]] library. Building a series of thematic models. Evaluation of their interpretability, stability and completeness. | ||

+ | * '''Novelty:''' Criteria for sustainability and completeness of thematic models are new. | ||

+ | * '''consultant''': Nikita Doikov, '''problem author''' Vorontsov K. V. | ||

+ | |||

+ | ===3. 2016=== | ||

+ | * '''Title:''' Coordination of logical and linear classification models in the information analysis of electrocardiosignals. | ||

+ | * '''Problem:''' There are logical classifiers based on the identification of diagnostic standards for each disease and built by the Expert in semi-manual mode. For these classifiers, estimates of disease activities are determined, which have been used in the diagnostic system for many years and satisfy physician users. We build linear classifiers that are trained completely automatically and are ahead of logical classifiers in terms of classification quality. However, a direct transfer of the activity estimation technique to linear classifiers turned out to be impossible. It is required to build a linear activity model, setting it to reproduce the known activity estimates of the logical classifier. | ||

+ | * '''Data:''' A selection of more than 10 thousand electrocardiograms with diagnoses for 32 diseases. | ||

+ | * '''References:''' will issue :) | ||

+ | * '''Base algorithm:''' Linear classifier. | ||

+ | * '''Solution:''' Methods of linear regression, linear classification, feature selection. | ||

+ | * '''Novelty:''' The problem of matching two models of different nature can be considered as learning with privileged information - a promising direction proposed by the machine learning classic VN Vapnik several years ago. | ||

+ | * '''consultant''': Vlada Tselykh, '''problem author''' Vorontsov K. V. | ||

+ | |||

+ | ===4. 2016=== | ||

+ | * '''Title:''' Thematic classification model for diagnosing diseases by electrocardiogram. | ||

+ | * '''Problem:''' [[Technology of information analysis of electrocardiosignals]] according to V.M.Uspensky is based on ECG conversion into a character string and selection of informative sets of words - diagnostic standards for each disease. The linear classifier builds one diagnostic standard for each disease. The Screenfax screening diagnostic system now uses four standards for each disease, built in a semi-manual mode. It is required to fully automate the process of constructing diagnostic standards and to determine their optimal number for each disease. To do this, it is supposed to finalize the thematic classification model of S. Tsyganova, to perform a new implementation under [[BigARTM]], to expand computational experiments, to improve the quality of classification. | ||

+ | * '''Data:''' A selection of more than 10 thousand electrocardiograms with diagnoses for 32 diseases. | ||

+ | * '''References:''' will issue :) | ||

+ | * '''Base algorithm:''' Classification models by V.Tselykh, thematic model by S.Tsyganova. | ||

+ | * '''Solution:''' Topic model implemented using the [[BigARTM]] library. | ||

+ | * '''Novelty:''' Topic models have not previously been used to classify sampled biomedical signals. | ||

+ | * '''consultant''': Svetlana Tsyganova, '''problem author''' Vorontsov K. V. | ||

+ | |||

+ | ===5. 2016=== | ||

+ | * '''Title:''' Thematic models of distributive semantics for highlighting ethno-relevant topics in social networks. | ||

+ | * '''Problem:''' Thematic modeling of social media text collections faces the problem of ultra-short documents. It is not always clear where to draw the boundaries between documents (possible options: a single post, a user's wall, all posts by a given user, all posts for a given day in a given region, and so on). Topic models give interpretable vector representations of words and documents, but their quality depends on the distribution of document lengths. The word2vec model is independent of document lengths, since it takes into account only the local contexts of words, but the coordinates of vector representations do not allow thematic interpretation. The objective of the project is to build a hybrid model that combines the advantages and is free from the disadvantages of both models. | ||

+ | * '''Data:''' Collections of social networks LJ and VK. | ||

+ | * '''References:''' will issue :) | ||

+ | * '''Base algorithm:''' Topic models previously built on this data. | ||

+ | * '''Solution:''' Implementation of a distributive semantics regularizer similar to the vord2vec language model in the [[BigARTM]] library. | ||

+ | * '''Novelty:''' So far, there are no language models in the literature that combine the main advantages of probabilistic topic models and the word2vec model. | ||

+ | * '''consultant''': Anna Potapenko, on technical issues Murat Apishev, '''problem author''' Vorontsov K. V. | ||

+ | |||

+ | ===7. 2016=== | ||

+ | * '''Title:''' Determining the position of proteins using an electronic map | ||

+ | * '''Problem:''' informally --- there are sets of experimentally determined maps of the location of proteins in complexes, some of them are known in high resolution, it is necessary to restore the entire map in high resolution; formally --- there are matrices and energy vectors corresponding to each map of the protein complex, it is necessary to determine which set of proteins minimizes the quadratic form formed by the matrix and vector. | ||

+ | * '''Data:''' experimental data from the site http://www.emdatabank.org/ will be converted into matrices into energy vectors. Understanding the biophysical nature is not necessary. | ||

+ | * '''References:''' articles on methods for solving quadratic programming problems and various relaxations | ||

+ | * '''Base algorithm:''' quadratic programming methods with various relaxations | ||

+ | * '''Solution:''' minimizing the total energy of the protein complex | ||

+ | * '''Novelty:''' the application of quadratic programming methods and the study of their accuracy in The problems of restoring electronic maps | ||

+ | * '''consultant''': Alexander Katrutsa, problem author: Sergei Grudinin. | ||

+ | * '''Desirable skills''': understanding and interest in optimization methods, working with CVX package | ||

+ | |||

+ | ===8. 2016=== | ||

+ | * '''Title:''' Classification of Physical Activity: Investigation of Parameter Space Variation in Retraining and Modification of Deep Learning Models | ||

+ | * '''Problem:''' Given a classification model for a sample of time segments recorded from a mobile phone's accelerometer. The model is a multilayer neural network. It is required 1) to investigate the variance and covariance matrix of the neural network parameters under different optimization schedules (i.e., under dif |