• A
  • A
  • A
  • ABC
  • ABC
  • ABC
  • А
  • А
  • А
  • А
  • А
Regular version of the site
Student
Title
Supervisor
Faculty
Educational Programme
Final Grade
Year of Graduation
Polina Kirichenko
Study of Bayesian Regularization of Neural Networks
9
2018
Deep neural networks have shown state-of-the-art performance in many machine learning tasks. However, such models with a large number of parameters are prone to overfitting, and this problem must be addressed for obtaining good generalization ability. Commonly used regularization techniques to tackle this issue are binary dropout or its fast approximation Gaussian dropout. In this work, we study a novel approach for neural network regularization via injecting noise on weight vector magnitude and direction instead of independently perturbing individual scalar weights. We consider several direction noise distributions and, further, propose a probabilistic model where variational inference can be applied for automatic hyperparameter tuning. Incorporating appropriate prior distributions in such a model can also potentially lead to structural sparsity and model compression.

Student Theses at HSE must be completed in accordance with the University Rules and regulations specified by each educational programme.

Summaries of all theses must be published and made freely available on the HSE website.

The full text of a thesis can be published in open access on the HSE website only if the authoring student (copyright holder) agrees, or, if the thesis was written by a team of students, if all the co-authors (copyright holders) agree. After a thesis is published on the HSE website, it obtains the status of an online publication.

Student theses are objects of copyright and their use is subject to limitations in accordance with the Russian Federation’s law on intellectual property.

In the event that a thesis is quoted or otherwise used, reference to the author’s name and the source of quotation is required.

Search all student theses