Семинары
RUS  ENG    ЖУРНАЛЫ   ПЕРСОНАЛИИ   ОРГАНИЗАЦИИ   КОНФЕРЕНЦИИ   СЕМИНАРЫ   ВИДЕОТЕКА   ПАКЕТ AMSBIB  
Календарь
Поиск
Регистрация семинара

RSS
Ближайшие семинары




Коллоквиум Факультета компьютерных наук НИУ ВШЭ
4 марта 2025 г. 16:20–17:40, г. Москва, Покровский бульвар 11
 


Mirror-Prox Algorithm with Linear Convergence Rate and its Application for Dynamic Loss Scaling

Alexandr Beznosikov



Аннотация: During this talk, we extend the classical machine learning problem of the empirical risk minimization by adding weights to the loss function of each training object. We select these weights adaptively during the optimization process by modifying the original minimization problem to a min-max formulation. The maximization is performed over the above weights. To solve this saddle-point problem, we explore different variations of the Mirror-Prox algorithm: classical, optimistic, stochastic and with variance reduction. For all these methods, we obtain the convergence rates for the original min- max problem. Based on these algorithms we propose a novel deep learning (DL) optimizer ALSO – Adaptive Loss Scaling Optimizer. To justify our approach, we consider a wide range of DL problems, from Tabular DL to image generation tasks.

Website: https://vk.com/cshse?z=video-69306530_456239983
 
  Обратная связь:
 Пользовательское соглашение  Регистрация посетителей портала  Логотипы © Математический институт им. В. А. Стеклова РАН, 2025