Підтримка
www.wikidata.uk-ua.nina.az
U mashinnomu navchanni te mp navcha nnya angl learning rate abo koeficiye nt shvi dkosti navcha nnya ce giperparametr algoritmu optimizaciyi yakij zadaye rozmir kroku na kozhnij iteraciyi poshuku minimumu funkciyi vtrat Oskilki vin viznachaye te yakoyu miroyu nova otrimana informaciya zaminyuye poperednyu informaciyu vin obrazno predstavlyaye shvidkist z yakoyu model mashinnogo navchannya navchayetsya U literaturi z adaptivnogo keruvannya shvidkist navchannya zazvichaj nazivayut koeficiye ntom pidsi lennya angl gain Pri vstanovlenni tempu navchannya isnuye problema kompromisu mizh shvidkistyu zbizhnosti ta perestribuvannyam minimumu V toj chas yak en zazvichaj otrimuyetsya z gradiyenta funkciyi vtrat koeficiyent shvidkosti navchannya viznachaye naskilki velikij krok bude zrobleno v comu napryamku Zanadto visokij temp navchannya zmusit algoritm perestribnuti cherez minimum a navchannya z zanadto nizkim koeficiyentom abo zajme zanadto bagato chasu abo zastryagne u nebazhanomu lokalnomu minimumi Shob dosyagti shvidshoyi zbizhnosti zapobigti gojdannyu i zastryagannyu v nebazhanih lokalnih minimumah temp navchannya chasto zminyuyetsya pid chas navchannya abo vidpovidno do grafika tempu navchannya abo za dopomogoyu algoritmiv adaptivnogo tempu navchannya Koeficiyent shvidkosti navchannya ta jogo pidbir mozhe vidriznyatisya dlya riznih parametriv modeli i v takomu vipadku otrimuyemo diagonalnu matricyu yaku mozhna rozglyadati yak nablizhennya obernenoyi matrici Gese sho vikoristovuyetsya v metodi Nyutona Temp navchannya podiben do dovzhini kroku otrimuvanoyi netochnim linijnim poshukom u kvazinyutonovih metodah i vidpovidnih algoritmah optimizaciyi Minipaketna pidvibirka MPPV angl mini batch sub sampling MBSS pid chas linijnogo poshuku torkayetsya harakteristik funkcij vtrat za yakimi neobhidno viznachati temp navchannya Statichna MPPV utrimuye minipaket nezminnim uzdovzh napryamku poshuku sho prizvodit do plavnosti funkciyi vtrat uzdovzh napryamku poshuku Dinamichna MPPV utochnyuye minipaket na kozhnomu obchislenni funkciyi sho prizvodit do potochkovoyi rozrivnosti funkciyi vtrat uzdovzh napryamku poshuku Do vidiv linijnogo poshuku yaki adaptivno vstanovlyuyut temp navchannya dlya funkcij vtrat statichnoyi MPPV nalezhit parabolichno nablizhuvalnij linijnij PNL angl parabolic approximation line PAL poshuk Do vidiv linijnogo poshuku yaki adaptivno vstanovlyuyut temp navchannya dlya funkcij vtrat dinamichnoyi MPPV nalezhat imovirnisni vidi linijnogo poshuku viklyuchno gradiyentni vidi linijnogo poshuku angl gradient only line searches GOLS ta kvadratichni nablizhennya Grafik tempu navchannyaPochatkovij koeficiyent mozhna zalishiti za zamovchuvannyam abo vibrati za dopomogoyu nizki metodiv Grafik tempu navchannya zminyuye koeficiyent shvidkosti navchannya pid chas navchannya i najchastishe onovlyuyetsya mizh epohami iteraciyami V osnovnomu ce robitsya z dvoma parametrami zagasannyam ta impulsom Isnuye bagato riznih grafikiv shvidkosti navchannya ale najposhirenishimi ye chasovi pokrokovi ta eksponencijni Zagasannya angl decay ce giperparametr sho sluzhit dlya togo shob uniknuti gojdan situaciyi yaka mozhe viniknuti koli zanadto visokij postijnij temp navchannya zmushuye algoritm perestribuvati vpered i nazad cherez minimum Impuls angl momentum analogichnij kuli sho kotitsya z pagorba yaksho mi hochemo shob m yach opustivsya v najnizhchu tochku pagorba vidpovidaye najmenshij pohibci Impuls priskoryuye navchannya zbilshuyuchi koeficiyent shvidkosti koli gradiyent funkciyi vtrat ruhayetsya v odnomu napryamku protyagom trivalogo chasu a takozh unikaye lokalnih minimumiv perekochuyuchis cherez neveliki nerivnosti Impuls kontrolyuyetsya giperparametrom analogichnim masi m yacha yakij potribno pidibrati vruchnu zanadto visoka i m yach perekotitsya cherez minimumi yaki mi hochemo znajti zanadto nizka i vin ne dopomozhe optimizuvati poshuk Formula dlya viboru impulsu ye skladnishoyu nizh dlya zagasannya ale najchastishe vbudovana v biblioteki glibokogo navchannya taki yak Keras Chasovij angl time based grafik tempu navchannya zminyuye koeficiyent shvidkosti navchannya zalezhno vid tempu navchannya na poperednomu promizhku chasu Z urahuvannyam zagasannya formula tempu navchannya v nastupnomu promizhku chasu viglyadaye tak h n 1 h n 1 d n displaystyle eta n 1 frac eta n 1 dn de h displaystyle eta ce koeficiyent shvidkosti navchannya d displaystyle d ye parametrom zagasannya a n displaystyle n ce nomer kroku Krokovij angl step based grafik tempu navchannya zminyuye temp navchannya vidpovidno do deyakih poperedno viznachenih krokiv Formula z urahuvannyam zagasannya viznachayetsya yak h n h 0 d 1 n r displaystyle eta n eta 0 d left lfloor frac 1 n r right rfloor de h n displaystyle eta n temp navchannya na iteraciyi n displaystyle n h 0 displaystyle eta 0 pochatkovij temp navchannya d displaystyle d ce naskilki temp navchannya povinen zminyuvatisya na kozhnomu kroci a r displaystyle r vidpovidaye shvidkosti zmenshennya abo tomu yak chasto slid skidati shvidkist 10 vidpovidaye padinnyu kozhni 10 krokiv Funkciya floor displaystyle lfloor dots rfloor okruglyuye vsi znachennya menshi za 1 do 0 Eksponencijnij angl exponential grafik tempu navchannya shozhij na pokrokovij ale zamist krokiv vikoristovuyetsya eksponencijno spadna funkciya Formula eksponencijnogo grafika viglyadaye yak h n h 0 e d n displaystyle eta n eta 0 e dn de d displaystyle d ye parametrom zagasannya Adaptivnij temp navchannyaProblema z grafikami tempu navchannya polyagaye v tomu sho vsi voni zalezhat vid giperparametriv yaki potribno obirati vruchnu dlya kozhnogo konkretnogo seansu navchannya i voni mozhut silno vidriznyatisya zalezhno vid zadachi abo vikoristovuvanoyi modeli Dlya podolannya ciyeyi problemi isnuye bagato riznih adaptivnih algoritmiv gradiyentnogo spusku takih yak en Adadelta en ta en yaki zazvichaj vbudovuyutsya v biblioteki glibokogo navchannya taki yak Keras Div takozhGiperparametr mashinne navchannya Optimizaciya giperparametriv Stohastichnij gradiyentnij spusk Metodi zminnoyi metriki Perenavchannya Zvorotne poshirennya AvtoMN Obirannya modeli AvtonalashtuvannyaPrimitkiMurphy Kevin P 2012 Machine Learning A Probabilistic Perspective Cambridge MIT Press s 247 ISBN 978 0 262 01802 9 angl Delyon Bernard 2000 Stochastic Approximation with Decreasing Gain Convergence and Asymptotic Theory Unpublished Lecture Notes Universite de Rennes CiteSeerX 10 1 1 29 4428 angl Buduma Nikhil Locascio Nicholas 2017 Fundamentals of Deep Learning Designing Next Generation Machine Intelligence Algorithms O Reilly s 21 ISBN 978 1 4919 2558 4 angl Patterson Josh Gibson Adam 2017 Understanding Learning Rates Deep Learning A Practitioner s Approach O Reilly s 258 263 ISBN 978 1 4919 1425 0 angl Ruder Sebastian 2017 An Overview of Gradient Descent Optimization Algorithms arXiv 1609 04747 Nesterov Y 2004 Introductory Lectures on Convex Optimization A Basic Course Boston Kluwer s 25 ISBN 1 4020 7553 7 angl Dixon L C W 1972 The Choice of Step Length a Crucial Factor in the Performance of Variable Metric Algorithms Numerical Methods for Non linear Optimization London Academic Press s 149 170 ISBN 0 12 455650 7 angl An empirical study into finding optima in stochastic optimization of neural networks Information Sciences 2021 T 560 2 chervnya S 235 255 arXiv 1903 08552 angl Mutschler Maximus Zell Andreas 2019 Parabolic Approximation Line Search for DNNs arXiv 1903 11991 angl Mahsereci Maren Hennig Phillip 2016 Probabilistic Line Searches for Stochastic Optimization arXiv 1502 02846v4 angl Resolving learning rates adaptively by locating stochastic non negative associated gradient projection points using line searches Journal of Global Optimization 2021 T 79 2 chervnya S 111 152 arXiv 2001 05113 angl Chae Younghwan Wilke Daniel N 2019 Empirical study towards understanding line search approximations for training neural networks arXiv 1909 06893 angl Smith Leslie N 4 kvitnya 2017 Cyclical Learning Rates for Training Neural Networks arXiv 1506 01186 cs CV angl Murphy Kevin 2021 Probabilistic Machine Learning An Introduction MIT Press Procitovano 10 kvitnya 2021 angl Brownlee Jason 22 sichnya 2019 How to Configure the Learning Rate When Training Deep Learning Neural Networks Machine Learning Mastery Procitovano 4 sichnya 2021 angl LiteraturaGeron Aurelien 2017 Gradient Descent Hands On Machine Learning with Scikit Learn and TensorFlow O Reilly s 113 124 ISBN 978 1 4919 6229 9 angl Plagianakos V P Magoulas G D Vrahatis M N 2001 Learning Rate Adaptation in Stochastic Gradient Descent Advances in Convex Analysis and Global Optimization Kluwer s 433 444 ISBN 0 7923 6942 4 angl Posilannyade Freitas Nando 12 lyutogo 2015 Optimization Deep Learning Lecture 6 University of Oxford cherez YouTube angl
Топ