<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="ko">
	<id>https://wiki.mathnt.net/index.php?action=history&amp;feed=atom&amp;title=Learning_rate</id>
	<title>Learning rate - 편집 역사</title>
	<link rel="self" type="application/atom+xml" href="https://wiki.mathnt.net/index.php?action=history&amp;feed=atom&amp;title=Learning_rate"/>
	<link rel="alternate" type="text/html" href="https://wiki.mathnt.net/index.php?title=Learning_rate&amp;action=history"/>
	<updated>2026-04-04T21:27:28Z</updated>
	<subtitle>이 문서의 편집 역사</subtitle>
	<generator>MediaWiki 1.35.0</generator>
	<entry>
		<id>https://wiki.mathnt.net/index.php?title=Learning_rate&amp;diff=51354&amp;oldid=prev</id>
		<title>2021년 2월 17일 (수) 08:21에 Pythagoras0님의 편집</title>
		<link rel="alternate" type="text/html" href="https://wiki.mathnt.net/index.php?title=Learning_rate&amp;diff=51354&amp;oldid=prev"/>
		<updated>2021-02-17T08:21:18Z</updated>

		<summary type="html">&lt;p&gt;&lt;/p&gt;
&lt;table class=&quot;diff diff-contentalign-left diff-editfont-monospace&quot; data-mw=&quot;interface&quot;&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;tr class=&quot;diff-title&quot; lang=&quot;ko&quot;&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #202122; text-align: center;&quot;&gt;← 이전 판&lt;/td&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #202122; text-align: center;&quot;&gt;2021년 2월 17일 (수) 08:21 판&lt;/td&gt;
				&lt;/tr&gt;&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot; id=&quot;mw-diff-left-l109&quot; &gt;109번째 줄:&lt;/td&gt;
&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;109번째 줄:&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&#039;diff-marker&#039;&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;  &amp;lt;references /&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;td class=&#039;diff-marker&#039;&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;  &amp;lt;references /&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&#039;diff-marker&#039;&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;td class=&#039;diff-marker&#039;&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&#039;diff-marker&#039;&gt;−&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;== 메타데이터 ==&lt;/div&gt;&lt;/td&gt;&lt;td class=&#039;diff-marker&#039;&gt;+&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;==메타데이터==&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&#039;diff-marker&#039;&gt;−&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt; &lt;/div&gt;&lt;/td&gt;&lt;td colspan=&quot;2&quot;&gt; &lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&#039;diff-marker&#039;&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;===위키데이터===&lt;/div&gt;&lt;/td&gt;&lt;td class=&#039;diff-marker&#039;&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;===위키데이터===&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&#039;diff-marker&#039;&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;* ID :  [https://www.wikidata.org/wiki/Q65121812 Q65121812]&lt;/div&gt;&lt;/td&gt;&lt;td class=&#039;diff-marker&#039;&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;* ID :  [https://www.wikidata.org/wiki/Q65121812 Q65121812]&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot;&gt; &lt;/td&gt;&lt;td class=&#039;diff-marker&#039;&gt;+&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;===Spacy 패턴 목록===&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot;&gt; &lt;/td&gt;&lt;td class=&#039;diff-marker&#039;&gt;+&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;* [{&amp;#039;LOWER&amp;#039;: &amp;#039;learning&amp;#039;}, {&amp;#039;LEMMA&amp;#039;: &amp;#039;rate&amp;#039;}]&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot;&gt; &lt;/td&gt;&lt;td class=&#039;diff-marker&#039;&gt;+&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;* [{&amp;#039;LOWER&amp;#039;: &amp;#039;step&amp;#039;}, {&amp;#039;LEMMA&amp;#039;: &amp;#039;size&amp;#039;}]&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot;&gt; &lt;/td&gt;&lt;td class=&#039;diff-marker&#039;&gt;+&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;* [{&amp;#039;LOWER&amp;#039;: &amp;#039;step&amp;#039;}, {&amp;#039;LEMMA&amp;#039;: &amp;#039;length&amp;#039;}]&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot;&gt; &lt;/td&gt;&lt;td class=&#039;diff-marker&#039;&gt;+&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;* [{&amp;#039;LEMMA&amp;#039;: &amp;#039;gain&amp;#039;}]&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;/table&gt;</summary>
		<author><name>Pythagoras0</name></author>
	</entry>
	<entry>
		<id>https://wiki.mathnt.net/index.php?title=Learning_rate&amp;diff=47151&amp;oldid=prev</id>
		<title>Pythagoras0: /* 메타데이터 */ 새 문단</title>
		<link rel="alternate" type="text/html" href="https://wiki.mathnt.net/index.php?title=Learning_rate&amp;diff=47151&amp;oldid=prev"/>
		<updated>2020-12-26T12:24:12Z</updated>

		<summary type="html">&lt;p&gt;&lt;span dir=&quot;auto&quot;&gt;&lt;span class=&quot;autocomment&quot;&gt;메타데이터: &lt;/span&gt; 새 문단&lt;/span&gt;&lt;/p&gt;
&lt;table class=&quot;diff diff-contentalign-left diff-editfont-monospace&quot; data-mw=&quot;interface&quot;&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;tr class=&quot;diff-title&quot; lang=&quot;ko&quot;&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #202122; text-align: center;&quot;&gt;← 이전 판&lt;/td&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #202122; text-align: center;&quot;&gt;2020년 12월 26일 (토) 12:24 판&lt;/td&gt;
				&lt;/tr&gt;&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot; id=&quot;mw-diff-left-l108&quot; &gt;108번째 줄:&lt;/td&gt;
&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;108번째 줄:&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&#039;diff-marker&#039;&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;===소스===&lt;/div&gt;&lt;/td&gt;&lt;td class=&#039;diff-marker&#039;&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;===소스===&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&#039;diff-marker&#039;&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;  &amp;lt;references /&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;td class=&#039;diff-marker&#039;&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;  &amp;lt;references /&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot;&gt; &lt;/td&gt;&lt;td class=&#039;diff-marker&#039;&gt;+&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot;&gt; &lt;/td&gt;&lt;td class=&#039;diff-marker&#039;&gt;+&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;== 메타데이터 ==&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot;&gt; &lt;/td&gt;&lt;td class=&#039;diff-marker&#039;&gt;+&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot;&gt; &lt;/td&gt;&lt;td class=&#039;diff-marker&#039;&gt;+&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;===위키데이터===&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot;&gt; &lt;/td&gt;&lt;td class=&#039;diff-marker&#039;&gt;+&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;* ID :  [https://www.wikidata.org/wiki/Q65121812 Q65121812]&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;/table&gt;</summary>
		<author><name>Pythagoras0</name></author>
	</entry>
	<entry>
		<id>https://wiki.mathnt.net/index.php?title=Learning_rate&amp;diff=46165&amp;oldid=prev</id>
		<title>Pythagoras0: /* 노트 */ 새 문단</title>
		<link rel="alternate" type="text/html" href="https://wiki.mathnt.net/index.php?title=Learning_rate&amp;diff=46165&amp;oldid=prev"/>
		<updated>2020-12-21T08:54:03Z</updated>

		<summary type="html">&lt;p&gt;&lt;span dir=&quot;auto&quot;&gt;&lt;span class=&quot;autocomment&quot;&gt;노트: &lt;/span&gt; 새 문단&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;b&gt;새 문서&lt;/b&gt;&lt;/p&gt;&lt;div&gt;== 노트 ==&lt;br /&gt;
&lt;br /&gt;
===위키데이터===&lt;br /&gt;
* ID :  [https://www.wikidata.org/wiki/Q65121812 Q65121812]&lt;br /&gt;
===말뭉치===&lt;br /&gt;
# The learning rate is a hyperparameter that controls how much we are adjusting the weights of our network with respect to the loss gradient.&amp;lt;ref name=&amp;quot;ref_f07907a5&amp;quot;&amp;gt;[https://blog.dataiku.com/the-learning-rate-finder-technique-how-reliable-is-it The Learning Rate Finder Technique: How Reliable Is It?]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# Indeed, one of the many challenges in training deep neural networks has historically been the selection of a good learning rate — that’s where the Learning Rate Range Test (LRRT) comes in.&amp;lt;ref name=&amp;quot;ref_f07907a5&amp;quot; /&amp;gt;&lt;br /&gt;
# The LRRT was proposed in 2015 and made popular by the fast.ai’s deep learning library as the Learning Rate Finder, also known as the LRFinder.&amp;lt;ref name=&amp;quot;ref_f07907a5&amp;quot; /&amp;gt;&lt;br /&gt;
# The LRRT consists of, at most, one epoch of training iterations, where the learning rate is increased at every mini-batch of data.&amp;lt;ref name=&amp;quot;ref_f07907a5&amp;quot; /&amp;gt;&lt;br /&gt;
# Under a weight-update algorithm, a low learning rate would make the network learning slowly, and a high learning rate would make the weights and error function diverge.&amp;lt;ref name=&amp;quot;ref_77548098&amp;quot;&amp;gt;[https://link.springer.com/chapter/10.1007/978-3-540-92137-0_76 Theoretical and Empirical Analysis of the Learning Rate and Momentum Factor in Neural Network Modeling for Stock Prediction]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# A backpropagation learning algorithm for feedforward neural networks withan adaptive learning rate is derived.&amp;lt;ref name=&amp;quot;ref_749364d0&amp;quot;&amp;gt;[https://link.springer.com/article/10.1023/A:1009686825582 Towards the Optimal Learning Rate for Backpropagation]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# Indeed,the derived optimal adaptive learning rate of a neural network trainedby backpropagation degenerates to the learning rate of the NLMS for a linear activation function of a neuron.&amp;lt;ref name=&amp;quot;ref_749364d0&amp;quot; /&amp;gt;&lt;br /&gt;
# By continuity, the optimal adaptive learning rate for neural networks imposes additional stabilisationeffects to the traditional backpropagation learning algorithm.&amp;lt;ref name=&amp;quot;ref_749364d0&amp;quot; /&amp;gt;&lt;br /&gt;
# The mathematical considerations that go into the derivation of backpropagation require that the learning rate be as small as possible.&amp;lt;ref name=&amp;quot;ref_99830a96&amp;quot;&amp;gt;[https://www.spiedigitallibrary.org/conference-proceedings-of-spie/1966/0000/Adaptive-learning-rate-for-increasing-learning-speed-in-backpropagation-networks/10.1117/12.152622.full Adaptive learning rate for increasing learning speed in backpropagation networks]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# Our simulation results suggest that the adaptive learning rate modification helps substantially speed up the convergence of backpropagation algorithm.&amp;lt;ref name=&amp;quot;ref_99830a96&amp;quot; /&amp;gt;&lt;br /&gt;
# Furthermore, it makes the initial choice of the learning rate fairly unimportant as our method allows the learning rate to change and settle at a reasonable value for the specific problem.&amp;lt;ref name=&amp;quot;ref_99830a96&amp;quot; /&amp;gt;&lt;br /&gt;
# Simulation results indicate that our heuristic modification matches the performance of backpropagation with the quasi-optimal learning rate.&amp;lt;ref name=&amp;quot;ref_99830a96&amp;quot; /&amp;gt;&lt;br /&gt;
# This can be counteracted by slowly decreasing the learning rate.&amp;lt;ref name=&amp;quot;ref_99fb1f7b&amp;quot;&amp;gt;[https://wiki.tum.de/display/lfdv/Adaptive+Learning+Rate+Method Adaptive Learning Rate Method]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# Learning rate is one such hyperparameter that defines the adjustment in the weights of our network with respect to the loss gradient.&amp;lt;ref name=&amp;quot;ref_76e0046a&amp;quot;&amp;gt;[https://heartbeat.fritz.ai/introduction-to-learning-rates-in-machine-learning-6ed685c16506 Introduction to Learning Rates in Machine Learning]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# With a desirable learning rate, he would quickly understand that black color is not an important feature of birds and would look for another feature.&amp;lt;ref name=&amp;quot;ref_76e0046a&amp;quot; /&amp;gt;&lt;br /&gt;
# But with a low learning rate, he would consider the yellow bird an outlier and would continue to believe that all birds are black.&amp;lt;ref name=&amp;quot;ref_76e0046a&amp;quot; /&amp;gt;&lt;br /&gt;
# It’s really important to achieve a desirable learning rate because both low and high learning rates result in wasted time and resources.&amp;lt;ref name=&amp;quot;ref_76e0046a&amp;quot; /&amp;gt;&lt;br /&gt;
# The learning rate warmup for Adam is a must-have trick for stable training in certain situations (or eps tuning).&amp;lt;ref name=&amp;quot;ref_77a92449&amp;quot;&amp;gt;[https://github.com/LiyuanLucasLiu/RAdam LiyuanLucasLiu/RAdam: On the Variance of the Adaptive Learning Rate and Beyond]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# The variance of the adaptive learning rate is simulated and plotted in Figure 1 (blue curve).&amp;lt;ref name=&amp;quot;ref_77a92449&amp;quot; /&amp;gt;&lt;br /&gt;
# A better idea is to select the dynamic learning rate which decreases over time because it allows the algorithm to swiftly identify the point.&amp;lt;ref name=&amp;quot;ref_d788f200&amp;quot;&amp;gt;[https://medium.com/analytics-vidhya/cost-function-learning-rate-and-gradient-descent-in-machine-learning-3dfd033e2d59 Cost Function, Learning rate, and Gradient Descent in Machine learning]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# It is common practice to decay the learning rate.&amp;lt;ref name=&amp;quot;ref_90c13e42&amp;quot;&amp;gt;[https://research.google/pubs/pub46644/ Don&amp;#039;t decay the learning rate, increase the batch size – Google Research]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# Operationally, stochastic inference iteratively subsamples from the data, analyzes the subsample, and updates parameters with a decreasing learning rate.&amp;lt;ref name=&amp;quot;ref_e8f2b32e&amp;quot;&amp;gt;[http://proceedings.mlr.press/v28/ranganath13.html An Adaptive Learning Rate for Stochastic Variational Inference]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# We solve this problem by developing an adaptive learning rate for stochastic inference.&amp;lt;ref name=&amp;quot;ref_e8f2b32e&amp;quot; /&amp;gt;&lt;br /&gt;
# Inference with the adaptive learning rate converges faster and to a better approximation than the best settings of hand-tuned rates.&amp;lt;ref name=&amp;quot;ref_e8f2b32e&amp;quot; /&amp;gt;&lt;br /&gt;
# Abstract: Intriguing empirical evidence exists that deep learning can work well with exotic schedules for varying the learning rate.&amp;lt;ref name=&amp;quot;ref_5a29a5ab&amp;quot;&amp;gt;[https://iclr.cc/virtual_2020/poster_rJg8TeSFDH.html ICLR: An Exponential Learning Rate Schedule for Deep Learning]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# • Training can be done using SGD with momentum and an exponentially in- creasing learning rate schedule, i.e., learning rate increases by some (1 + α) factor in every epoch for some α &amp;gt; 0.&amp;lt;ref name=&amp;quot;ref_5a29a5ab&amp;quot; /&amp;gt;&lt;br /&gt;
# The learning rate controls how quickly the model is adapted to the problem.&amp;lt;ref name=&amp;quot;ref_fd35f135&amp;quot;&amp;gt;[https://stackoverflow.com/questions/58266988/learning-rate-gradient-descent-difference Learning rate &amp;amp; gradient descent difference?]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# The challenge of training deep learning neural networks involves carefully selecting the learning rate.&amp;lt;ref name=&amp;quot;ref_fd35f135&amp;quot; /&amp;gt;&lt;br /&gt;
# It is otherefore often necessary to reduce the global learning rate µ when using a lot of momentum (m close to 1).&amp;lt;ref name=&amp;quot;ref_02a92368&amp;quot;&amp;gt;[https://cnl.salk.edu/~schraudo/teach/NNcourse/momrate.html Momentum and Learning Rate Adaptation]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# A useful batch method for adapting the global learning rate µ is the bold driver algorithm.&amp;lt;ref name=&amp;quot;ref_02a92368&amp;quot; /&amp;gt;&lt;br /&gt;
# In order to actually reach the minimum, and stay there, we must anneal (gradually lower) the global learning rate.&amp;lt;ref name=&amp;quot;ref_02a92368&amp;quot; /&amp;gt;&lt;br /&gt;
# At time t, we would like to change the learning rate (before changing the weight) such that the loss E(t+1) at the next time step is reduced.&amp;lt;ref name=&amp;quot;ref_02a92368&amp;quot; /&amp;gt;&lt;br /&gt;
# Learning rate is an important hyperparameter that controls how much we adjust the weights in the network according to the gradient.&amp;lt;ref name=&amp;quot;ref_7e86914f&amp;quot;&amp;gt;[https://analyticsindiamag.com/comprehensive-guide-to-learning-rate-algorithms-with-python-codes/ Comprehensive Guide To Learning Rate Algorithms (With Python Codes)]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# Here, η represents the learning rate.&amp;lt;ref name=&amp;quot;ref_7e86914f&amp;quot; /&amp;gt;&lt;br /&gt;
# Adaptive learning rate algorithm – Here, the optimizers help in changing the learning rate throughout the process of training.&amp;lt;ref name=&amp;quot;ref_7e86914f&amp;quot; /&amp;gt;&lt;br /&gt;
# One solution to this is fixing our learning rate large enough to escape the saddle.&amp;lt;ref name=&amp;quot;ref_7e86914f&amp;quot; /&amp;gt;&lt;br /&gt;
# # Decay Learning Rate, pass validation accuracy for tracking at every epoch print ( &amp;#039;Epoch {} completed&amp;#039; .&amp;lt;ref name=&amp;quot;ref_1e72376a&amp;quot;&amp;gt;[https://www.deeplearningwizard.com/deep_learning/boosting_models_pytorch/lr_scheduling/ Learning Rate Scheduling]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# Learning rate decay (lrDecay) is a \emph{de facto} technique for training modern neural networks.&amp;lt;ref name=&amp;quot;ref_6d161af3&amp;quot;&amp;gt;[https://openreview.net/forum?id=r1eOnh4YPB How Does Learning Rate Decay Help Modern Neural Networks?]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# It starts with a large learning rate and then decays it multiple times.&amp;lt;ref name=&amp;quot;ref_6d161af3&amp;quot; /&amp;gt;&lt;br /&gt;
# We provide another novel explanation: an initially large learning rate suppresses the network from memorizing noisy data while decaying the learning rate improves the learning of complex patterns.&amp;lt;ref name=&amp;quot;ref_6d161af3&amp;quot; /&amp;gt;&lt;br /&gt;
# In this post, I’m describing a simple and powerful way to find a reasonable learning rate that I learned from fast.ai Deep Learning course.&amp;lt;ref name=&amp;quot;ref_fe63c7de&amp;quot;&amp;gt;[https://www.kdnuggets.com/2017/11/estimating-optimal-learning-rate-deep-neural-network.html Estimating an Optimal Learning Rate For a Deep Neural Network]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# All of them let you set the learning rate.&amp;lt;ref name=&amp;quot;ref_fe63c7de&amp;quot; /&amp;gt;&lt;br /&gt;
# If the learning rate is high, then training may not converge or even diverge.&amp;lt;ref name=&amp;quot;ref_fe63c7de&amp;quot; /&amp;gt;&lt;br /&gt;
# There are multiple ways to select a good starting point for the learning rate.&amp;lt;ref name=&amp;quot;ref_fe63c7de&amp;quot; /&amp;gt;&lt;br /&gt;
# Learning rate is a hyper-parameter that controls the weights of our neural network with respect to the loss gradient.&amp;lt;ref name=&amp;quot;ref_b7f61b67&amp;quot;&amp;gt;[https://www.educative.io/edpresso/learning-rate-in-machine-learning Learning Rate in Machine learning]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# A desirable learning rate is low enough that the network converges to something useful, but high enough that it can be trained in a reasonable amount of time.&amp;lt;ref name=&amp;quot;ref_b7f61b67&amp;quot; /&amp;gt;&lt;br /&gt;
# Selecting a learning rate is an example of a &amp;quot;meta-problem&amp;quot; known as hyperparameter optimization.&amp;lt;ref name=&amp;quot;ref_781c625d&amp;quot;&amp;gt;[https://datascience.stackexchange.com/questions/410/choosing-a-learning-rate Choosing a learning rate]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# The best learning rate depends on the problem at hand, as well as on the architecture of the model being optimized, and even on the state of the model in the current optimization process!&amp;lt;ref name=&amp;quot;ref_781c625d&amp;quot; /&amp;gt;&lt;br /&gt;
# Simulated annealing is a technique for optimizing a model whereby one starts with a large learning rate and gradually reduces the learning rate as optimization progresses.&amp;lt;ref name=&amp;quot;ref_781c625d&amp;quot; /&amp;gt;&lt;br /&gt;
# This can be combined with early stopping to optimize the model with one learning rate as long as progress is being made, then switch to a smaller learning rate once progress appears to slow.&amp;lt;ref name=&amp;quot;ref_781c625d&amp;quot; /&amp;gt;&lt;br /&gt;
# The learning rate is the most important hyperparameter for tuning neural networks.&amp;lt;ref name=&amp;quot;ref_9c16b1d5&amp;quot;&amp;gt;[https://mlexplained.com/2018/01/29/learning-rate-tuning-in-deep-learning-a-practical-guide/ Learning Rate Tuning in Deep Learning: A Practical Guide]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# The learning rate represents how large a step you take in that direction.&amp;lt;ref name=&amp;quot;ref_9c16b1d5&amp;quot; /&amp;gt;&lt;br /&gt;
# When we look at the left diagram, we see that both steps are in the direction of the gradient; the only difference is the step size.&amp;lt;ref name=&amp;quot;ref_9c16b1d5&amp;quot; /&amp;gt;&lt;br /&gt;
# If the step size is too large, your parameters go back and forth between points with a large loss and easily overshoot the minima (the point where the loss function is at its lowest).&amp;lt;ref name=&amp;quot;ref_9c16b1d5&amp;quot; /&amp;gt;&lt;br /&gt;
# Step #1: We start by defining an upper and lower bound on our learning rate.&amp;lt;ref name=&amp;quot;ref_1bfd15cf&amp;quot;&amp;gt;[https://www.pyimagesearch.com/2019/08/05/keras-learning-rate-finder/ Keras Learning Rate Finder]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# At 1e-10 the learning rate will be too small for our network to learn, while at 1e+1 the learning rate will be too large and our model will overfit.&amp;lt;ref name=&amp;quot;ref_1bfd15cf&amp;quot; /&amp;gt;&lt;br /&gt;
# We start by defining an upper and lower bound on our learning rate.&amp;lt;ref name=&amp;quot;ref_1bfd15cf&amp;quot; /&amp;gt;&lt;br /&gt;
# After each batch update, we exponentially increase our learning rate.&amp;lt;ref name=&amp;quot;ref_1bfd15cf&amp;quot; /&amp;gt;&lt;br /&gt;
# With standard steepest descent, the learning rate is held constant throughout training.&amp;lt;ref name=&amp;quot;ref_495bbd62&amp;quot;&amp;gt;[https://www.mathworks.com/help/deeplearning/ref/traingda.html Gradient descent with adaptive learning rate backpropagation]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# If the learning rate is set too high, the algorithm can oscillate and become unstable.&amp;lt;ref name=&amp;quot;ref_495bbd62&amp;quot; /&amp;gt;&lt;br /&gt;
# If the learning rate is too small, the algorithm takes too long to converge.&amp;lt;ref name=&amp;quot;ref_495bbd62&amp;quot; /&amp;gt;&lt;br /&gt;
# You can improve the performance of the steepest descent algorithm if you allow the learning rate to change during the training process.&amp;lt;ref name=&amp;quot;ref_495bbd62&amp;quot; /&amp;gt;&lt;br /&gt;
# One of the simplest learning rate strategies is to have a fixed learning rate throughout the training process.&amp;lt;ref name=&amp;quot;ref_5b3f35e0&amp;quot;&amp;gt;[https://mxnet.apache.org/versions/1.7/api/python/docs/tutorials/packages/gluon/training/learning_rates/learning_rate_schedules.html Learning Rate Schedules — Apache MXNet documentation]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# Choosing a small learning rate allows the optimizer find good solutions, but this comes at the expense of limiting the initial speed of convergence.&amp;lt;ref name=&amp;quot;ref_5b3f35e0&amp;quot; /&amp;gt;&lt;br /&gt;
# Schedules define how the learning rate changes over time and are typically specified for each epoch or iteration (i.e. batch) of training.&amp;lt;ref name=&amp;quot;ref_5b3f35e0&amp;quot; /&amp;gt;&lt;br /&gt;
# All of these schedules define the learning rate for a given iteration, and it is expected that iterations start at 1 rather than 0.&amp;lt;ref name=&amp;quot;ref_5b3f35e0&amp;quot; /&amp;gt;&lt;br /&gt;
# When training a model, it is often recommended to lower the learning rate as the training progresses.&amp;lt;ref name=&amp;quot;ref_89df335a&amp;quot;&amp;gt;[https://www.tensorflow.org/api_docs/python/tf/compat/v1/train/exponential_decay tf.compat.v1.train.exponential_decay]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# This function applies an exponential decay function to a provided initial learning rate.&amp;lt;ref name=&amp;quot;ref_89df335a&amp;quot; /&amp;gt;&lt;br /&gt;
# It requires a global_step value to compute the decayed learning rate.&amp;lt;ref name=&amp;quot;ref_89df335a&amp;quot; /&amp;gt;&lt;br /&gt;
# The function returns the decayed learning rate.&amp;lt;ref name=&amp;quot;ref_89df335a&amp;quot; /&amp;gt;&lt;br /&gt;
# computing a more accurate learning rate.&amp;lt;ref name=&amp;quot;ref_6d3e8006&amp;quot;&amp;gt;[https://www.researchgate.net/publication/331208233_A_novel_adaptive_learning_rate_scheduler_for_deep_neural_networks (PDF) A novel adaptive learning rate scheduler for deep neural networks]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# Learning rate is used to scale the magnitude of parameter updates during gradient descent.&amp;lt;ref name=&amp;quot;ref_b5a56aaf&amp;quot;&amp;gt;[https://www.mygreatlearning.com/blog/understanding-learning-rate-in-machine-learning/ Understanding Learning Rate in Machine Learning]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# The choice of the value for learning rate can impact two things: 1) how fast the algorithm learns and 2) whether the cost function is minimized or not.&amp;lt;ref name=&amp;quot;ref_b5a56aaf&amp;quot; /&amp;gt;&lt;br /&gt;
# It can be seen that for an optimal value of the learning rate, the cost function value is minimized in a few iterations (smaller time).&amp;lt;ref name=&amp;quot;ref_b5a56aaf&amp;quot; /&amp;gt;&lt;br /&gt;
# If the learning rate used is lower than the optimal value, the number of iterations/epochs required to minimize the cost function is high (takes longer time).&amp;lt;ref name=&amp;quot;ref_b5a56aaf&amp;quot; /&amp;gt;&lt;br /&gt;
# Gradient descent algorithms multiply the gradient by a scalar known as the learning rate (also sometimes called step size) to determine the next point.&amp;lt;ref name=&amp;quot;ref_af5c77ad&amp;quot;&amp;gt;[https://developers.google.com/machine-learning/crash-course/reducing-loss/learning-rate Reducing Loss: Learning Rate]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# Most machine learning programmers spend a fair amount of time tuning the learning rate.&amp;lt;ref name=&amp;quot;ref_af5c77ad&amp;quot; /&amp;gt;&lt;br /&gt;
# The learning rate hyperparameter controls the rate or speed at which the model learns.&amp;lt;ref name=&amp;quot;ref_d9f725a8&amp;quot;&amp;gt;[https://machinelearningmastery.com/learning-rate-for-deep-learning-neural-networks/ How to Configure the Learning Rate When Training Deep Learning Neural Networks]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# Generally, a large learning rate allows the model to learn faster, at the cost of arriving on a sub-optimal final set of weights.&amp;lt;ref name=&amp;quot;ref_d9f725a8&amp;quot; /&amp;gt;&lt;br /&gt;
# When the learning rate is too large, gradient descent can inadvertently increase rather than decrease the training error.&amp;lt;ref name=&amp;quot;ref_d9f725a8&amp;quot; /&amp;gt;&lt;br /&gt;
# Therefore, we should not use a learning rate that is too large or too small.&amp;lt;ref name=&amp;quot;ref_d9f725a8&amp;quot; /&amp;gt;&lt;br /&gt;
# One of the key hyperparameters to set in order to train a neural network is the learning rate for gradient descent.&amp;lt;ref name=&amp;quot;ref_db5af4c3&amp;quot;&amp;gt;[https://www.jeremyjordan.me/nn-learning-rate/ Setting the learning rate of your neural network.]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# If your learning rate is set too low, training will progress very slowly as you are making very tiny updates to the weights in your network.&amp;lt;ref name=&amp;quot;ref_db5af4c3&amp;quot; /&amp;gt;&lt;br /&gt;
# However, if your learning rate is set too high, it can cause undesirable divergent behavior in your loss function.&amp;lt;ref name=&amp;quot;ref_db5af4c3&amp;quot; /&amp;gt;&lt;br /&gt;
# 3e-4 is the best learning rate for Adam, hands down.&amp;lt;ref name=&amp;quot;ref_db5af4c3&amp;quot; /&amp;gt;&lt;br /&gt;
# In setting a learning rate, there is a trade-off between the rate of convergence and overshooting.&amp;lt;ref name=&amp;quot;ref_5e947bec&amp;quot;&amp;gt;[https://en.wikipedia.org/wiki/Learning_rate Learning rate]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# While the descent direction is usually determined from the gradient of the loss function, the learning rate determines how big a step is taken in that direction.&amp;lt;ref name=&amp;quot;ref_5e947bec&amp;quot; /&amp;gt;&lt;br /&gt;
# A learning rate schedule changes the learning rate during learning and is most often changed between epochs/iterations.&amp;lt;ref name=&amp;quot;ref_5e947bec&amp;quot; /&amp;gt;&lt;br /&gt;
# Time-based learning schedules alter the learning rate depending on the learning rate of the previous time iteration.&amp;lt;ref name=&amp;quot;ref_5e947bec&amp;quot; /&amp;gt;&lt;br /&gt;
# This algorithm needs differentiable transfer function and the adaptive step size is recommended for Elman&amp;#039;s RNN because of the delays involved in the training scheme.&amp;lt;ref name=&amp;quot;ref_33cd6b8d&amp;quot;&amp;gt;[https://www.sciencedirect.com/topics/computer-science/adaptive-learning-rate Adaptive Learning Rate - an overview]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# Learning rate is a hyper-parameter that controls how much we are adjusting the weights of our network with respect the loss gradient.&amp;lt;ref name=&amp;quot;ref_e999e86c&amp;quot;&amp;gt;[https://towardsdatascience.com/understanding-learning-rates-and-how-it-improves-performance-in-deep-learning-d0d4059c1c10 Understanding Learning Rates and How It Improves Performance in Deep Learning]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# Furthermore, the learning rate affects how quickly our model can converge to a local minima (aka arrive at the best accuracy).&amp;lt;ref name=&amp;quot;ref_e999e86c&amp;quot; /&amp;gt;&lt;br /&gt;
# In practice, our learning rate should ideally be somewhere to the left to the lowest point of the graph (as demonstrated in below graph).&amp;lt;ref name=&amp;quot;ref_e999e86c&amp;quot; /&amp;gt;&lt;br /&gt;
# One only needs to type in the following command to start finding the most optimal learning rate to use before training a neural network.&amp;lt;ref name=&amp;quot;ref_e999e86c&amp;quot; /&amp;gt;&lt;br /&gt;
# this would be the gradient multiplied by the learning rate).&amp;lt;ref name=&amp;quot;ref_1f6a2613&amp;quot;&amp;gt;[https://cs231n.github.io/neural-networks-3/ CS231n Convolutional Neural Networks for Visual Recognition]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# If it is lower than this then the learning rate might be too low.&amp;lt;ref name=&amp;quot;ref_1f6a2613&amp;quot; /&amp;gt;&lt;br /&gt;
# Unconverged network, improperly set learning rate, very low weight regularization penalty.&amp;lt;ref name=&amp;quot;ref_1f6a2613&amp;quot; /&amp;gt;&lt;br /&gt;
# In training deep networks, it is usually helpful to anneal the learning rate over time.&amp;lt;ref name=&amp;quot;ref_1f6a2613&amp;quot; /&amp;gt;&lt;br /&gt;
# The learning rate is a hyperparameter that controls how much to change the model in response to the estimated error each time the model weights are updated.&amp;lt;ref name=&amp;quot;ref_0320309d&amp;quot;&amp;gt;[https://machinelearningmastery.com/understand-the-dynamics-of-learning-rate-on-deep-learning-neural-networks/ Understand the Impact of Learning Rate on Neural Network Performance]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# The learning rate may be the most important hyperparameter when configuring your neural network.&amp;lt;ref name=&amp;quot;ref_0320309d&amp;quot; /&amp;gt;&lt;br /&gt;
# The callbacks operate separately from the optimization algorithm, although they adjust the learning rate used by the optimization algorithm.&amp;lt;ref name=&amp;quot;ref_0320309d&amp;quot; /&amp;gt;&lt;br /&gt;
# Keras provides the ReduceLROnPlateau that will adjust the learning rate when a plateau in model performance is detected, e.g. no change for a given number of training epochs.&amp;lt;ref name=&amp;quot;ref_0320309d&amp;quot; /&amp;gt;&lt;br /&gt;
===소스===&lt;br /&gt;
 &amp;lt;references /&amp;gt;&lt;/div&gt;</summary>
		<author><name>Pythagoras0</name></author>
	</entry>
</feed>