<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="ko">
	<id>https://wiki.mathnt.net/index.php?action=history&amp;feed=atom&amp;title=%EC%95%8C%ED%8C%8C%EA%B3%A0_%EC%A0%9C%EB%A1%9C</id>
	<title>알파고 제로 - 편집 역사</title>
	<link rel="self" type="application/atom+xml" href="https://wiki.mathnt.net/index.php?action=history&amp;feed=atom&amp;title=%EC%95%8C%ED%8C%8C%EA%B3%A0_%EC%A0%9C%EB%A1%9C"/>
	<link rel="alternate" type="text/html" href="https://wiki.mathnt.net/index.php?title=%EC%95%8C%ED%8C%8C%EA%B3%A0_%EC%A0%9C%EB%A1%9C&amp;action=history"/>
	<updated>2026-04-04T21:36:30Z</updated>
	<subtitle>이 문서의 편집 역사</subtitle>
	<generator>MediaWiki 1.35.0</generator>
	<entry>
		<id>https://wiki.mathnt.net/index.php?title=%EC%95%8C%ED%8C%8C%EA%B3%A0_%EC%A0%9C%EB%A1%9C&amp;diff=51394&amp;oldid=prev</id>
		<title>2021년 2월 17일 (수) 08:27에 Pythagoras0님의 편집</title>
		<link rel="alternate" type="text/html" href="https://wiki.mathnt.net/index.php?title=%EC%95%8C%ED%8C%8C%EA%B3%A0_%EC%A0%9C%EB%A1%9C&amp;diff=51394&amp;oldid=prev"/>
		<updated>2021-02-17T08:27:03Z</updated>

		<summary type="html">&lt;p&gt;&lt;/p&gt;
&lt;table class=&quot;diff diff-contentalign-left diff-editfont-monospace&quot; data-mw=&quot;interface&quot;&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;tr class=&quot;diff-title&quot; lang=&quot;ko&quot;&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #202122; text-align: center;&quot;&gt;← 이전 판&lt;/td&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #202122; text-align: center;&quot;&gt;2021년 2월 17일 (수) 08:27 판&lt;/td&gt;
				&lt;/tr&gt;&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot; id=&quot;mw-diff-left-l97&quot; &gt;97번째 줄:&lt;/td&gt;
&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;97번째 줄:&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&#039;diff-marker&#039;&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;  &amp;lt;references /&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;td class=&#039;diff-marker&#039;&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;  &amp;lt;references /&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&#039;diff-marker&#039;&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;td class=&#039;diff-marker&#039;&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&#039;diff-marker&#039;&gt;−&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;== 메타데이터 ==&lt;/div&gt;&lt;/td&gt;&lt;td class=&#039;diff-marker&#039;&gt;+&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;==메타데이터==&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&#039;diff-marker&#039;&gt;−&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt; &lt;/div&gt;&lt;/td&gt;&lt;td colspan=&quot;2&quot;&gt; &lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&#039;diff-marker&#039;&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;===위키데이터===&lt;/div&gt;&lt;/td&gt;&lt;td class=&#039;diff-marker&#039;&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;===위키데이터===&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&#039;diff-marker&#039;&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;* ID :  [https://www.wikidata.org/wiki/Q42259287 Q42259287]&lt;/div&gt;&lt;/td&gt;&lt;td class=&#039;diff-marker&#039;&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;* ID :  [https://www.wikidata.org/wiki/Q42259287 Q42259287]&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot;&gt; &lt;/td&gt;&lt;td class=&#039;diff-marker&#039;&gt;+&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;===Spacy 패턴 목록===&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot;&gt; &lt;/td&gt;&lt;td class=&#039;diff-marker&#039;&gt;+&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;* [{&amp;#039;LOWER&amp;#039;: &amp;#039;alphago&amp;#039;}, {&amp;#039;LEMMA&amp;#039;: &amp;#039;zero&amp;#039;}]&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot;&gt; &lt;/td&gt;&lt;td class=&#039;diff-marker&#039;&gt;+&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;* [{&amp;#039;LEMMA&amp;#039;: &amp;#039;AG0&amp;#039;}]&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot;&gt; &lt;/td&gt;&lt;td class=&#039;diff-marker&#039;&gt;+&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;* [{&amp;#039;LEMMA&amp;#039;: &amp;#039;AGZ&amp;#039;}]&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;/table&gt;</summary>
		<author><name>Pythagoras0</name></author>
	</entry>
	<entry>
		<id>https://wiki.mathnt.net/index.php?title=%EC%95%8C%ED%8C%8C%EA%B3%A0_%EC%A0%9C%EB%A1%9C&amp;diff=47191&amp;oldid=prev</id>
		<title>Pythagoras0: /* 메타데이터 */ 새 문단</title>
		<link rel="alternate" type="text/html" href="https://wiki.mathnt.net/index.php?title=%EC%95%8C%ED%8C%8C%EA%B3%A0_%EC%A0%9C%EB%A1%9C&amp;diff=47191&amp;oldid=prev"/>
		<updated>2020-12-26T12:26:46Z</updated>

		<summary type="html">&lt;p&gt;&lt;span dir=&quot;auto&quot;&gt;&lt;span class=&quot;autocomment&quot;&gt;메타데이터: &lt;/span&gt; 새 문단&lt;/span&gt;&lt;/p&gt;
&lt;table class=&quot;diff diff-contentalign-left diff-editfont-monospace&quot; data-mw=&quot;interface&quot;&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;tr class=&quot;diff-title&quot; lang=&quot;ko&quot;&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #202122; text-align: center;&quot;&gt;← 이전 판&lt;/td&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #202122; text-align: center;&quot;&gt;2020년 12월 26일 (토) 12:26 판&lt;/td&gt;
				&lt;/tr&gt;&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot; id=&quot;mw-diff-left-l96&quot; &gt;96번째 줄:&lt;/td&gt;
&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;96번째 줄:&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&#039;diff-marker&#039;&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;===소스===&lt;/div&gt;&lt;/td&gt;&lt;td class=&#039;diff-marker&#039;&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;===소스===&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&#039;diff-marker&#039;&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;  &amp;lt;references /&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;td class=&#039;diff-marker&#039;&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;  &amp;lt;references /&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot;&gt; &lt;/td&gt;&lt;td class=&#039;diff-marker&#039;&gt;+&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot;&gt; &lt;/td&gt;&lt;td class=&#039;diff-marker&#039;&gt;+&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;== 메타데이터 ==&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot;&gt; &lt;/td&gt;&lt;td class=&#039;diff-marker&#039;&gt;+&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot;&gt; &lt;/td&gt;&lt;td class=&#039;diff-marker&#039;&gt;+&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;===위키데이터===&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot;&gt; &lt;/td&gt;&lt;td class=&#039;diff-marker&#039;&gt;+&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;* ID :  [https://www.wikidata.org/wiki/Q42259287 Q42259287]&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;/table&gt;</summary>
		<author><name>Pythagoras0</name></author>
	</entry>
	<entry>
		<id>https://wiki.mathnt.net/index.php?title=%EC%95%8C%ED%8C%8C%EA%B3%A0_%EC%A0%9C%EB%A1%9C&amp;diff=46119&amp;oldid=prev</id>
		<title>Pythagoras0: /* 노트 */ 새 문단</title>
		<link rel="alternate" type="text/html" href="https://wiki.mathnt.net/index.php?title=%EC%95%8C%ED%8C%8C%EA%B3%A0_%EC%A0%9C%EB%A1%9C&amp;diff=46119&amp;oldid=prev"/>
		<updated>2020-12-21T06:05:45Z</updated>

		<summary type="html">&lt;p&gt;&lt;span dir=&quot;auto&quot;&gt;&lt;span class=&quot;autocomment&quot;&gt;노트: &lt;/span&gt; 새 문단&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;b&gt;새 문서&lt;/b&gt;&lt;/p&gt;&lt;div&gt;== 노트 ==&lt;br /&gt;
&lt;br /&gt;
===위키데이터===&lt;br /&gt;
* ID :  [https://www.wikidata.org/wiki/Q42259287 Q42259287]&lt;br /&gt;
===말뭉치===&lt;br /&gt;
# AlphaGo Zero was able to defeat its predecessor in only three days time with lesser processing power than AlphaGo.&amp;lt;ref name=&amp;quot;ref_6ff9d2c2&amp;quot;&amp;gt;[https://hub.packtpub.com/deepmind-alphago-zero-game-changer-for-ai-research/ Why DeepMind AlphaGo Zero is a game changer for AI research]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# All these facts beg the questions: what makes AlphaGo Zero so exceptional?&amp;lt;ref name=&amp;quot;ref_6ff9d2c2&amp;quot; /&amp;gt;&lt;br /&gt;
# Simply put, AlphaGo Zero is the strongest Go program in the world (with the exception of AlphaZero).&amp;lt;ref name=&amp;quot;ref_6ff9d2c2&amp;quot; /&amp;gt;&lt;br /&gt;
# But AlphaGo Zero didn’t use any human data whatsoever.&amp;lt;ref name=&amp;quot;ref_6ff9d2c2&amp;quot; /&amp;gt;&lt;br /&gt;
# In October 2017, the DeepMind team published details of a new Go-playing system, AlphaGo Zero, that studied no human games at all.&amp;lt;ref name=&amp;quot;ref_3b604e16&amp;quot;&amp;gt;[https://www.quantamagazine.org/why-alphazeros-artificial-intelligence-has-trouble-with-the-real-world-20180221/ Quanta Magazine]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# At the end of these scrimmages, AlphaGo Zero went head to head with the already superhuman version of AlphaGo that had beaten Lee Sedol.&amp;lt;ref name=&amp;quot;ref_3b604e16&amp;quot; /&amp;gt;&lt;br /&gt;
# But perhaps even more significant than these victories is how AlphaGo Zero became so dominant.&amp;lt;ref name=&amp;quot;ref_4c6a78a1&amp;quot;&amp;gt;[https://www.scientificamerican.com/article/ai-versus-ai-self-taught-alphago-zero-vanquishes-its-predecessor/ AI versus AI: Self-Taught AlphaGo Zero Vanquishes Its Predecessor]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# AlphaGo Zero even devised its own unconventional strategies.&amp;lt;ref name=&amp;quot;ref_4c6a78a1&amp;quot; /&amp;gt;&lt;br /&gt;
# This tutorial walks through a synchronous single-thread single-GPU (read malnourished) game-agnostic implementation of the recent AlphaGo Zero paper by DeepMind.&amp;lt;ref name=&amp;quot;ref_40b8489a&amp;quot;&amp;gt;[https://web.stanford.edu/~surag/posts/alphazero.html Simple Alpha Zero]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# The methods are fairly simple compared to previous papers by DeepMind, and AlphaGo Zero ends up beating AlphaGo (trained using data from expert games and beat the best human Go players) convincingly.&amp;lt;ref name=&amp;quot;ref_40b8489a&amp;quot; /&amp;gt;&lt;br /&gt;
# Recently, DeepMind published a preprint of Alpha Zero on arXiv that extends AlphaGo Zero methods to Chess and Shogi.&amp;lt;ref name=&amp;quot;ref_40b8489a&amp;quot; /&amp;gt;&lt;br /&gt;
# The aim of this post is to distil out the key ideas from the AlphaGo Zero paper and understand them concretely through code.&amp;lt;ref name=&amp;quot;ref_40b8489a&amp;quot; /&amp;gt;&lt;br /&gt;
# It is able to do this by using a novel form of reinforcement learning, in which AlphaGo Zero becomes its own teacher.&amp;lt;ref name=&amp;quot;ref_49f7785b&amp;quot;&amp;gt;[https://deepmind.com/blog/alphago-zero-learning-scratch/ AlphaGo Zero: Starting from scratch]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# This updated neural network is then recombined with the search algorithm to create a new, stronger version of AlphaGo Zero, and the process begins again.&amp;lt;ref name=&amp;quot;ref_49f7785b&amp;quot; /&amp;gt;&lt;br /&gt;
# Google later developed AlphaZero, a generalized version of AlphaGo Zero that could play chess and Shōgi in addition to Go.&amp;lt;ref name=&amp;quot;ref_2ba2291a&amp;quot;&amp;gt;[https://en.wikipedia.org/wiki/AlphaGo_Zero AlphaGo Zero]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# In December 2017, AlphaZero beat the 3-day version of AlphaGo Zero by winning 60 games to 40, and with 8 hours of training it outperformed AlphaGo Lee on an Elo scale.&amp;lt;ref name=&amp;quot;ref_2ba2291a&amp;quot; /&amp;gt;&lt;br /&gt;
# AlphaGo Zero&amp;#039;s neural network was trained using TensorFlow, with 64 GPU workers and 19 CPU parameter servers.&amp;lt;ref name=&amp;quot;ref_2ba2291a&amp;quot; /&amp;gt;&lt;br /&gt;
# AlphaGo Zero was widely regarded as a significant advance, even when compared with its groundbreaking predecessor, AlphaGo.&amp;lt;ref name=&amp;quot;ref_2ba2291a&amp;quot; /&amp;gt;&lt;br /&gt;
# The new AlphaGo Zero beat the previous version by 100 games to 0, and learned Go completely on its own.&amp;lt;ref name=&amp;quot;ref_96d48964&amp;quot;&amp;gt;[https://www.kdnuggets.com/2017/10/alphago-zero-biggest-ai-advance.html AlphaGo Zero: The Most Significant Research Advance in AI]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# In October, DeepMind published a paper describing a new version of AlphaGo, called AlphaGo Zero.&amp;lt;ref name=&amp;quot;ref_b16c50c1&amp;quot;&amp;gt;[https://www.yuzeh.com/data/agz-cost.html How much did AlphaGo Zero cost?]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# Not only that, AlphaGo Zero learned to play Go without any prior knowledge of the game (in other words, tabula rasa).&amp;lt;ref name=&amp;quot;ref_b16c50c1&amp;quot; /&amp;gt;&lt;br /&gt;
# AlphaGo Zero learns to play Go by simulating matches against itself in a procedure referred to as self-play.&amp;lt;ref name=&amp;quot;ref_b16c50c1&amp;quot; /&amp;gt;&lt;br /&gt;
# This is not to say that AlphaGo Zero is not an amazing achievement (it is!).&amp;lt;ref name=&amp;quot;ref_b16c50c1&amp;quot; /&amp;gt;&lt;br /&gt;
# Remarkably, during this self-teaching process AlphaGo Zero discovered many of the tricks and techniques that human Go players have developed over the past several thousand years.&amp;lt;ref name=&amp;quot;ref_5ed98bba&amp;quot;&amp;gt;[https://www.technologyreview.com/2017/10/18/148511/alphago-zero-shows-machines-can-become-superhuman-without-any-help/ AlphaGo Zero Shows Machines Can Become Superhuman Without Any Help]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# AlphaGo Zero simply played against itself, randomly at first.&amp;lt;ref name=&amp;quot;ref_5ed98bba&amp;quot; /&amp;gt;&lt;br /&gt;
# This is changed when DeepMind released AlphaGo Zero in late 2017.&amp;lt;ref name=&amp;quot;ref_eb307863&amp;quot;&amp;gt;[https://jonathan-hui.medium.com/alphago-zero-a-game-changer-14ef6e45eba5 AlphaGo Zero — a game changer. (How it works?)]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# We may expect AlphaGo Zero is more complicated and harder to train.&amp;lt;ref name=&amp;quot;ref_eb307863&amp;quot; /&amp;gt;&lt;br /&gt;
# In AlphaGo Zero, we train a single network f using self-play games.&amp;lt;ref name=&amp;quot;ref_eb307863&amp;quot; /&amp;gt;&lt;br /&gt;
# After three days of self-play using hugely powerful computer systems that let it play games at superhuman speeds, AlphaGo Zero was able to defeat its predecessor 100 games to nil.&amp;lt;ref name=&amp;quot;ref_c363e036&amp;quot;&amp;gt;[https://www.theverge.com/2019/11/27/20985260/ai-go-alphago-lee-se-dol-retired-deepmind-defeat Former Go champion beaten by DeepMind retires after declaring AI invincible]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# To address this challenge, we start by taking steps towards developing a formal understanding of AGZ.&amp;lt;ref name=&amp;quot;ref_921f15d8&amp;quot;&amp;gt;[https://openreview.net/forum?id=rkxtl3C5YX Understanding &amp;amp; Generalizing AlphaGo Zero]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# They had to make a newer and better version called AlphaGo Zero.&amp;lt;ref name=&amp;quot;ref_89498c98&amp;quot;&amp;gt;[https://towardsdatascience.com/enjoy-the-go-with-alpha-go-e1172afdb322 Enjoy the GO with Alpha Go]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# AlphaGo Zero uses only reinforcement learning.&amp;lt;ref name=&amp;quot;ref_89498c98&amp;quot; /&amp;gt;&lt;br /&gt;
# AlphaGo Zero uses only 1 neural network.&amp;lt;ref name=&amp;quot;ref_89498c98&amp;quot; /&amp;gt;&lt;br /&gt;
# AlphaGo Zero has some very similar features to AlphaGo Lee, but its distinct differences are what makes the new version so dominant.&amp;lt;ref name=&amp;quot;ref_89498c98&amp;quot; /&amp;gt;&lt;br /&gt;
# Among their accomplishments, particular focus will be placed upon the recent success of AlphaGo Zero which made waves in the machine learning and artificial intelligence communities.&amp;lt;ref name=&amp;quot;ref_85f6aa5b&amp;quot;&amp;gt;[https://dl.acm.org/doi/10.1145/3206157.3206174 Overview on DeepMind and Its AlphaGo Zero AI]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# The result, AlphaGo Zero, detailed in a paper published in October, 2017, was so called because it had zero knowledge of Go beyond the rules.&amp;lt;ref name=&amp;quot;ref_cdc1512a&amp;quot;&amp;gt;[https://www.newyorker.com/science/elements/how-the-artificial-intelligence-program-alphazero-mastered-its-games How the Artificial Intelligence Program AlphaZero Mastered Its Games]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# When the AlphaGo Zero and AlphaZero papers were published, a small army of enthusiasts began describing the systems in blog posts and YouTube videos and building their own copycat versions.&amp;lt;ref name=&amp;quot;ref_cdc1512a&amp;quot; /&amp;gt;&lt;br /&gt;
# The distributed LeelaZero community has had their system play more than ten million games against itself—a little more than AlphaGo Zero.&amp;lt;ref name=&amp;quot;ref_cdc1512a&amp;quot; /&amp;gt;&lt;br /&gt;
# The latest updates present in AlphaGo Zero left researchers in awe.&amp;lt;ref name=&amp;quot;ref_a23e3470&amp;quot;&amp;gt;[https://www.axcelerate.com.au/post/alphago-master-vs-alphago-zero-the-power-of-reinforcement-learning AlphaGo Master vs AlphaGo Zero - The Power of Reinforcement Learning]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# With AlphaGo Zero, DeepMind pushed RL’s independence from data further by starting with 100% randomness.&amp;lt;ref name=&amp;quot;ref_a23e3470&amp;quot; /&amp;gt;&lt;br /&gt;
# AlphaGo Zero is much less demanding than old Alphago, but running the same setup would still take 1700 GPU-years with ordinary hardware.&amp;lt;ref name=&amp;quot;ref_6d138ff8&amp;quot;&amp;gt;[https://rossum.ai/blog/building-our-own-version-of-alphago-zero/ Building Our Own Version of AlphaGo Zero]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# This is the setup where Nochi was the first AlphaGo Zero replication that achieved the level of the GNU Go baseline.&amp;lt;ref name=&amp;quot;ref_6d138ff8&amp;quot; /&amp;gt;&lt;br /&gt;
# Several other efforts to replicate the success of AlphaGo Zero are now underway – e.g. Leela Zero and Odin Zero.&amp;lt;ref name=&amp;quot;ref_6d138ff8&amp;quot; /&amp;gt;&lt;br /&gt;
# There were many advances in Deep Learning and AI in 2017, but few generated as much publicity and interest as DeepMind’s AlphaGo Zero.&amp;lt;ref name=&amp;quot;ref_8fcef0d5&amp;quot;&amp;gt;[https://hackernoon.com/the-3-tricks-that-made-alphago-zero-work-f3d47b6686ef dThe 3 Tricks That Made AlphaGo Zero Work]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# In this essay, I’ll try to give an intuitive idea of the techniques AlphaGo Zero used, what made them work, and what the implications for future AI research are.&amp;lt;ref name=&amp;quot;ref_8fcef0d5&amp;quot; /&amp;gt;&lt;br /&gt;
# This data, generated purely via lookahead and self-play, is what DeepMind used to train AlphaGo Zero.&amp;lt;ref name=&amp;quot;ref_8fcef0d5&amp;quot; /&amp;gt;&lt;br /&gt;
# AlphaGo Zero’s was its neural network architecture, a “two-headed” architecture.&amp;lt;ref name=&amp;quot;ref_8fcef0d5&amp;quot; /&amp;gt;&lt;br /&gt;
# A simplified, highly flexible, commented and (hopefully) easy to understand implementation of self-play based reinforcement learning based on the AlphaGo Zero paper (Silver et al).&amp;lt;ref name=&amp;quot;ref_743bb4be&amp;quot;&amp;gt;[https://github.com/suragnair/alpha-zero-general suragnair/alpha-zero-general: A clean implementation based on AlphaZero for any game in any framework + tutorial + Othello/Gobang/TicTacToe/Connect4 and more]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# If you follow artificial intelligence research, you probably saw last week&amp;#039;s Nature article on DeepMind&amp;#039;s AlphaGo Zero.&amp;lt;ref name=&amp;quot;ref_d7b1fb7e&amp;quot;&amp;gt;[https://blog.benchsci.com/alphago-zero-artificial-intelligence-drug-discovery What AlphaGo Zero Means for Artificial Intelligence Drug Discovery]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# This is the promise that AlphaGo Zero represents.&amp;lt;ref name=&amp;quot;ref_d7b1fb7e&amp;quot; /&amp;gt;&lt;br /&gt;
# While previous versions learned from thousands of human amateur and professional games, AlphaGo Zero taught itself to play by receiving only the goal (win), the rules, and feedback on its success.&amp;lt;ref name=&amp;quot;ref_d7b1fb7e&amp;quot; /&amp;gt;&lt;br /&gt;
# Even so, he notes, AlphaGo Zero is an important advance.&amp;lt;ref name=&amp;quot;ref_d7b1fb7e&amp;quot; /&amp;gt;&lt;br /&gt;
# AI research arm, DeepMind, announced in an article in the journal Nature today that it has built a final version of its prolific digital Go master: AlphaGo Zero.&amp;lt;ref name=&amp;quot;ref_4f3d7f61&amp;quot;&amp;gt;[https://qz.com/1105509/deepminds-new-alphago-zero-artificial-intelligence-is-ready-for-more-than-board-games/ DeepMind has a bigger plan for its newest Go-playing AI]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# The most important new modification is how AlphaGo Zero learned to master the game.&amp;lt;ref name=&amp;quot;ref_4f3d7f61&amp;quot; /&amp;gt;&lt;br /&gt;
# AlphaGo Zero doesn’t have hints from humans that previous systems had, like which pieces are whose or how to interpret the board.&amp;lt;ref name=&amp;quot;ref_4f3d7f61&amp;quot; /&amp;gt;&lt;br /&gt;
# AlphaGo Zero could beat the version of AlphaGo that faced Lee Sedol after training for just 36 hours and earned its 100-o score after 72 hours.&amp;lt;ref name=&amp;quot;ref_4f3d7f61&amp;quot; /&amp;gt;&lt;br /&gt;
# AlphaGo Zero used a Deep Reinforcement Learning approach and a tree-based search strategy.&amp;lt;ref name=&amp;quot;ref_6033617b&amp;quot;&amp;gt;[https://analyticsindiamag.com/deepminds-alphago-zero-really-scientific-breakthrough-step-towards-general-intelligence/ Is DeepMind’s AlphaGo Zero Really A Scientific Breakthrough?]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# Besides the human-like ingenuity displayed in learning, the reason why AlphaGo Zero is a big step forward is because the system got rid of the supervision and feature engineering.&amp;lt;ref name=&amp;quot;ref_6033617b&amp;quot; /&amp;gt;&lt;br /&gt;
# At NIPS 2017, Deep Reinforcement Learning was the most popular topic and DeepMind has delivered great results with AlphaGo Zero which plays at superhuman level.&amp;lt;ref name=&amp;quot;ref_6033617b&amp;quot; /&amp;gt;&lt;br /&gt;
# A new and much more powerful version of the program called AlphaGo Zero unveiled Wednesday is even more capable of surprises.&amp;lt;ref name=&amp;quot;ref_bc19db6f&amp;quot;&amp;gt;[https://www.wired.com/story/this-more-powerful-version-of-alphago-learns-on-its-own/ This More Powerful Version of AlphaGo Learns On Its Own]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# AlphaGo Zero showcases an approach to teaching machines new tricks that makes them less reliant on humans.&amp;lt;ref name=&amp;quot;ref_bc19db6f&amp;quot; /&amp;gt;&lt;br /&gt;
# DeepMind CEO Demis Hassabis said in a press briefing Monday that the guts of AlphaGo Zero should be adaptable to scientific problems such as drug discovery, or understanding protein folding.&amp;lt;ref name=&amp;quot;ref_bc19db6f&amp;quot; /&amp;gt;&lt;br /&gt;
# AlphaGo Zero is so-named because it doesn’t need human knowledge to get started, relying solely on that self-play mechanism.&amp;lt;ref name=&amp;quot;ref_bc19db6f&amp;quot; /&amp;gt;&lt;br /&gt;
# AlphaGo Zero was built on an improved reinforcement-learning system, and it trained itself from scratch without any input from human games.&amp;lt;ref name=&amp;quot;ref_b802908c&amp;quot;&amp;gt;[https://livebook.manning.com/book/deep-learning-and-the-game-of-go/chapter-14/ Chapter 14. AlphaGo Zero: Integrating tree search with reinforcement learning · Deep Learning and the Game of Go]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# AlphaGo Zero AI program just became the Go champion of the world without human data or guidance.&amp;lt;ref name=&amp;quot;ref_ed46dbf7&amp;quot;&amp;gt;[https://futureoflife.org/2017/10/18/deepminds-alphago-zero-becomes-go-champion-without-human-assistance/ DeepMind’s AlphaGo Zero Becomes Go Champion Without Human Input]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# AlphaGo Zero was not shown a single human game of Go from which to learn.&amp;lt;ref name=&amp;quot;ref_ed46dbf7&amp;quot; /&amp;gt;&lt;br /&gt;
# AlphaGo Zero learned entirely from playing against itself, with no prior knowledge of the game.&amp;lt;ref name=&amp;quot;ref_ed46dbf7&amp;quot; /&amp;gt;&lt;br /&gt;
# After just three days of playing against itself (4.9 million times), AlphaGo Zero beat AlphaGo by 100 games to 0.&amp;lt;ref name=&amp;quot;ref_ed46dbf7&amp;quot; /&amp;gt;&lt;br /&gt;
# DeepMind has shaken the world of Reinforcement Learning and Go with its creation AlphaGo, and later AlphaGo Zero.&amp;lt;ref name=&amp;quot;ref_08505231&amp;quot;&amp;gt;[https://dylandjian.github.io/alphago-zero/ AlphaGo Zero demystified]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# AlphaGo Zero pipeline is divided into three main components (just like the previous article on World Models), each in a different process that runs the code asynchronously.&amp;lt;ref name=&amp;quot;ref_08505231&amp;quot; /&amp;gt;&lt;br /&gt;
# This is also why AlphaGo Zero is sometimes called the two headed beast : a body, which is the feature extractor, and two heads : policy and value.&amp;lt;ref name=&amp;quot;ref_08505231&amp;quot; /&amp;gt;&lt;br /&gt;
# I want to thank my school’s AI association for letting me use the server to try to train this implementation of AlphaGo Zero.&amp;lt;ref name=&amp;quot;ref_08505231&amp;quot; /&amp;gt;&lt;br /&gt;
# Using less computing power and only three days of training time, AlphaGo Zero beat the original AlphaGo in a 100-game match by 100 to 0.&amp;lt;ref name=&amp;quot;ref_02da8cfa&amp;quot;&amp;gt;[https://qbi.uq.edu.au/blog/2017/10/google-alphago-zero-masters-game-three-days Google AlphaGo Zero masters the game in three days]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# By contrast, AlphaGo Zero never saw humans play.&amp;lt;ref name=&amp;quot;ref_02da8cfa&amp;quot; /&amp;gt;&lt;br /&gt;
# AlphaGo Zero achieved this feat by approaching the problem differently from the original AlphaGo.&amp;lt;ref name=&amp;quot;ref_02da8cfa&amp;quot; /&amp;gt;&lt;br /&gt;
# In the pure reinforcement learning approach of AlphaGo Zero, the only information available to learn policies and values was for it to predict who might ultimately win.&amp;lt;ref name=&amp;quot;ref_02da8cfa&amp;quot; /&amp;gt;&lt;br /&gt;
# That Go-playing virtual intelligence was called AlphaGo Zero, and it managed to rediscover over 3,000 years of human knowledge around the game in just 72 hours.&amp;lt;ref name=&amp;quot;ref_00ad3329&amp;quot;&amp;gt;[https://techcrunch.com/2017/11/02/deepmind-has-yet-to-find-out-how-smart-its-alphago-zero-ai-could-be/ DeepMind has yet to find out how smart its AlphaGo Zero AI could be – TechCrunch]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# Perhaps the most interesting thing about AlphaGo Zero, though, isn’t how fast it was able to do what it did, or with such efficacy, but also that it ultimately didn’t even achieve its full potential.&amp;lt;ref name=&amp;quot;ref_00ad3329&amp;quot; /&amp;gt;&lt;br /&gt;
# Named AlphaGo Zero, the AI program has been hailed as a major advance because it mastered the ancient Chinese board game from scratch, and with no human help beyond being told the rules.&amp;lt;ref name=&amp;quot;ref_683fd64d&amp;quot;&amp;gt;[https://www.theguardian.com/science/2017/oct/18/its-able-to-create-knowledge-itself-google-unveils-ai-learns-all-on-its-own &amp;#039;It&amp;#039;s able to create knowledge itself&amp;#039;: Google unveils AI that learns on its own]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# Twitter Pinterest David Silver describes how the Go playing AI program, AlphaGo Zero, discovers new knowledge from scratch.&amp;lt;ref name=&amp;quot;ref_683fd64d&amp;quot; /&amp;gt;&lt;br /&gt;
# When AlphaGo Zero plays a good move, it is more likely to be rewarded with a win.&amp;lt;ref name=&amp;quot;ref_683fd64d&amp;quot; /&amp;gt;&lt;br /&gt;
# Though far better than previous versions, AlphaGo Zero is a simpler program and mastered the game faster despite training on less data and running on a smaller computer.&amp;lt;ref name=&amp;quot;ref_683fd64d&amp;quot; /&amp;gt;&lt;br /&gt;
# In this way, AlphaGo Zero begins as a clean slate and learns from itself.&amp;lt;ref name=&amp;quot;ref_26a5d85a&amp;quot;&amp;gt;[https://www.technologynetworks.com/neuroscience/articles/deepminds-alphago-zero-learns-from-scratch-without-any-human-input-293412 New AI Learns From Scratch]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# Revealed in October 2017, AlphaGo Zero was the first computer program that learns to play simply by playing games against itself, starting from completely random play.&amp;lt;ref name=&amp;quot;ref_daefbd94&amp;quot;&amp;gt;[http://www.xinhuanet.com/english/2018-04/09/c_137097436.htm Feature: One man&amp;#039;s Go program looks to remake AlphaGo Zero - and beyond - Xinhua]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# Point being: AlphaGo Zero (which we’ll go ahead and shorten to AG0) is arguably the most impressive and definitely the most praised recent AI accomplishment.&amp;lt;ref name=&amp;quot;ref_6d4649f9&amp;quot;&amp;gt;[https://www.skynettoday.com/editorials/is-alphago-zero-overrated AlphaGo Zero Is Not A Sign of Imminent Human-Level AI]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# Roughly speaking, AG0 is just a Deep Neural Network that takes the current state of a Go board as input, and outputs a Go move.&amp;lt;ref name=&amp;quot;ref_6d4649f9&amp;quot; /&amp;gt;&lt;br /&gt;
# With those positive things having been said, some perspective: AG0 is not really a testament to the usefulness of such techniques for solving the hard problems of AI.&amp;lt;ref name=&amp;quot;ref_6d4649f9&amp;quot; /&amp;gt;&lt;br /&gt;
# AG0 is a definite example of Weak AI, also known as narrow AI.&amp;lt;ref name=&amp;quot;ref_6d4649f9&amp;quot; /&amp;gt;&lt;br /&gt;
# AlphaGo Zero scored 20-0.&amp;lt;ref name=&amp;quot;ref_b8b9e0b4&amp;quot;&amp;gt;[https://homepages.cwi.nl/~aeb/go/games/games/AlphaGo/ AlphaGo]&amp;lt;/ref&amp;gt;&lt;br /&gt;
# AlphaGo Zero scored 17-3.&amp;lt;ref name=&amp;quot;ref_b8b9e0b4&amp;quot; /&amp;gt;&lt;br /&gt;
===소스===&lt;br /&gt;
 &amp;lt;references /&amp;gt;&lt;/div&gt;</summary>
		<author><name>Pythagoras0</name></author>
	</entry>
</feed>