由买买提看人间百态

boards

本页内容为未名空间相应帖子的节选和存档,一周内的贴子最多显示50字,超过一周显示500字 访问原贴
Go版 - AlphaGo is not the solution to AI
相关主题
阿法狗生怕人类还不够绝望从第四盘棋看狗狗的弱点
想赢AlphaGo的唯一机会是走已有定式写过程序的都知道
AlphaGo通俗的解释看大家讨论中比较少提这个 reinforcement learning
围棋软件Zen,Pachi作者对AlphaGo提的一些问题和看法Facebook’s AI tech mimics how humans learn
锵锵请的港大计算机系主任不懂啊AlphaGo之父:下个版本会让电脑从零学围棋
神经网络的结构决定了他的极限FB的出来踩AlphaGO了
金明完9段点评AlphaGo: 缺点明显Nature重磅封面:谷歌人工智能击败欧洲围棋冠军
原来阿尔法下的是5秒版的ZEN喆理围棋---关于Google人工智能围棋的访谈
相关话题的讨论汇总
话题: ai话题: go话题: alphago话题: solution
进入Go版参与讨论
1 (共1页)
a***m
发帖数: 5037
1
http://hunch.net/?p=3692542
Congratulations are in order for the folks at Google Deepmind who have
mastered Go.
However, some of the discussion around this seems like giddy overstatement.
Wired says Machines have conquered the last games and Slashdot says We know
now that we don’t need any big new breakthroughs to get to true AI. The
truth is nowhere close.
For Go itself, it’s been well-known for a decade that Monte Carlo tree
search (i.e. valuation by assuming randomized playout) is unusually
effective in Go. Given this, it’s unclear that the AlphaGo algorithm
extends to other board games where MCTS does not work so well. Maybe? It
will be interesting to see.
Delving into existing computer games, the Atari results (see figure 3) are
very fun but obviously unimpressive on about ¼ of the games. My
hypothesis for why is that their solution does only local (epsilon-greedy
style) exploration rather than global exploration so they can only learn
policies addressing either very short credit assignment problems or with
greedily accessible polices. Global exploration strategies are known to
result in exponentially more efficient strategies in general for
deterministic decision process(1993), Markov Decision Processes (1998), and
for MDPs without modeling (2006).
The reason these strategies are not used is because they are based on
tabular learning rather than function fitting. That’s why I shifted to
Contextual Bandit research after the 2006 paper. We’ve learned quite a bit
there, enough to start tackling a Contextual Deterministic Decision Process,
but that solution is still far from practical. Addressing global
exploration effectively is only one of the significant challenges between
what is well known now and what needs to be addressed for what I would
consider a real AI.
This is generally understood by people working on these techniques but seems
to be getting lost in translation to public news reports. That’s dangerous
because it leads to disappointment. The field will be better off without an
overpromise/bust cycle so I would encourage people to keep and inform a
balanced view of successes and their extent. Mastering Go is a great
accomplishment, but it is quite far from everything.
1 (共1页)
进入Go版参与讨论
相关主题
喆理围棋---关于Google人工智能围棋的访谈锵锵请的港大计算机系主任不懂啊
AlphaGo的开发团队信心满满啊神经网络的结构决定了他的极限
浏览了Deepmind 关于AlphaGo的技术说明文件金明完9段点评AlphaGo: 缺点明显
罗洗河让AlphaGo四子的说法没有什么错原来阿尔法下的是5秒版的ZEN
阿法狗生怕人类还不够绝望从第四盘棋看狗狗的弱点
想赢AlphaGo的唯一机会是走已有定式写过程序的都知道
AlphaGo通俗的解释看大家讨论中比较少提这个 reinforcement learning
围棋软件Zen,Pachi作者对AlphaGo提的一些问题和看法Facebook’s AI tech mimics how humans learn
相关话题的讨论汇总
话题: ai话题: go话题: alphago话题: solution