DataScience 板


LINE

https://blog.piekniewski.info/2018/10/13/deep-learning-the-why-question/ There are many many deep learning models out there doing various things. Depending on the exact task they are solving, they may be constructed differently. Some will use convolution followed by pooling. Some will use several convolutional layers before there is any pooling layer. Some will use max-pooling. Some will use mean-pooling. Some will have a dropout added. Some will have a batch-norm layer here and there. Some will use sigmoid neurons, some will use half-recitfiers. Some will classify and therefore optimize for cross-entropy. Others will minimize mean-squared error. Some will use unpooling layers. Some will use deconvolutional layers. Some will use stochastic gradient descent with momentum. Some will use ADAM. Some will have RESNET layers, some will use Inception. The choices are plentiful (see e.g.here). Reading any of these particular papers, one is faced with a set of choices the authors had made, followed by the evaluation on the dataset of their choice. The discussion of choices typically refers strongly to papers where given techniques were first introduced, whereas the results section typically discusses in detail the previous state of the art. The shape of the architecture is often broken down into obvious and non obvious decisions. The obvious ones are dictated by the particular task the authors are trying to solve (e.g. when they have an autoencoding like task, they obviously use a form of an autoencoder). The non obvious choices would include questions similar to those: Why did they use 3x3 conv followed by 1x1 conv and only then pooling? Why did they only replaced the 3 middle layers with MobileNet layers (ridiculous name BTW)? Why did they slap batch-norm only in the middle two layers and not all of them? Why did they use max-pooling in the first two layers and no pooling whatsoever in the following three? Obvious stuff is not discussed because it is obvious, the non-obvious stuff is not discussed because ... let me get back to that in a moment. In my opinion discussing these questions separates a paper from something at least shallowly scientific from complete charlatanry, even if thecharlatanry appears to improve the results on the given dataset. The sad truth, that few even talk about, is that in the vast majority of cases the answers to the why questions are purely empirical: they tried a bunch of models and these worked best - it is called "hyperparameter tuning" (or meta-parameter tuning). What does that tell us? A few things, first the authors are completely ignoring the danger of multiple hypothesis testing and generally piss on any statistical foundations of their "research". Second, they probably have more GPU's accessible than they know what to do with (very often they case in big companies these days). Third, they just want to stamp their names on some new record breaking benchmark, that obviously will be broken two weeks later by somebody who takes their model and does some extra blind tweaking, utilizing even more GPU power. This is not science. This has more to do with people who build beefy PC's and submit their 3dMark results to hold a record for a few days. It is a craft, no doubt, but it is not science. The PC-builders don't make a pretense for this to be any science. The deep learning people do. They write what appears to be research papers, just to describe their GPU-rig and the result of their random meta-parameter search, with perhaps some tiny shreds of real scientific discussion. Benchmark results provide a nice cover, to claim that the paper is in some way "novel" and interesting, but truth to the mater is, they just overfitted that dataset some more. They might just as well memorize the entire dataset in their model and achieve 100% accuracy, who cares? (read my AI winter addendum post for some interesting literature on the subject). Similarly to the difference between chemistry andalchemy, the scientific discussion is about building a concept, a theory that will enable one to make accurate predictions. Something to guide their experimental actions. Science does not need to make gold out of lead every time, or in the case of machine learning, a real scientific paper in this field does not need to beat some current benchmark. A scientific paper does not even need to answer any questions, if it happens to ask some good ones. Now obviously there are exceptions, a small fraction of papers have interesting stuff in them. These are mostly the ones which try to show the deficits of deep learning and engage into a discussion as to why that might be the case. So next time you read a deep learning paper, try to contemplate these quiet and never explained choices the authors have made. You'll be shocked to see how many of those are hidden between the lines. ---------------------------- 这篇文章打到我心坎里 就算上了这麽多课,实作了好几十个model 也打了好几场Kaggle 甚至即将发布paper 对自己还是有很深的imposter syndrome 到底,为什麽这个kernel是3不是2、filter为什麽设64成绩就是比128好看 工作最多的时间除了 data preprocessing 就是弄到质疑人生的 hyperparameter tuning 怀疑自己到底是在搞学术研究,还是穷举各种可能性以结果论的资料炼金术 去年韩国人拿掉了几层BN就夺了NTIRE冠军 除了大家津津乐道地事後诸葛BN本来就不适合super resolution云云 难到这背後到底有什麽经得起严谨检验的基础吗? 这种自我质疑是来自我自己过深的自卑感还是其实没那麽罕见呢? ----- Sent from JPTT on my Asus ASUS_Z01KDA. --



※ 发信站: 批踢踢实业坊(ptt.cc), 来自: 61.231.154.103
※ 文章网址: https://webptt.com/cn.aspx?n=bbs/DataScience/M.1539515084.A.E47.html
1F:推 GCWShadow: 推这个想法 10/14 19:18
2F:→ GCWShadow: 现在这个领域就是各种乱兜model拼publication 10/14 19:18
3F:→ GCWShadow: 但是底子里是不是真的有东西就难说了 10/14 19:18
4F:→ Kazimir: 还好吧 他自己也提到化学 拉塞福也呛过化学是集邮活动啊 10/14 19:25
5F:→ Kazimir: 效果在那边 边往前边探讨罢了 化学生物医学什麽的哪个没 10/14 19:28
6F:→ Kazimir: 经历过这一段 10/14 19:29
7F:推 yiefaung: 目前无法解决 因为投资报酬率不成比例 大概要等出现某 10/14 21:21
8F:→ yiefaung: 种瓶颈或是热潮泡沫化 10/14 21:21
9F:推 yiefaung: 这也牵涉到水论文很多 还有很大一部分无法reproduce之 10/14 21:23
10F:→ yiefaung: 类的问题 总之有这个问题意识是好事… 10/14 21:23
11F:推 outrunner: 事实如此, 不用怀疑了 10/14 22:37
12F:推 germun: 研究不就这样 没有这些垃圾产出的经验 也找不到正确的路 10/14 23:22
13F:推 ap954212: 发现後再用数学证明很普遍吧 10/15 18:16
14F:→ baseguard: 以业界来说,能work就好,不会要求严谨证明 10/18 20:29
15F:→ Mchord: 业界要求比学界paper高多了,尤其是非纯吸金炒作要出产品 10/19 23:17
16F:→ Mchord: 的公司 10/19 23:17
17F:推 sma1033: 业界要求面向不同啦,业界诉求是1.稳定2.速度快3.可预期 10/20 10:57
18F:→ sma1033: 对於「不知道什麽时候会吐出坏掉东西」的Model实务上其实 10/20 10:58
19F:→ sma1033: 是非常难真的上线应用的,偏偏DL Model常常很容易出现无 10/20 10:58
20F:→ sma1033: 法预测的output,然後出错的时候又很容易不确定原因在哪 10/20 10:59
21F:→ sma1033: 是不是足够创新反而不是主要诉求的点,有人愿意付钱就好 10/20 11:00
22F:推 wheado: 我认为可能需要严谨的理论基础,不过也不一定对 10/23 19:24







like.gif 您可能会有兴趣的文章
icon.png[问题/行为] 猫晚上进房间会不会有憋尿问题
icon.pngRe: [闲聊] 选了错误的女孩成为魔法少女 XDDDDDDDDDD
icon.png[正妹] 瑞典 一张
icon.png[心得] EMS高领长版毛衣.墨小楼MC1002
icon.png[分享] 丹龙隔热纸GE55+33+22
icon.png[问题] 清洗洗衣机
icon.png[寻物] 窗台下的空间
icon.png[闲聊] 双极の女神1 木魔爵
icon.png[售车] 新竹 1997 march 1297cc 白色 四门
icon.png[讨论] 能从照片感受到摄影者心情吗
icon.png[狂贺] 贺贺贺贺 贺!岛村卯月!总选举NO.1
icon.png[难过] 羡慕白皮肤的女生
icon.png阅读文章
icon.png[黑特]
icon.png[问题] SBK S1安装於安全帽位置
icon.png[分享] 旧woo100绝版开箱!!
icon.pngRe: [无言] 关於小包卫生纸
icon.png[开箱] E5-2683V3 RX480Strix 快睿C1 简单测试
icon.png[心得] 苍の海贼龙 地狱 执行者16PT
icon.png[售车] 1999年Virage iO 1.8EXi
icon.png[心得] 挑战33 LV10 狮子座pt solo
icon.png[闲聊] 手把手教你不被桶之新手主购教学
icon.png[分享] Civic Type R 量产版官方照无预警流出
icon.png[售车] Golf 4 2.0 银色 自排
icon.png[出售] Graco提篮汽座(有底座)2000元诚可议
icon.png[问题] 请问补牙材质掉了还能再补吗?(台中半年内
icon.png[问题] 44th 单曲 生写竟然都给重复的啊啊!
icon.png[心得] 华南红卡/icash 核卡
icon.png[问题] 拔牙矫正这样正常吗
icon.png[赠送] 老莫高业 初业 102年版
icon.png[情报] 三大行动支付 本季掀战火
icon.png[宝宝] 博客来Amos水蜡笔5/1特价五折
icon.pngRe: [心得] 新鲜人一些面试分享
icon.png[心得] 苍の海贼龙 地狱 麒麟25PT
icon.pngRe: [闲聊] (君の名は。雷慎入) 君名二创漫画翻译
icon.pngRe: [闲聊] OGN中场影片:失踪人口局 (英文字幕)
icon.png[问题] 台湾大哥大4G讯号差
icon.png[出售] [全国]全新千寻侘草LED灯, 水草

请输入看板名称,例如:Tech_Job站内搜寻

TOP