Exploiting Heavy Tails in Training Times of Multilayer Perceptrons. A Case Study with the UCI Thyroid Disease Database

Clinical Orthopaedics and Related Research(2007)

引用 25|浏览3
暂无评分
摘要
The random initialization of weights of a multilayer perceptron makes it possible to model its training process as a Las Vegas algorithm, i.e. a randomized algorithm which stops when some required training error is obtained, and whose execution time is a random variable. This model- ing is used to perform a case study on a well-known pattern recognition benchmark: the UCI Thyroid Disease Database. Empirical evidence is presented of the training time probability distribution exhibiting a heavy tail behavior, meaning a big probability mass of long executions. This fact is exploited to reduce the training time cost by applying two simple restart strategies. The first assumes full knowledge of the distribution yielding a 40% cut down in expected time with respect to the training without restarts. The second, assumes null knowledge, yielding a reduc- tion ranging from 9% to 23%.
更多
查看译文
关键词
heavy tail distri- bution,stochastic modeling,uci thyroid disease database.,multilayer perceptron,restart strategy,pattern recognition,probability distribution,random variable,stochastic model,empirical evidence,heavy tail,randomized algorithm
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要