Training a neural network to learn other dimensionality reduction removes data size restrictions in bioinformatics and provides a new route to exploring data representations

Alex Dexter, Thomas Sa, Steven Rt, Robinson Kn, Andy Taylor,Efstathios A. Elia,Chelsea Nikula, Campbell Ad,Yulia Panina, Najumudeen Ak,Teresa Murta, Bo Yan,Piotr Grabowski,Grégory Hamm,John M. Swales, Gilmore Is, Yuneva Mo, Goodwin Rj,Simon T. Barry,Owen J. Sansom,Zoltán Takáts,Josephine Bunch

bioRxiv (Cold Spring Harbor Laboratory)(2020)

引用 0|浏览1
暂无评分
摘要
Abstract High dimensionality omics and hyperspectral imaging datasets present difficult challenges for feature extraction and data mining due to huge numbers of features that cannot be simultaneously examined. The sample numbers and variables of these methods are constantly growing as new technologies are developed, and computational analysis needs to evolve to keep up with growing demand. Current state of the art algorithms can handle some routine datasets but struggle when datasets grow above a certain size. We present a training deep learning via neural networks on non-linear dimensionality reduction, in particular t-distributed stochastic neighbour embedding (t-SNE), to overcome prior limitations of these methods. One Sentence Summary Analysis of prohibitively large datasets by combining deep learning via neural networks with non-linear dimensionality reduction.
更多
查看译文
关键词
other dimensionality reduction,bioinformatics,neural network,data size restrictions,representations
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要