Machine Learning at Facebook: Understanding Inference at the Edge

2019 IEEE International Symposium on High Performance Computer Architecture (HPCA)(2019)

引用 508|浏览434
暂无评分
摘要
At Facebook, machine learning provides a wide range of capabilities that drive many aspects of user experience including ranking posts, content understanding, object detection and tracking for augmented and virtual reality, speech and text translations. While machine learning models are currently trained on customized data-center infrastructure, Facebook is working to bring machine learning inference to the edge. By doing so, user experience is improved with reduced latency (inference time) and becomes less dependent on network connectivity. Furthermore, this also enables many more applications of deep learning with important features only made available at the edge. This paper takes a data-driven approach to present the opportunities and design challenges faced by Facebook in order to enable machine learning inference locally on smartphones and other edge platforms.
更多
查看译文
关键词
Facebook,Smart phones,Performance evaluation,Graphics processing units,Optimization,Hardware,Machine learning
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要