Deadline-Aware Offloading for High-Throughput Accelerators

2021 IEEE International Symposium on High-Performance Computer Architecture (HPCA)(2021)

引用 12|浏览29
暂无评分
摘要
Contemporary GPUs are widely used for throughput-oriented data-parallel workloads and increasingly are being considered for latency-sensitive applications in datacenters. Examples include recurrent neural network (RNN) inference, network packet processing, and intelligent personal assistants. These data parallel applications have both high throughput demands and real-time deadlines (40μs-7ms). Mor...
更多
查看译文
关键词
Recurrent neural networks,Scheduling algorithms,Instruction sets,Graphics processing units,Throughput,Real-time systems,Kernel
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要