|速度超快!字节跳动开源序列推理引擎LightSeq( 五 )
传送门:
GitHub项目地址:
https://github.com/bytedance/lightseq
[1] Vaswani, Ashish, et al. ''Attention is all you need.'' Advances in neural information processing systems. 2017.
[2] Devlin, Jacob, et al. ''Bert: Pre-training of deep bidirectional transformers for language understanding.'' arXiv preprint arXiv:1810.04805 (2018).
[3] Brown, Tom B., et al. ''Language models are few-shot learners.'' arXiv preprint arXiv:2005.14165 (2020).
[4] WMT2020, http://www.statmt.org/wmt20/
[5] Li, Jiwei, Will Monroe, and Dan Jurafsky. ''A simple, fast diverse decoding algorithm for neural generation.'' arXiv preprint arXiv:1611.08562 (2016).
[6] TurboTransformers, https://github.com/Tencent/TurboTransformers
[7] FasterTransformer, https://github.com/NVIDIA/DeepLearningExamples/tree/master/FasterTransformer
[8] NVIDIA Triton Inference Server, https://github.com/triton-inference-server/server
[9] LightSeq proto, https://github.com/bytedance/lightseq/tree/master/proto
[10] LightSeq性能评测报告, https://github.com/bytedance/lightseq/blob/master/docs/performance.md
[11] LightSeq Layer Normalization, https://github.com/bytedance/lightseq/blob/master/kernels/transformerKernels.cu.cc#L269
[12] cuBLAS, https://docs.nvidia.com/cuda/cublas/index.html
【|速度超快!字节跳动开源序列推理引擎LightSeq】[13] GPT2,''Language Models are Unsupervised Multitask Learners''
推荐阅读
- 砍柴网|苹果 M1 没想象快,启动 Office 要多花 20 秒 苹果承认:速度会有点慢
- 雷帝触网|字节跳动今年广告流水收入或超1700亿 抖音贡献大
- 华为|下载速度每秒1TB,苹果、高通等联手预研6G:华为早已开始
- 趣投稿|OPPO 125W超快充电器或将于明年第一季度推出
- 科技|掌阅获字节跳动11亿元入股:尚未深化合作,建立品牌矩阵还需时日
- 教育|小台灯切入大教育 字节跳动全面布局教育领域
- 晚点LatePost|晚点独家|字节将电商列为明年头号孵化项目,将进一步整合
- 科技密探频道|哈耶克给张一鸣这点启发,是字节走向规模化的基础之一
- |保障“双十一”海关有速度
- 新系统|macOS Big Sur正式版11月12日推送 响应速度快1.9倍