Int8 with TRT

baidu "caffe int8 "

 

https://blog.csdn.net/qq_33431368/article/details/85029041 利用量化工具caffe-int8-convert-tools实现caffemodel量化(double32->int8)

https://note.youdao.com/ynoteshare1/index.html?id=829ba6cabfde990e2832b048a4f492b3&type=note#/ (good)  基于tensorRT方案的INT8量化实现

http://on-demand.gputechconf.com/gtc/2017/presentation/s7310-8-bit-inference-with-tensorrt.pdf nvidia pdf   8-bit inference with tensorrt

 

https://cloud.tencent.com/developer/article/1435233 CNN模型 INT8 量化实现方式(一)

https://blog.csdn.net/u014644466/article/details/83278954 基于Caffe-Int8-Convert-Tools进行caffe模型转int8量化日常记录

https://github.com/lyk125/caffe-int8-convert-tools  Caffe-Int8-Convert-Tools

 

https://github.com/BUG1989/caffe-int8-convert-tools 

https://github.com/Tencent/ncnn/pull/749

 

 

"mobile inference"

https://blog.csdn.net/weixin_34038293/article/details/86752375

 

"TRT sorce code"

https://blog.csdn.net/bingbob/article/details/92802444

https://devblogs.nvidia.com/speed-up-inference-tensorrt/   How to Speed Up Deep Learning Inference Using TensorRT

 

 

posted on 2019-08-13 17:49  cdekelon  阅读(345)  评论(0)    收藏  举报

导航