原推:@O42nl @kirillgroshkov Tesla does INT8 inference. Way more efficient than FP16, but took us a lot of effort to overcome quantization errors.
https://twitter.com/elonmusk/status/1630657812530339855
翻译英文优质信息和名人推特
原推:@O42nl @kirillgroshkov Tesla does INT8 inference. Way more efficient than FP16, but took us a lot of effort to overcome quantization errors.
https://twitter.com/elonmusk/status/1630657812530339855