pytorch-quantization
Techniques for model size reduction and inference acceleration using INT8 quantization, including Post-Training Quantization (PTQ) and Quantization Aware Training (QAT). (quantization, int8, qat, fbgemm, qnnpack, ptq, dequantize)
$ Installer
git clone https://github.com/majiayu000/claude-skill-registry /tmp/claude-skill-registry && cp -r /tmp/claude-skill-registry/skills/data/pytorch-quantization ~/.claude/skills/claude-skill-registry// tip: Run this command in your terminal to install the skill
Repository

majiayu000
Author
majiayu000/claude-skill-registry/skills/data/pytorch-quantization
0
Stars
0
Forks
Updated6d ago
Added6d ago