CVT, a Computer Vision Toolkit.
-
Updated
Aug 24, 2022 - C
CVT, a Computer Vision Toolkit.
Winner solution of mobile AI (CVPRW 2021).
A header-only neural network library for microcontrollers, with partial bare-metal & native-os support.
FrostNet: Towards Quantization-Aware Network Architecture Search
Quantization Aware Training
ATtiny85 arduino example, running an RNN MNIST model via the (internal) 512-Byte EEPROM with ~95% accuracy
将端上模型部署过程中,常见的问题以及解决办法记录并汇总,希望能给其他人带来一点帮助。
VB.NET api wrapper for llm-inference chatllm.cpp
C# api wrapper for llm-inference chatllm.cpp
Garuda: CVXIF coprocessor optimizing batch-1 attention microkernels with 7.5-9× lower p99 latency. RISC-V INT8 MAC accelerator for transformer inference.
Generating tensorrt model using onnx
Corrects your grammar in 5 languages directly in your browser. Powered by an open-source AI model.
A fork of convert_to_quant that adds QuIP quantization for INT‑8 models.
TinyML project. This system monitors your room or surrounding with an onboard microphone of Arduino nano BLE sense. Still Under Developement
Python ML for training a custom on-device cry model (knowledge-distilled from YAMNet, INT8, deployed on ESP32-S3)
CPU face-embedding engine: 13 ms/face ArcFace INT8, 99.65% LFW 10-fold (beats FP32), 96 KB binary, 2.4x faster than ONNX Runtime. C99 + AVX-VNNI.
gemma-2-2b-it int8 cpu inference in one file of pure C#
g023's TurboXInf 🚀: 2x+ faster inference for Qwen3-1.77B or Qwen3.5-2B on RTX 3060! Custom Triton INT8 GEMV kernels halve memory traffic by fusing dequantization, paired with torch.compile. Hits 113 tok/s (vs 56.4 baseline) with no quality loss with INT8 even better results for INT4. MIT License.
Edge-deployable keyword spotter: INT8-quantized DS-CNN on Google Speech Commands, exported to ONNX, with fp32 vs INT8 benchmarks, a live mic demo, and a C++ inference harness.
Add a description, image, and links to the int8-quantization topic page so that developers can more easily learn about it.
To associate your repository with the int8-quantization topic, visit your repo's landing page and select "manage topics."