gemma-2-2b-it int8 cpu inference in one file of pure C#
csharp inference quantization gemma int8 inference-engine model-serving int8-inference int8-quantization cpu-inference llm llms llm-serving llm-inference gemma2 gemma2-2b-it
-
Updated
Jun 14, 2025 - C#