Model Depot - ONNX
Collection
Leading Models packaged in ONNX format optimized for use with AI PCs
•
20 items
•
Updated
gemma-2b-it-onnx is an ONNX int4 quantized version of Google's Gemma-2B with Instruct Training (IT), providing a very fast, very small inference implementation, optimized for AI PCs using Intel GPU, CPU and NPU.
gemma-2b-it is a leading open source foundation model from Google.
Base model
google/gemma-2b-it