Model Card: vital-ai/watt-tool-70B-awq

Model Description

This model, vital-ai/watt-tool-70B-awq, is a quantized version of the base model watt-ai/watt-tool-70B. The quantization process was performed to reduce the model size and improve inference speed while maintaining high performance.

Base Model: watt-ai/watt-tool-70B

Quantization Method: 4-bit AWQ

Downloads last month
4
Safetensors
Model size
11B params
Tensor type
I32
BF16
F16
Inference Providers NEW
This model isn't deployed by any Inference Provider. 馃檵 Ask for provider support

Model tree for vital-ai/watt-tool-70B-awq

Quantized
(6)
this model