Quantized using Intel's AutoRound quantization tool.
auto-round-best --model facebook/opt-350m --scheme "w4a16"
- Downloads last month
- 53
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
🙋
Ask for provider support
Model tree for Emanresu/opt-350m-w4g128-AutoRound
Base model
facebook/opt-350m