Skip to content

4bit ONNX models support #861

@Mihaiii

Description

@Mihaiii

Feature request

It look like ONNX now supports 4bit: https://onnx.ai/onnx/technical/int4.html

It would be nice if we could use 4bit models with transformers.js .

Motivation

Make models for semantic-autocomplete that are even smaller when it comes to disk size

Your contribution

Maybe, but for the moment is not even clear to me how to convert models to 4bit. I think ONNX needs to add more docs.

Metadata

Metadata

Assignees

No one assigned

    Labels

    enhancementNew feature or request

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions