Skip to content

Conversation

Copy link

Copilot AI commented Jan 11, 2026

GGML supports TQ2_0 quantization but the whisper.cpp quantize tool and model loader did not expose it.

Changes

ggml/include/ggml.h

  • Add GGML_FTYPE_MOSTLY_TQ2_0 = 26 enum value

examples/common-ggml.cpp

  • Map "tq2_0" string to GGML_FTYPE_MOSTLY_TQ2_0
  • Accept 't' prefix in ggml_parse_ftype() (alongside existing 'q')
  • Add GGML_FTYPE_MOSTLY_TQ2_0GGML_TYPE_TQ2_0 mapping
  • Move GGML_TYPE_TQ2_0 to supported quantization types

ggml/src/ggml.c

  • Add GGML_FTYPE_MOSTLY_TQ2_0 case to ggml_ftype_to_ggml_type()

Usage

./build/bin/quantize model-f32.bin model-tq2_0.bin tq2_0
# or
./build/bin/quantize model-f32.bin model-tq2_0.bin 26
Original prompt

ggml支持tq2_0,但quant应用不支持,请提交一下pr,支持tq2_0推理


✨ Let Copilot coding agent set things up for you — coding agent works faster and does higher quality work when set up for your repo.

Copilot AI and others added 2 commits January 11, 2026 12:26
Co-authored-by: lhpqaq <63844184+lhpqaq@users.noreply.github.com>
Co-authored-by: lhpqaq <63844184+lhpqaq@users.noreply.github.com>
Copilot AI changed the title [WIP] Add support for tq2_0 inference in quant application Add TQ2_0 quantization support to whisper.cpp Jan 11, 2026
Copilot AI requested a review from lhpqaq January 11, 2026 12:31
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants