Skip to content

Commit

Permalink
Add quantized qwen2-0.5b (#490)
Browse files Browse the repository at this point in the history
Add quantized(q4f16) qwen2-0.5b to the list of supported models.
[PR](mlc-ai/binary-mlc-llm-libs#128) must be
merged before merging this.
  • Loading branch information
bil-ash authored Jun 27, 2024
1 parent b7537a6 commit 1da0f76
Showing 1 changed file with 13 additions and 0 deletions.
13 changes: 13 additions & 0 deletions src/config.ts
Original file line number Diff line number Diff line change
Expand Up @@ -601,6 +601,19 @@ export const prebuiltAppConfig: AppConfig = {
},
},
// Qwen-2
{
model: "https://huggingface.co/mlc-ai/Qwen2-0.5B-Instruct-q4f16_1-MLC",
model_id: "Qwen2-0.5B-Instruct-q4f16_1-MLC",
model_lib:
modelLibURLPrefix +
modelVersion +
"/Qwen2-0.5B-Instruct-q4f16_1-ctx4k_cs2k-webgpu.wasm",
low_resource_required: true,
vram_required_MB: 944.62,
overrides: {
context_window_size: 4096,
},
},
{
model: "https://huggingface.co/mlc-ai/Qwen2-0.5B-Instruct-q0f16-MLC",
model_id: "Qwen2-0.5B-Instruct-q0f16-MLC",
Expand Down

0 comments on commit 1da0f76

Please sign in to comment.