second-state
second-state/MiniCPM-V-2_6-GGUF
No description available.
Model Documentation
header start -->
200823 -->
header end -->
openbmb/MiniCPM-V-2_6
LlamaEdge version: v0.14.17 and above
Prompt template
Prompt type:
Prompt string
The
Context size:
Run as LlamaEdge service
| Name | Quant method | Bits | Size | Use case | | ---| ---- | ---- | ---- | ----- |
| MiniCPM-V-2_6-Q2_K.gguf | Q2_K | 2 | 3.01 GB| smallest, significant quality loss not recommended for most purposes |
| MiniCPM-V-2_6-Q3_K_L.gguf | Q3_K_L | 3 | 4.09 GB| small, substantial quality loss |
| MiniCPM-V-2_6-Q3_K_M.gguf | Q3_K_M | 3 | 3.81 GB| very small, high quality loss |
| MiniCPM-V-2_6-Q3_K_S.gguf | Q3_K_S | 3 | 3.49 GB| very small, high quality loss |
| MiniCPM-V-2_6-Q4_0.gguf | Q4_0 | 4 | 4.43 GB| legacy; small, very high quality loss prefer using Q3_K_M |
| MiniCPM-V-2_6-Q4_K_M.gguf | Q4_K_M | 4 | 4.68 GB| medium, balanced quality recommended |
| MiniCPM-V-2_6-Q4_K_S.gguf | Q4_K_S | 4 | 4.46 GB| small, greater quality loss |
| MiniCPM-V-2_6-Q5_0.gguf | Q5_0 | 5 | 5.31 GB| legacy; medium, balanced quality prefer using Q4_K_M |
| MiniCPM-V-2_6-Q5_K_M.gguf | Q5_K_M | 5 | 5.44 GB| large, very low quality loss recommended |
| MiniCPM-V-2_6-Q5_K_S.gguf | Q5_K_S | 5 | 5.31 GB| large, low quality loss recommended |
| MiniCPM-V-2_6-Q6_K.gguf | Q6_K | 6 | 6.25 GB| very large, extremely low quality loss |
| MiniCPM-V-2_6-Q8_0.gguf | Q8_0 | 8 | 8.10 GB| very large, extremely low quality loss not recommended |
| MiniCPM-V-2_6-f16.gguf | f16 | 16 | 15.2 GB| |
*Quantized with llama.cpp b4120.*
header end -->
MiniCPM-V-2_6-GGUF
Original Model
openbmb/MiniCPM-V-2_6
Run with LlamaEdge
minicpmvtext
<|system|>
{system_message}<|end|>
<|user|>
{user_message_1}<|end|>
<|assistant|>
{assistant_message_1}<|end|>
<|user|>
{user_message_2}<|end|>
<|assistant|>
The
{user_message_n} has the format: {image_base64_encoding_string}\n{user_question}.128000bash
wasmedge --dir .:. \
--nn-preload default:GGML:AUTO:MiniCPM-V-2_6-Q5_K_M.gguf \
llama-api-server.wasm \
--prompt-template minicpmv \
--ctx-size 128000 \
--llava-mmproj mmproj-model-f16.gguf \
--model-name minicpmv-26
Quantized GGUF Models
| Name | Quant method | Bits | Size | Use case | | ---
*Quantized with llama.cpp b4120.*
Files & Weights
| Filename | Size | Action |
|---|---|---|
| MiniCPM-V-2_6-Q2_K.gguf | 2.81 GB | |
| MiniCPM-V-2_6-Q3_K_L.gguf | 3.81 GB | |
| MiniCPM-V-2_6-Q3_K_M.gguf | 3.55 GB | |
| MiniCPM-V-2_6-Q3_K_S.gguf | 3.25 GB | |
| MiniCPM-V-2_6-Q4_0.gguf | 4.13 GB | |
| MiniCPM-V-2_6-Q4_K_M.gguf | 4.36 GB | |
| MiniCPM-V-2_6-Q4_K_S.gguf | 4.15 GB | |
| MiniCPM-V-2_6-Q5_0.gguf | 4.95 GB | |
| MiniCPM-V-2_6-Q5_K_M.gguf | 5.07 GB | |
| MiniCPM-V-2_6-Q5_K_S.gguf | 4.95 GB | |
| MiniCPM-V-2_6-Q6_K.gguf | 5.82 GB | |
| MiniCPM-V-2_6-Q8_0.gguf | 7.54 GB | |
| MiniCPM-V-2_6-f16.gguf | 14.19 GB | |
| mmproj-model-f16.gguf | 0.97 GB |