aitorrent/dolphin-2.9.2-Phi-3-Medium-abliterated-GGUF-torrent

Photo by Jason Leung on Unsplash

aitorrent/dolphin-2.9.2-Phi-3-Medium-abliterated-GGUF-torrent

https://huggingface.co/aitorrent/dolphin-2.9.2-Phi-3-Medium-abliterated-GGUF-torrent

ko-fi

About

static quants of huggingface.co/cognitivecomputations/dolphi..

weighted/imatrix quants are available at huggingface.co/mradermacher/dolphin-2.9.2-P..

Usage

If you are unsure how to use GGUF files, refer to one of TheBloke's READMEs for more details, including on how to concatenate multi-part files.

Provided Quants

(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants)

LinkTypeSize/GBNotes
GGUFQ2_K5.3
GGUFIQ3_XS5.9
GGUFQ3_K_S6.2
GGUFIQ3_S6.2beats Q3_K*
GGUFIQ3_M6.4
GGUFQ3_K_M6.9lower quality
GGUFQ3_K_L7.4
GGUFIQ4_XS7.7
GGUFQ4_K_S8.1fast, recommended
GGUFQ4_K_M8.5fast, recommended
GGUFQ5_K_S9.7
GGUFQ5_K_M10.0
GGUFQ6_K11.6very good quality
GGUFQ8_014.9fast, best quality

Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

image.png

And here are Artefact2's thoughts on the matter: gist.github.com/Artefact2/b5f810600771265fc..