⚠️ Warning: This model can produce narratives and RP that contain violent and graphic erotic content. Adjust your system prompt accordingly, and use Llama 3 chat template.

πŸ™ Cthulhu 70B v1 GGUF

Only the Q4_0 is available for now. This runs faster than Q4_K_M with low VRAM.

Update: I have added a few more essential quants, mirrored from mradermacher: IQ4_XS, IQ3_XXS, IQ2_XXS

Team mradermacher has uploaded several more quants, so if you need them in other sizes go here:

https://huggingface.co/mradermacher/Cthulhu-70B-v1-GGUF

https://huggingface.co/mradermacher/Cthulhu-70B-v1-i1-GGUF

Cthulhu

Downloads last month
1,093
GGUF
Model size
71B params
Architecture
llama
Hardware compatibility
Log In to add your hardware

2-bit

3-bit

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for Naphula/Cthulhu-70B-v1-GGUF

Quantized
(4)
this model