ggml causal-lm gpt2 text-generation

image/png

This repository contains quantized conversions of the AI Dungeon 2 checkpoint, "model_v5".

For use with frontends that support GGML quantized GPT-2 models. This model works best with KoboldCpp's "Adventure" mode.

Last updated on 2023-09-23.

Model RAM usage (KoboldCpp) RAM usage (Oobabooga)
aid2classic-ggml-q4_0.bin 984.1 MiB 1.4 GiB
aid2classic-ggml-q4_1.bin 1.1 GiB 1.5 GiB
aid2classic-ggml-q5_0.bin 1.2 GiB 1.6 GiB
aid2classic-ggml-q5_1.bin 1.2 GiB 1.7 GiB
aid2classic-ggml-q8_0.bin 1.7 GiB 2.2 GiB
aid2classic-ggml-f16.bin 3.2 GiB 3.6 GiB

Description:

Notes: