pytorch llama llama-2 alpaca chinese 中文

LLaMA-v2-chinese-alpaca-13B-GGML (ymcui)

Here are the GGML converted and/or quantized models for ymcui's Chinese LLaMA-v2 Alpaca 13B.

!NOTE! The GGML filetype is outdated. Prefer GGUF format going forward.

Explanation of quantisation methods

<details> <summary>Click to see details</summary>

Methods:

The new methods available are:

This is exposed via llama.cpp quantization types that define various "quantization mixes" as follows:

Provided files

Name Quant method Bits Size Max RAM required Use case
llama-v2-chinese-alpaca-13B-Q2_K.ggml Q2_K 2 5.65 GB 8.15 GB smallest, significant quality-loss - not recommended for most purposes
llama-v2-chinese-alpaca-13B-Q3_K_S.ggml Q3_K_S 3 5.81 GB 8.31 GB very small, high quality-loss
llama-v2-chinese-alpaca-13B-Q3_K_M.ggml Q3_K_M 3 6.46 GB 7.96 GB very small, high quality-loss
llama-v2-chinese-alpaca-13B-Q3_K_L.ggml Q3_K_L 3 7.08 GB 9.58 GB small, substantial quality-loss
llama-v2-chinese-alpaca-13B-Q4_0.ggml Q4_0 4 7.53 GB 10.03 GB legacy; small, very high quality-loss - prefer using Q3_K_M
llama-v2-chinese-alpaca-13B-Q4_1.ggml Q4_1 4 8.34 GB 10.84 GB legacy; small, very high quality-loss - prefer using Q3_K_M
llama-v2-chinese-alpaca-13B-Q4_K_S.ggml Q4_K_S 4 7.53 GB 10.03 GB small, greater quality-loss
llama-v2-chinese-alpaca-13B-Q4_K_M.ggml Q4_K_M 4 8.03 GB 10.53 GB medium, balanced quality - recommended
llama-v2-chinese-alpaca-13B-Q5_0.ggml Q5_0 5 9.15 GB 11.65 GB legacy; medium, balanced quality - prefer using Q4_K_M
llama-v2-chinese-alpaca-13B-Q5_1.ggml Q5_1 5 9.96 GB 12.46 GB legacy; medium, balanced quality - prefer using Q4_K_M
llama-v2-chinese-alpaca-13B-Q5_K_S.ggml Q5_K_S 5 9.15 GB 11.65 GB large, low quality-loss - recommended
llama-v2-chinese-alpaca-13B-Q5_K_M.ggml Q5_K_M 5 9.41 GB 11.91 GB large, very low quality-loss - recommended
llama-v2-chinese-alpaca-13B-Q6_K.ggml Q6_K 6 10.9 GB 13.4 GB very large, extremely low quality-loss
llama-v2-chinese-alpaca-13B-Q8_0.ggml Q8_0 8 14 GB 16.5 GB very large, extremely low quality-loss - not recommended
llama-v2-chinese-alpaca-13B-f16.ggml f16 16 26.5 GB 29 GB very large, almost no quality-loss - not recommended

Model Sources