Model Card for Carpincho-13b

<!-- Provide a quick summary of what the model is/does. -->

This is Carpincho-13B an Instruction-tuned LLM based on LLama-13B. It is trained to answer in colloquial spanish Argentine language. It's based on LLama-13b (https://huggingface.co/decapoda-research/llama-13b-hf).

Model Details

The model is provided in ggml format, for use with the llama.cpp CPU-only LLM inference (https://github.com/ggerganov/llama.cpp)

Usage

Clone the llama.cpp repository:

git clone https://github.com/ggerganov/llama.cpp

Compile the tool:

make

Download the file carpincho-13b-ggml-model-q4_0.bin into the llama.cpp directory and run this command:

./main -m ./carpincho-13b-ggml-model-q4_0.bin -i -ins -t 4

Change -t 4 to the number of physical CPU cores you have.

This model requires at least 8GB of free RAM. No GPU is needed to run llama.cpp.

Model Description

<!-- Provide a longer summary of what this model is. -->

Model Sources [optional]

<!-- Provide the basic links for the model. -->

Uses

<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> This is a generic LLM chatbot that can be used to interact directly with humans.

Bias, Risks, and Limitations

<!-- This section is meant to convey both technical and sociotechnical limitations. --> This bot is uncensored and may provide shocking answers. Also it contains bias present in the training material.

Recommendations

<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->

Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model.

Model Card Contact

Contact the creator at @ortegaalfredo on twitter/github