Model Card for Apollo-13B-pt

Model Details

Model Description

This model is a fine-tuned version of the Llama-2-13b-chat-hf model. It's trained to perform causal language modeling tasks. The fine-tuning is performed on a dataset named "guanaco-llama2-1k".

Model Sources

Uses

Direct Use

This model can be used for various NLP tasks like text generation, summarization, question-answering, etc.

Downstream Use

The model can be further fine-tuned for more specific tasks such as sentiment analysis, translation, etc.

Out-of-Scope Use

The model is not intended for generating harmful or biased content.

Bias, Risks, and Limitations

The model inherits the biases present in the training data and the base model. Users should be cautious while using the model in sensitive applications.

Recommendations

Users should evaluate the model for biases and other ethical considerations before deploying it for real-world applications.

How to Get Started with the Model

The model can be loaded using the Transformers library and can be used for generating text, among other things.

Environmental Impact

Technical Specifications

Model Architecture

The architecture is based on the Llama-2-13b-chat-hf model with causal language modeling as the primary objective.

Training Details

QLoRA Parameters

Bitsandbytes Parameters

Additional Features

Evaluation

Testing Data

[Soon...]

Factors

[Soon...]

Metrics

[Soon...]