LLM Panther Transformers llama PyTorch Tensorboard Text Generation

<h1 style='text-align: center '>Panther</h1> <h2 style='text-align: center '><em>Rardilit Large Open-access Language Model</em> </h2> <h3 style='text-align: center '>Model Card</h3>

Panther Logo

Version 1.0 / 29.May.2023

Model Card for Bloom-560m

<!-- Provide a quick summary of what the model is/does. -->

Table of Contents

  1. Model Details
  2. Uses
  3. Bias, Risks, and Limitations
  4. Recommendations
  5. Training Details

Model Details

Model Description

This section provides information for anyone who wants to know about the model.

Uses

This section addresses questions around how the model is intended to be used, discusses the foreseeable users of the model (including those affected by the model), and describes uses that are considered out of scope or misuse of the model. It provides information for anyone considering using the model or who is affected by the model.

Intended Use

This model is being created in order to enable public research on large language models (LLMs). LLMs are intended to be used for language generation or as a pretrained base model that can be further fine-tuned for specific tasks. Use cases below are not exhaustive.

Direct Use

Downstream Use

Misuse and Out-of-scope Use

This section addresses what users ought not do with the model.

Out-of-scope Uses

Using the model in high-stakes settings is out of scope for this model.  The model is not designed for critical decisions nor uses with any material consequences on an individual's livelihood or wellbeing. The model outputs content that appears factual but is not correct.

Out-of-scope Uses Include:

Misuse

Intentionally using the model for harm, violating human rights, or other kinds of malicious activities, is a misuse of this model. This includes:

Intended Users

Direct Users

Indirect Users

Others Affected (Parties Prenantes)

Bias, Risks and Limitations

This section identifies foreseeable harms and misunderstandings.

Model may:

Recommendations

This section provides information on warnings and potential mitigations.

Training Details

This repo contains a low-rank adapter for LLaMA-7b with just 4194304 parameters fit on the Rardilit/Panther-dataset_v1 dataset with 20k prompts and responses.

This version of the weights was trained with the following hyperparameters:

Training Time

The time in training this model with 1 x T4 16gb vRAM was approx. 45 min.