Model Overview

Model license: Llama-2<br> This model is trained based on NousResearch/Llama-2-7b-chat-hf model that is QLoRA finetuned on Photolens/oasst1-langchain-llama-2-formatted dataset.<br>

Prompt Template: Llama-2

<s>[INST] Prompter Message [/INST] Assistant Message </s>

Intended Use

Dataset that is used to finetune base model is optimized for langchain applications.<br> So this model is intended for a langchain LLM.

Training Details

This model took 1:14:16 to train in QLoRA on a single A100 40gb GPU.<br>

Models in this series

Model Train time Size (in params) Base Model
llama-2-7b-langchain-chat 1:14:16 7 billion NousResearch/Llama-2-7b-chat-hf
llama-2-13b-langchain-chat 2:50:27 13 billion TheBloke/Llama-2-13B-Chat-fp16
Photolens/OpenOrcaxOpenChat-2-13b-langchain-chat 2:56:54 13 billion Open-Orca/OpenOrcaxOpenChat-Preview2-13B