Model Overview
Model license: Llama-2<br> This model is trained based on TheBloke/Llama-2-13B-Chat-fp16 model that is QLoRA finetuned on Photolens/oasst1-langchain-llama-2-formatted dataset.<br>
Prompt Template: Llama-2
<s>[INST] Prompter Message [/INST] Assistant Message </s>
Intended Use
Dataset that is used to finetune base model is optimized for langchain applications.<br> So this model is intended for a langchain LLM.
Training Details
This model took 2:50:27 to train in QLoRA on a single A100 40gb GPU.<br>
- epochs:
1 - train batch size:
8 - eval batch size:
8 - gradient accumulation steps:
1 - maximum gradient normal:
0.3 - learning rate:
2e-4 - weight decay:
0.001 - optimizer:
paged_adamw_32bit - learning rate schedule:
cosine - warmup ratio (linear):
0.03
Models in this series
| Model | Train time | Size (in params) | Base Model |
|---|---|---|---|
| llama-2-7b-langchain-chat | 1:14:16 | 7 billion | NousResearch/Llama-2-7b-chat-hf |
| llama-2-13b-langchain-chat | 2:50:27 | 13 billion | TheBloke/Llama-2-13B-Chat-fp16 |
| Photolens/OpenOrcaxOpenChat-2-13b-langchain-chat | 2:56:54 | 13 billion | Open-Orca/OpenOrcaxOpenChat-Preview2-13B |