llama-2 code

<p><h1> speechless-tora-code-7b-v1.0 </h1></p>

Use the following dataset to fine-tune llm_agents/tora-code-7b-v1.0 in order to improve the model's reasoning and planning abilities.

Total 201,981 samples.

HumanEval

Metric Value
humaneval-python 51.829

Big Code Models Leaderboard

CodeLlama-34B-Python: 53.29

CodeLlama-34B-Instruct: 50.79

CodeLlama-13B-Instruct: 50.6

CodeLlama-34B: 45.11

CodeLlama-13B-Python: 42.89

CodeLlama-13B: 35.07

LM-Evaluation-Harness

Open LLM Leaderboard

Metric Value
ARC 42.66
HellaSwag 65.16
MMLU 38.56
TruthfulQA 42.06
Average 47.11

Parameters

lr 2e-4
lr_scheduler_type cosine
weight_decay 0.0
optim paged_adamw_8bit
flash_attention True
rerope False
max_new_tokens 4096
num_train_epochs 2
bits 4
lora_r 64
lora_alpha 16
lora_dropout 0.05
double_quant True
quant_type nf4
dataset_format airoboros
mini_batch_size 2
grandient_accumulation_steps 32
bf16 True

A800-80G x 2

epoch 2.0
etrain_loss 0.5891
etrain_runtime 19:24:49.43
etrain_samples_per_second 5.664
etrain_steps_per_second 0.044
eeval_loss 0.5872
eeval_runtime 0:00:15.59
eeval_samples_per_second 12.822
eeval_steps_per_second 6.411