chemistry biology molecule instructions

This repo contains a low-rank adapter for LLaMA2-7b-chat, trained on the ๐Ÿฅผ biomolecule text instructions from the ๐Ÿงช Mol-Instructions dataset.

Instructions for running it can be found at https://github.com/zjunlp/Mol-Instructions.

Please refer to our paper for more details.

image.png

<h3> ๐Ÿฅผ Tasks</h3>

<details> <summary><b>Chemical entity recognition</b></summary>

</details>

<details> <summary><b>Chemical-disease interaction extraction</b></summary>

</details>

<details> <summary><b>Chemical-protein interaction extraction</b></summary>

</details>

<details> <summary><b>Multiple-choice question</b></summary>

</details>

<details> <summary><b>True or False question</b></summary>

</details>

<details> <summary><b>Open question</b></summary>

</details>

<h3> ๐Ÿ“ Demo</h3>

As illustrated in our repository, we provide an example to perform generation.

>> python generate.py \
    --CLI True \
    --protein False\
    --load_8bit \
    --base_model $BASE_MODEL_PATH \
    --lora_weights $FINETUNED_MODEL_PATH \

Please download Llama-2-7b-chat to obtain the pre-training weights of LlamA-2-7b-chat, refine the --base_model to point towards the location where the model weights are saved.

For model fine-tuned on biomolecular text instructions, set $FINETUNED_MODEL_PATH to 'zjunlp/llama2-molinst-molecule-7b'.

<h3> ๐Ÿšจ Limitations</h3>

The current state of the model, obtained via instruction tuning, is a preliminary demonstration. Its capacity to handle real-world, production-grade tasks remains limited.

<h3> ๐Ÿ“š References</h3> If you use our repository, please cite the following related paper:

@article{molinst,
  title={Mol-Instructions: A Large-Scale Biomolecular Instruction Dataset for Large Language Models},
  author={Fang, Yin and Liang, Xiaozhuan and Zhang, Ningyu and Liu, Kangwei and Huang, Rui and Chen, Zhuo and Fan, Xiaohui and Chen, Huajun},
  journal={arXiv preprint arXiv:2306.08018},
  year={2023}
}

<h3> ๐Ÿซฑ๐Ÿปโ€๐Ÿซฒ๐Ÿพ Acknowledgements</h3>

We appreciate LLaMA-2, LLaMA, Huggingface Transformers Llama, Alpaca, Alpaca-LoRA, Chatbot Service and many other related works for their open-source contributions.