Pygmalion 6B
Model description
Pymalion 6B is a proof-of-concept dialogue model based on EleutherAI's GPT-J-6B.
Training data
The fine-tuning dataset consisted of 500MB of dialogue data gathered from multiple sources, which includes both real and partially machine-generated conversations.
The training datasets:
- dolly_v2
- sharegpt
- unified_abstract_infill
- unified_chip2
- unified_hc3_human
- unified_unifiedskg_instructions
- Anh datasets
- oa_v3_fixed_plus_safety
- oasst_ready
Intended use
The model can be used as a regular text generation model, but it'll perform best if the input prompt adheres to the following format:
A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions. USER: Hello ASSISTANT:
Known issues
We haven't played around with the model enough to enumerate them. Feel free to give us some feedback!