An attempt at a Generalist Model, using chunks spliced from the Airoboros and OpenHermes dataset, and some of my own stuff which I have since discarded (chat stuff, basic shit)forgot i added CollectiveCognition to the mix.

Base Model used is a gradient merge between OpenHermes & Airoboros 2.2, lora trained on the two spliced datasets and my own one, which was reapplied to them (dumb move I know)

Tests: Uncensored, not bad at RP I Guess, it felt decent. Good? who knows, tbh. try it urself or smth