metadata
license: cc-by-4.0
Experimental frankenmerge using multiple 7B models using the Dare-ties method.
Including jondurbin--bagel-dpo-7b-v0.1 to maywell--Synatra-7B-v0.3-RP for "part 1"
mlabonne--NeuralBeagle14-7B to openaccess-ai-collective/DPOpenHermes-7B-v2 for "part 2"
merged part 1 and part 2 together
then took the first 26 layers of FelixChao--WestSeverus-7B-DPO-v2 and added them before the 32 layers of part 3 to make the final model
seems to work well with alpaca for instructions, and chatML format for just normal conversation.