# Faint tecnnique, crest-e2 clash-e1 # # review: # - Instruction-following: # - Swerve: # - Word choice: # - Rhythm, cadence: # - Notes: # - # # - Design: # The idea here is to cut crush -- formerly the very cornerstone # of our merges -- completely out. it's very good for word choice # but crest is, too. The only problem is I seem to remember that # crest is overfit. So, we make it faint. # # Note: nearly two years later I'm trying to bring Mixtral # back from the dead. There are multiple reasons: # 1. Mistral-Small is kind of crap and smells like slop. # Hell, even the comprehension felt weak but maybe that's # just how I tried to sample it. # 2. Llama3 hasn't been interesting and is definitely crammed # with slop. # 3. Mixtral is probably the least synthetic-trained sounding # of all the OG models. Even when I tried the Quen shit # it seemed to be just openai. Mixtral is still sloppy. # # So, the pieces that are ours are uphill: non-instruct lora # being applied to the instruct rawdog without an intermediate # step. # # Obviously we're using pure elemental antisoc loras, hush's shit # but not her merge because the merges aren't "uphill", as in, # a lora made with "mixtral non-instruct" applied straight to # the instruct with loraize. # # The notion, which came to me in the middle of the night, is # to have the hush loras be only barely present layer-wise but # weighted heavily. Likewise with LimaRP, send uphill from # doctor-shotgun's qlora straight into mixtral-instruct # # My hypothesis is that we should get really fucking close to # pure-ass mixtral-instruct in terms of attention, but that # we're weighting really hard not to write like it. I have no # idea if that's how it works--I'm a fucking caveman. # # What I'm given to understand, and I'm way out of my depth, # is that the antisoc layers won't have blotched the instruct # as badly as they usually do, but when they're triggered they # are dominant. It's entirely possible I've got no idea what # I'm saying. # Model descriptions: # - crush: poetry; we have all checkpoints # - crest: fic; we only have e2 for this # - clash: novels (I think); we have all checkpoints for 0.2 models: # I wonder what happens if we just hurl this out the window # - model: mistralai/Mixtral-8x7B-Instruct-v0.1 # parameters: # density: 0.9 # weight: 0.55 # # crest is fic - model: ./uphill-instruct-crest-e2-nolime # i found lima in this, I need to cook another parameters: density: 0.4 weight: 0.3 # This is actually an uphill lima but I didn't name it that way. - model: ./Mixtral-8x7B-Yes-Instruct-LimaRP parameters: # Still just a breath of layers from the thing density: 0.2 # I am gimping its weight compared to hush tunes because limarp has too # much ai-slop and amateur-smut cliche slop. Honestly, if there were # something better than limarp I'd try to train it myself but I don't # know if there is. weight: 0.1 # Pure uphill clash at e2. Also more weight. - model: ./uphill-pure-clash-0.2-e2 parameters: density: 0.5 weight: 0.6 # della sucked ass so dare_ties it is merge_method: dare_ties # I know all of these look like instruct but the lora # is actually not so we go to the base base base_model: mistralai/Mixtral-8x7B-v0.1 parameters: normalize: true int8_mask: true dtype: bfloat16