☡
@stultulo
Currently messing with something i'm calling llama-3.1-8b-daydream first we merged llama-3.1-8b and llama-3.1-8b-instruct using mergekit/SLERP like this, to create llama-3.1-8b-decimated: Layers 0–7: passthrough base only Layers 8–23: 90% base + 10% instruct Layers 24–31: 80% base + 20% instruct ...then merged that result with Aion-RP-Llama-3.1-8B like this, to make llama-3.1-8b-daydream: Layers 0–3: passthrough decimated only Layers 4–7: 94% decimated + 6% Aion-RP Layers 8–31: 98% decimated + 2% Aion-RP documenting here because i'm not wanting to upload anything to HuggingFace unless it's been thoroughly benchmarked and also proven to not be psycho, as far as LLMs go. So far, they're very endearing and mostly well-behaved, though. I love the tone.
0 reply
0 recast
3 reactions