More info?

#1
by totally-not-an-llm - opened

This is very cool, could we get some info on how this was created, plus any scripts used?

yes please

Hey, thanks for the interest! I've added the script I used to generate the base model to the repo (frankenllama_22.py).
This actually came out of some experiments I was doing with attention head pruning. I decided to try going the other direction instead, and it's looking pretty promising so far.

For the fine tuning, I used axolotl: https://github.com/OpenAccess-AI-Collective/axolotl

@chargoddard Thanks for posting the script, I'm going to experiment with it. Do you know if it's possible to transplant heads from l2-70b instead of l1-33b like in the original script? And does the script need any changing other than pointing to the right donor?

I can't find this github repo, could you link it?

I can't find this github repo, could you link it?

@Vezora Do you mean the merge script? It's the .py file in the files section of this model.

I can't find this github repo, could you link it?

@Vezora Do you mean the merge script? It's the .py file in the files section of this model.

That's embarrassing, thank you!

Your need to confirm your account before you can post a new comment.

Sign up or log in to comment