Skip to content

Extract a single expert from a Mixture Of Experts model using slerp interpolation.

License

Notifications You must be signed in to change notification settings

cognitivecomputations/extract-expert

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

11 Commits
 
 
 
 
 
 

Repository files navigation

Uses slerp to extract a single expert from multiple experts in a MoE model. This process requires a significant amount of system memory and can take some time. Be patient and make sure you have about five times the system RAM as the model's parameter count from which you're extracting. -Lucas Atkins (Crystalcareai)

Example usage:

python extract.py --model-name mistralai/Mixtral-8x7B-v0.1 --output-dir ./out

Note

Only works for models using Mixtral architecture. There is no guarantee that we'll expand this to support other architectures in the future.

About

Extract a single expert from a Mixture Of Experts model using slerp interpolation.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages