Replies: 2 comments 3 replies
-
@epolewsky did this in https://github.com/epolewski/EricLLM |
Beta Was this translation helpful? Give feedback.
-
I actually had my own version of this in mind before the recent PR. Not sure which direction to take this right now. I'm a little hesitant because the idea of easily creating Frankenstein models puts me off a little bit. I mean, I wouldn't be too surprised if HF starts banning merged models soon because they don't actually have an unlimited budget for hosting all the files. I've yet to hear a compelling reason for why Frankenstein models should work in the first place, and I've yet to see an objective (or blind) test showing that they do. But they're definitely crowding the space, so... idk. |
Beta Was this translation helpful? Give feedback.
-
I really liked the discussion on Reddit localllama about the potential of easily creating Frankenstein models using exllama.
I think that would open up new areas of research without huge compute and memory requirements.
Could we start a discussion on 1) minimal Proof of Concept code modifications to test the ideas, and if the results are positive, 2) how we could make the inference efficient.
On the first topic, where in codebase would I start?
Beta Was this translation helpful? Give feedback.
All reactions