I’ve playing with a lot of models around 7B but I’m now prototyping something that would be fine with a 1B model I think, but there’s just Phi-1.5 that I’ve seen of this size, and I haven’t seen a way to run it efficiently so far. llama.cpp has still not implemented it for instance.

Anyone has an idea of what to use?

  • palpapeenOPB
    link
    fedilink
    English
    arrow-up
    1
    ·
    10 months ago

    I mean yeah but it’s not done training AFAIK, and not fine-tuned either