So RWKV 7b v5 is 60% trained now, saw that multilingual parts are better than mistral now, and the english capabilities are close to mistral, except for hellaswag and arc, where its a little behind. all the benchmarks are on rwkv discor, and you can google the pro/cons of rwkv, though most of them are v4.

Thoughts?

  • MoffKalastB
    link
    fedilink
    English
    arrow-up
    1
    ·
    10 months ago

    Well it seems a lot better at Slovenian than LLamas or Mistral, especially for a 3B model, although it mostly just rambles about stuff that’s vaguely related to the prompt and makes lots of grammatical mistakes. The 7B one ought to be interesting once it’s done.

    • vatsadevOPB
      link
      fedilink
      English
      arrow-up
      1
      ·
      10 months ago

      Its trained on 100+ languages, the focus is multilingual

      • alchemist1e9B
        link
        fedilink
        English
        arrow-up
        1
        ·
        10 months ago

        Will that make it a good translator? I remember seeing somewhere a 400+ language translation model but not an LLM somewhere. Wonder what the best many language open source fast high quality translation solutions might look like.