Alien Top
  • Communities
  • Create Post
  • heart
    Support Lemmy
  • search
    Search
  • Login
  • Sign Up
ThistleknotB to LocalLLaMA@poweruser.forum · 2 years ago

Felladrin/TinyMistral-248M-Alpaca

huggingface.co

external-link
message-square
1
link
fedilink
1
external-link

Felladrin/TinyMistral-248M-Alpaca

huggingface.co

ThistleknotB to LocalLLaMA@poweruser.forum · 2 years ago
message-square
1
link
fedilink
Felladrin/TinyMistral-248M-Alpaca · Hugging Face
huggingface.co
external-link
We’re on a journey to advance and democratize artificial intelligence through open source and open science.
  • ThistleknotOPB
    link
    fedilink
    English
    arrow-up
    1
    ·
    2 years ago

    I was going to try to knowledge distill but they modified their tokenizer.

    Either way neo has a 125M model, so a 248M model is x2 that. I imagine this could be useful for shorter context tasks. Idk, or to continue training for very tight uses cases

    I came across it while looking for tiny mistral config jsons to replicate⁸

    https://preview.redd.it/l9l7a39u3a1c1.jpeg?width=720&format=pjpg&auto=webp&s=80589cb6fbb2268b0d8af65b4ec27647185b4780

LocalLLaMA@poweruser.forum

localllama@poweruser.forum

Subscribe from Remote Instance

Create a post
You are not logged in. However you can subscribe from another Fediverse account, for example Lemmy or Mastodon. To do this, paste the following into the search field of your instance: !localllama@poweruser.forum

Community to discuss about Llama, the family of large language models created by Meta AI.

Visibility: Public
globe

This community can be federated to other instances and be posted/commented in by their users.

  • 4 users / day
  • 4 users / week
  • 4 users / month
  • 4 users / 6 months
  • 3 local subscribers
  • 4 subscribers
  • 1.03K Posts
  • 5.96K Comments
  • Modlog
  • mods:
  • communick@poweruser.forum
  • BE: 0.19.11
  • Modlog
  • Instances
  • Docs
  • Code
  • join-lemmy.org