Hey everyone, I’m looking for a way to use an open source local large language model (LLM) on Linux, particularly on low-spec hardware like Raspberry Pi, to generate lengthy, coherent stories of 10k+ words from a single prompt. I recall reading about methods described in scientific papers such as “Re3: Generating Longer Stories With Recursive Reprompting and Revision”, announced in this Twitter thread from October 2022 and “DOC: Improving Long Story Coherence With Detailed Outline Control”, announced in this Twitter thread from December 2022. These papers used GPT-3, and since it’s been a while since then, I was hoping there might be something similar made using only open source tools. Does anyone have experience with this or know of any resources that could help me achieve long, coherent story generation with an open source LLM? Any advice or pointers would be greatly appreciated. Thank you!

  • atzanteol@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    18
    ·
    9 months ago

    Raspberry Pi

    This is a famously high-compute problem and you want to chuck it on a pi? Most llm models require a good GPU.

    • Mixel@feddit.de
      link
      fedilink
      arrow-up
      4
      ·
      9 months ago

      Just want to piggyback this. You will probably need more than 6gb vram to run good enough models with a acceptable speed and coherent output, but the more the better.

      • exu@feditown.com
        link
        fedilink
        English
        arrow-up
        1
        ·
        9 months ago

        Using a CPU is possible thanks to llama,cpp, but you’d still want more RAM than the Pi offers.