TL;DR:

Hey everyone, I am excited to share with you the first release of “DreamGen Opus”, an uncensored model that lets you write stories in collaborative fashion, but also works nicely for chat / (E)RP.

Specifically, it understands the following prompt syntax (yes, another one — please don’t hate :D):

(Description of the story, can also optionally include information about characters) 

...


(Instructions as you write the story, to guide the next few sentences / paragraphs)

You can find more details about prompting the model in the official prompting guide, including a few examples (like for chat / ERP).

The initial model is based on Mistral 7B, but Llama 2 70B version is in the works and if things go well, should be out within 2 weeks (training is quite slow :)).

The model is based on a custom dataset that has >1M tokens of instructed examples like the above, and order of magnitude more examples that are a bit less instructed.

How to try it out

The model should work great with any tool that supports the Mistral 7B base model. It will work well with oobabooga/text-generation-webui and many other tools. I like vLLM.

Using vLLM

  • Install vLLM following the instructions in the repo
  • Run python -u -m vllm.entrypoints.openai.api_server --host 0.0.0.0 --model dreamgen/opus-v0-7b

Using DreamGen.com website (free)

You can also try the model on dreamgen.com for free (but it requires a registration with email).

What’s next

I believe that for story telling & character creation it’s especially important to have access to the model weights, otherwise you run the risk of losing your plot or virtual companion (as already happened a few times before on various closed platforms that suddenly changed their rules or got shut down by their API provider). Hence DreamGen.

Here’s a high level overview of what I would like to do next under the DreamGen umbrella:

On the model side:

  • (Soon) Larger story models
  • Fine tune the model for even better character chat & roleplay
  • Longer context windows, at least for smaller models (8-16K depending on how experiments go)

On the application side, I am thinking about these features:

  • Character editor, chat & roleplay
  • Ability to share your stories privately & publicly (not sure about this one, to be honest :))
  • Image generation to go alongside with story generation & chat
  • API so that you can use the model more easily if you don’t have a GPU

For all of these, I would love your input! You can vote on the roadmap here.

For more updates, join the community server or follow updates on Twitter.