https://huggingface.co/deepseek-ai/deepseek-llm-67b-chat
https://huggingface.co/deepseek-ai/deepseek-llm-67b-base
Knowledge cutoff May 2023, not bad.
Online demo: https://chat.deepseek.com/ (Google oauth login)
another Chinese model, demo is censored by keywords, not that censored on local.
I made it write about itself using LocalAI https://sfxworks.net/posts/deepseek/
I will post a how-to on using local-ai on my free time if anyone is interested
not that censored on local.
So… Some censoring?
I’m desensitized at this point. I wonder if this is yet another Pretraining on the Test Set Is All You Need marketing stunt or not, as most new models lately have been.
I wish there was a 13b model which can just fit in on my GPU with quant
Does it give refusals on base? 67B sounds like full foundation train.
GGUF via TheBloke: