So far, I have experimented with the following projects:

https://github.com/huggingface/chat-ui - Amazing clean UI with very good web search, my go to currently.

https://github.com/oobabooga/text-generation-webui - Best overall, supports any model format and has many extensions

https://github.com/ParisNeo/lollms-webui/ - Has PDF, stable diffusion and web search integration

https://github.com/h2oai/h2ogpt - Has PDF, Web search, best for files ingestion (supports many file formats)

https://github.com/SillyTavern/SillyTavern - Best for custom characters and roleplay

https://github.com/NimbleBoxAI/ChainFury - Has great UI and web search (experimental)

https://github.com/nomic-ai/gpt4all - Basic UI that replicated ChatGPT

https://github.com/imartinez/privateGPT - Basic UI that replicated ChatGPT with PDF integration

LM Studio - Clean UI, focuses on GGUF format

-

Really love them and wondering if there are any other great projects,

Some of them include full web search and PDF integrations, some are more about characters, or for example oobabooga is the best at trying every single model format there is as it supports anything.

What is your favorite project to interact with your large language models ?

Share your findings and il add them!

  • itsuka_devB
    link
    fedilink
    English
    arrow-up
    1
    ·
    10 months ago

    Thank you for the kind words! I stopped maintaining OAI UIs because things got a bit stagnant a few months ago (there were literally no new UIs for weeks, I think). But with the new features announced at DevDay, I’m expecting to see a surge in new UIs, especially those that leverages multi-modality. As a maintainer of list projects, this is such an exciting time.

    I think we can keep both at the meantime

    Absolutely. Someone needs to maintain an awesome list for local UIs, and I don’t think my projects fall into that category. Besides, my list got a mix of native UIs in there too, which is important to me since I’m building both web and native UIs.

    let me know what breakdown makes more sense from your end

    For local UIs, sorting/grouping by model backend (e.g. llama.cpp, ollama, ExLlama) makes the most sense, IMO - and the rest of what I mentioned above is optional.