Hello there,
I’m a student and me and my team have the assignment to make a chatbot for our university. We need to make a chatbot that can help other students find information about their course. We will get our data from manuals of multiple universtity websites (as pdf). This data will be turned into Q&A data using ChatGPT 4.
However, we are struggling to find a pre-trained LLM that fits our assignment. We’ve researched T5, BERT and GPT-2 but our teacher was surprised those were the models we researched, since there are more popular and newer models. Our chatbot must be in Dutch, but we can translate so the LLM doesn’t need to be trained on Dutch data. The LLM can’t be too big, because we don’t have the hardware for very large models.
My question is: is LLaMa a good LLM for making a chatbot?
I agree with finetuning + RAG, given that OP already seems to have Q&A pairs, so it should be a great starting point as a dataset.
The language (Dutch <-> English) could possibly be a barrier for reasonable performance with Llama or any other 7B model, but as OP stated they might be able to use translation for that. I’m not sure whether DeepL could be used for that, i.e., using the DeepL API as a wrapper around the code for user input and chatbot output. It should have pretty good perfomance for Dutch. I like the idea and would like to test this or see the results when properly implemented. So please keep us updated on your approach u/Flo501