Hey everyone, I work as a scientist at Microsoft Research and I spend a lot of my time training LLMs. I am super excited about the progress of open source llms and have I am happy to share some thoughts and ideas on how to scale them to GPT4 level.
Find my video lecture on this Here
https://youtu.be/gWJj-6udLWU?si=AqiJ-PpTQMBJAAm3
I plan on sharing a lot more ideas and code tutorials on building foundation models, instruction finetuning and alignment.
Excellent. Always nice to see big boys joining in. :)
Really cool, will check the video out. Since we found an actually qualified person though, let me ask a few layman questions, hope you have time to answer them!
sampling methods. Most of them look simple, but we still don’t really know how to tune them. Do you think novel sampling methods or specific combinations could improve output quality by a lot?
For instance, beam search. Does beam search provide a linear improvement in quality as you go up or not?
Do you think ideal numbers for temperature, top_k and top_p are context or model based, or both?