Hey everyone, I work as a scientist at Microsoft Research and I spend a lot of my time training LLMs. I am super excited about the progress of open source llms and have I am happy to share some thoughts and ideas on how to scale them to GPT4 level.

Find my video lecture on this Here

https://youtu.be/gWJj-6udLWU?si=AqiJ-PpTQMBJAAm3

I plan on sharing a lot more ideas and code tutorials on building foundation models, instruction finetuning and alignment.

  • FPhamB
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 year ago

    Excellent. Always nice to see big boys joining in. :)

  • xadiantB
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 year ago

    Really cool, will check the video out. Since we found an actually qualified person though, let me ask a few layman questions, hope you have time to answer them!

    sampling methods. Most of them look simple, but we still don’t really know how to tune them. Do you think novel sampling methods or specific combinations could improve output quality by a lot?

    For instance, beam search. Does beam search provide a linear improvement in quality as you go up or not?

    Do you think ideal numbers for temperature, top_k and top_p are context or model based, or both?