a_beautiful_rhindB to LocalLLaMA@poweruser.forumEnglish · 2 years agoYet another 120b. Trained on limarp.huggingface.coexternal-linkmessage-square1linkfedilinkarrow-up11arrow-down10
arrow-up11arrow-down1external-linkYet another 120b. Trained on limarp.huggingface.coa_beautiful_rhindB to LocalLLaMA@poweruser.forumEnglish · 2 years agomessage-square1linkfedilink
minus-squarettkciarBlinkfedilinkEnglisharrow-up1·2 years agoThis is a fine-tune of Goliath-120B. Didn’t the author hypothesize that Goliath’s interleaving of rows would degrade inference quality until it had been fine-tuned? It will be interesting to see if this fine-tune supports that hypothesis. Waiting for GUFF.
This is a fine-tune of Goliath-120B.
Didn’t the author hypothesize that Goliath’s interleaving of rows would degrade inference quality until it had been fine-tuned?
It will be interesting to see if this fine-tune supports that hypothesis.
Waiting for GUFF.