Two phenomenon affect reliability of LLMs – Grokking- models could seemingly fail to learn a task and then all of a sudden just get it and double descent- when certain models got bigger, their error rate went down, then up—and then down again
We use cookies to provide the best website experience for you. If you continue to use this site we will assume that you are happy with it.OkayPrivacy policy