Something I find really fascinating is the gap between GPT-3 and GPT-4. Why is it that dozens of companies are now capable of training a model better than GPT-3, but none have matched GPT-4? Clearly it isn't a problem of scale/capital - otherwise Google would have done it.
Yeah, I'm really looking forward to understanding the reason for this - I'm pretty sure we'll know within the next six months, someone's bound to have a GPT-4 beater by then, and hopefully they'll share at least some of the details of what it took to build.
Something I find really fascinating is the gap between GPT-3 and GPT-4. Why is it that dozens of companies are now capable of training a model better than GPT-3, but none have matched GPT-4? Clearly it isn't a problem of scale/capital - otherwise Google would have done it.
Yeah, I'm really looking forward to understanding the reason for this - I'm pretty sure we'll know within the next six months, someone's bound to have a GPT-4 beater by then, and hopefully they'll share at least some of the details of what it took to build.
Super valuable insights! Thanks so much for sharing - and it kind of motivated me to start experimenting with running LLMs on my devices📱