It may not be exactly as good, but it's also a 7B parameter model vs 175B parameters for GPT-3.5 (i.e., text-davinci-003). People are running the new model on their phones, laptops, and Raspberry PIs.
People's mouths were watering over the commercial implications of the recent 90% drop in cost for the new ChatGPT model. Now imagine if you can get similar performance on a model that requires <5% of the parameters.
People's mouths were watering over the commercial implications of the recent 90% drop in cost for the new ChatGPT model. Now imagine if you can get similar performance on a model that requires <5% of the parameters.