smc=# select model, sum(total_time_s)/count(*) from openai_chatgpt_log where time >= now() - interval '1 weeks' group by model;
model | ?column?
---------------+--------------------
gpt-4 | 64.17583870967742
gpt-3.5-turbo | 22.513887411945003
(2 rows)
smc=# select model, sum(total_time_s)/count(*) from openai_chatgpt_log where time >= now() - interval '8 weeks' and time <= now() - interval '7 weeks' group by model;
model | ?column?
---------------+--------------------
gpt-4 | 30.74102777777778
gpt-3.5-turbo | 10.309548475729441
(2 rows)
The times have more than doubled on average! (I checked and the average total tokens hasn't changed at all.) Does openai publish any stats about API response times?I also subscribed to ChatGPT plus, and anecdotally it does seem much faster than the API for us. So maybe OpenAI is increasingly throttling API access for customers who are not marked as special? I wonder if some API users get much faster response times?
Given how valuable LLM's are for products like ours, what does this mean for us? Does it mean that relying on api access as a longterm solution isn't a way to stay competitive? There are other LLM api providers like Anthropic (and potentially Google), but so far they are vaporware for us, since it's just waitlists forever.
This gives me new appreciation for the approach repl.it is taking of building their own open source models.
It sounds like they are just swamped with usage and are just trying to keep it working at all…