Hacker News new | past | comments | ask | show | jobs | submit login

32k context is absolutely huge. There's all sorts of techniques for summarizing large documents down to get into 4k right now with 3.5, but it's incredibly lossy.

But boy, not cheap at all - $2 per api call on a 32k token document + whatever the output.

gpt-3.5-turbo is going to be around for a long time. At this price, your use case is going to need to be replacing a large cost center. Which based on their released results on common benchmarks, is absolutely going to happen.




3.5 might be their loss leader to keep people in their ecosystem for most use cases and to create a unique wall in terms of the training dataset they made via ChatGPT, GPT-4 they must be confident enough that nobody can compete that they can charge much more. Plus the use cases it can be used to replace cost centers like you said




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: