Discussion about this post

User's avatar
Andrew Bartholomew's avatar

Really enjoyed this post. I was wondering, as I read it, "Why not both?" AWS's origins were serving some of Amazon's consumer-facing business lines. Amazon was (is?) AWS's first and best customer, which ensured immediate demand for its most advanced features.

What would prevent a similar framing for OpenAI? Ie consumer-facing offerings in the short term push the backend forward as fast as possible, then rolling those features out at the platform level makes them widely available for recombination by developers. (And in the short term, keeping things more closed allows OpenAI to get a better handle on safety.)

Expand full comment
smakubi's avatar

This is a great read, Benn. But the logic breaks down when training and running an LLM stops being as expensive as building data centers :).

I am curious to hear what you think about OSS LLMs like Databrick’s Dolly that can do instruction following, brainstorming, and summarization just like ChatGPT.

BUT….

can be created for $30 using one server for 3 hours on a small dataset using a 2 year old open source base LLM as opposed to training for 100,000 GPU hours at ~$10 million price point.

Expand full comment
11 more comments...

No posts