Training a massive LLM on the scale of GPT-4 required a lot of lead time (less so nowadays due to various optimizations), so the timeframe makes sense.
Training a massive LLM on the scale of GPT-4 required a lot of lead time (less so nowadays due to various optimizations), so the timeframe makes sense.