Task’s comment about us not having 14k GPUs is a good one
The key thing for us is leveraging llama for the generalized training: let them spend to do general training
We will spend our resources on the specialized (hive) training instead
Task’s comment about us not having 14k GPUs is a good one
The key thing for us is leveraging llama for the generalized training: let them spend to do general training
We will spend our resources on the specialized (hive) training instead
You'll need 14K GPUs eventually, though...