Submitted by ravik_reddit_007 t3_zzitu1 in technology
EternalNY1 t1_j2cd4hm wrote
Reply to comment by go_comatose_for_me in There's now an open source alternative to ChatGPT, but good luck running it by ravik_reddit_007
They still estimate $87,000 per year on the low end to operate it yearly on AWS for 175 billion parameters.
I am assuming that is just the cost to train it though so it would be a "one time" cost every time you decided to train it.
Not exactly cheap, but something can can be budgeted for larger companies.
I asked it specifically how many GPUs it uses, and it replied with:
>For example, the largest version of GPT-3, called "GPT-3 175B," is trained on hundreds of GPUs and requires several dozen GPUs for inference.
aquamarine271 t1_j2ckpo1 wrote
That’s it? Companies pay like at least 100k a year on shitty business intelligence server space that is hardly ever used.
wskyindjar t1_j2cly2m wrote
seriously. Chump change for any company that could benefit from it in any way
aquamarine271 t1_j2cmsny wrote
This guy should put a deck together on the source of this 87k/yr and make it public if he wants every mid sized+ company to be sold on the idea
[deleted] t1_j2dax93 wrote
[deleted]
[deleted] t1_j2ckqon wrote
[deleted]
Tiny_Arugula_5648 t1_j2d910s wrote
It costs much less and trains in a fraction of the time when you can use a TPU instead of a GPU on Google Cloud.. that’s how Google trained the BERT & T5 models..
Viewing a single comment thread. View all comments