Submitted by singularpanda t3_1060gfk in MachineLearning
singularpanda OP t1_j3eohh7 wrote
Reply to comment by currentscurrents in [D] Will NLP Researchers Lose Our Jobs after ChatGPT? by singularpanda
Yes, it is quite costy. However, it seems not easy to modify it in our research as it is not open.
KBM_KBM t1_j3g7swj wrote
https://github.com/lucidrains/PaLM-rlhf-pytorch
Similar to chat get architecture you can play with this
singularpanda OP t1_j3gdv9p wrote
Thanks! Yes, there are many similar things. But the ChatGPT seems to have the most amazing performance.
Think_Olive_1000 t1_j3tnqyd wrote
I feel like you'd make a really bad research student
KBM_KBM t1_j3gere2 wrote
True but practically training a gpt model is not computationally cheap. I think instead of making such generalized language models we need to focus more one subject specific language models.
f_max t1_j3frhxs wrote
Megawatt sounds right for training. But kilowatts for inference. Take a look at tim dettmer’s work (he’s at UW) on int8 to see some of this kind of efficiency work. There’s definitely significant work happening in the open.
Viewing a single comment thread. View all comments