__lawless
__lawless t1_ja9sgcx wrote
Yes we did. It was a pleasant experience.
__lawless t1_ja7hdy6 wrote
Reply to About this week's US stock market trend by BeyondShroud
Read what you just posted. You wrote a lot of words without actually saying anything.
__lawless t1_ja5wlxe wrote
__lawless t1_j9eu7n2 wrote
r/learnmachinelearning
__lawless t1_j97v07m wrote
Reply to [D] Relu + sigmoid output activation by mrwafflezzz
Easiest solution no sigmoid no relu in the last layer just clamp it between 0 and 1. Works surprisingly well
__lawless t1_j96ycry wrote
Reply to [R] [N] In this paper, we show how a conversational model, 3.5x smaller than SOTA, can be optimized to outperform the baselines through Auxiliary Learning. Published in the ACL Anthology: "Efficient Task-Oriented Dialogue Systems with Response Selection as an Auxiliary Task." by radi-cho
Link?
__lawless t1_j9617qd wrote
Reply to [D] Is Google a language transformer like ChatGPT except without the G (Generative) part? by Lets_Gooo_123
You are not making any sense. Language transformer is not a thing. Google is a search engine, ChatGPT is a LLM. There is no quantum computing involved in chatGPT. It has nothing to do with it at all. I’m gonna leave you with a quote from Billy Madison.
what you've just said is one of the most insanely idiotic things I have ever heard. At no point in your rambling, incoherent response were you even close to anything that could be considered a rational thought. Everyone in this room is now dumber for having listened to it. I award you no points, and may God have mercy on your soul.
__lawless t1_j95ixov wrote
Reply to [D] Things you wish you knew before you started training on the cloud? by I_will_delete_myself
Use code-server (VS code in browser) it is amazing
__lawless t1_j926m4d wrote
Reply to comment by dojoteef in [D] Please stop by [deleted]
Can you add some rules to not let 1 day old accounts post. Also not let people post immediately after joining.
__lawless t1_j88tevq wrote
Reply to comment by AsIAm in [D] Have their been any attempts to create a programming language specifically for machine learning? by throwaway957280
Very true. It’s just a shame that it did not make it. I was very looking forward to it
__lawless t1_j88fo2u wrote
I don’t think you have a sound idea what you are trying to do. So you want chatGPT + extra!!! What you are asking does not exist, at least currently. Making a model size of chatGPT will cost at the very least $5M and absolutely not possible locally. You need a distributed setup. Not to mention all the technical difficulties of making such a setup.
__lawless t1_j86xhtr wrote
Reply to comment by wittfm in [D] Have their been any attempts to create a programming language specifically for machine learning? by throwaway957280
That was swift for tensorflow. Did not pan out.
__lawless t1_j86qlmc wrote
Reply to [D] Have their been any attempts to create a programming language specifically for machine learning? by throwaway957280
Swift for tensorflow. Didn’t workout though
__lawless t1_j7ii2h1 wrote
Reply to comment by aicharades in [P] ChatGPT without size limits: upload any pdf and apply any prompt to it by aicharades
I see it is not a prompt per se, it is an analogue of map operation in ETL.
__lawless t1_j7ihgh5 wrote
Can you explain what map prompt is? Not sure I understood that part
__lawless t1_j76xpgk wrote
Reply to comment by jaqws in [R] Multimodal Chain-of-Thought Reasoning in Language Models - Amazon Web Services Zhuosheng Zhang et al - Outperforms GPT-3.5 by 16% (75%->91%) and surpasses human performance on ScienceQA while having less than 1B params! by Singularian2501
Just 2 points a) They fine tuned this model to death. Where as GPT3.5 has a handful of examples to fine tune b) This is a multi modal model which consumes the image directly. Where as GPT can only consume text, so they fed it caption of the image
__lawless t1_j76wlwb wrote
Reply to comment by zbyte64 in [R] Multimodal Chain-of-Thought Reasoning in Language Models - Amazon Web Services Zhuosheng Zhang et al - Outperforms GPT-3.5 by 16% (75%->91%) and surpasses human performance on ScienceQA while having less than 1B params! by Singularian2501
They did it on 4 V100 with 32GB RAM
__lawless t1_j76vq7h wrote
Reply to [R] Multimodal Chain-of-Thought Reasoning in Language Models - Amazon Web Services Zhuosheng Zhang et al - Outperforms GPT-3.5 by 16% (75%->91%) and surpasses human performance on ScienceQA while having less than 1B params! by Singularian2501
Just finished reading. Although imho not a very fair comparison with GPT it still is super impressive
__lawless t1_j4r9ebs wrote
Reply to comment by monkeysingmonkeynew in [D] Is it possible to update random forest parameters with new data instead of retraining on all data? by monkeysingmonkeynew
Ok let me elaborate a bit. Imagine the old model is called m_0
. Your newly obtained training data is X
, y
, features and labels, respectively. Now calculate the residual error which is the difference between y
and prediction of m_0
: dy = y - m_0(X)
.
Now train a new model m_1
. The labels and features are X
, dy
. Finally at inference time the prediction is the sum of the two models: y_pred = m_0(X_new) + m_1(X_new)
.
__lawless t1_j4pzotl wrote
Reply to [D] Is it possible to update random forest parameters with new data instead of retraining on all data? by monkeysingmonkeynew
A lot of folks here already mentioned online learning and the resources for it. However I am going to offer a very hacky solution inspired by idea of boosting. Suppose you had a regression model already trained. Make prediction for the new training batch and calculate the errors. Now train a new random forest model for the residual errors. For inference pass the features into the first model. For inference just pass the features to both models and sum the results.
__lawless t1_j4muji7 wrote
Reply to [D] I’m a Machine Learning Engineer for FAANG companies. What are some places I can get started doing freelance work for ML? by doctorjuice
I am a data scientist in one of the FAANGs and feel the same. I went looking for freelance work on some websites, but they pay so little. I would be curious on how to find freelance work too.
__lawless t1_j0yuzij wrote
Reply to comment by Awekonti in [D] Deep Learning based Recommendation Systems by Awekonti
It really depends on the problem/situation. Of course MF is very powerful and interpretable despite being very simple. But for my job we did not have success with it, on the other hand we had great success with transformer based sequential recommendation models. Another method you want to look into is GNN. We did not invest it in because it is harder (not impossible) to scale. For example look into pinsage
__lawless t1_j0ymvpl wrote
There are some methods that are matrix factorization free like sequential recombination models
__lawless t1_jbadl23 wrote
Reply to [R] Where can I train a deep learning algorithm with a $1 million budget? by coderdd
Yes