cantfindaname2take
cantfindaname2take t1_j9qov0f wrote
Reply to comment by Seankala in [D] 14.5M-15M is the smallest number of parameters I could find for current pretrained language models. Are there any that are smaller? by Seankala
For simple NER tasks some simpler models might work too,like conditional random fields. The crfsuite package has a very easy to use implementation of it and it is using a C lib under the hood for the model training.
cantfindaname2take t1_j9qohtm wrote
Reply to comment by Friktion in [D] 14.5M-15M is the smallest number of parameters I could find for current pretrained language models. Are there any that are smaller? by Seankala
Yeah, I would try FastText before LLMs.
cantfindaname2take t1_j8rdyje wrote
Reply to comment by Cool-Refrigerator141 in [OC] Beyoncé now holds the record for most Grammy wins by any artist by giteam
Meh, I'm sure some would say society is in a decline since pop culture appeared.
cantfindaname2take t1_j8onl6h wrote
Reply to comment by Cool-Refrigerator141 in [OC] Beyoncé now holds the record for most Grammy wins by any artist by giteam
I think nowadays he is mostly known by music nerds, boomers and hip hop producers.
cantfindaname2take t1_j8omdij wrote
Reply to comment by terath in [D] Is anyone working on ML models that infer and train at the same time? by Cogwheel
This! Online learning is a very common term that is used in time series modeling, for example in anomaly or change point detection.
cantfindaname2take t1_j8m5sry wrote
Mostly implementing change point detection algorithms that in some way utilize ordinal pattern analysis.
cantfindaname2take t1_j8e4x9q wrote
Reply to comment by Cherubin0 in [D] Can Google sue OpenAI for using the Transformer in their products? by t0t0t4t4
No, it's not like that at all. IMO that analogy does not make any sense. First, r&d is not just thinking up stuff and then making them. In drug discovery it involves expensive trials. In other fields it may involve a lot of building and scraping things, sometimes from expensive material. Patent should be an incentive to do all that knowing that once that it's done it can monetized in a way that does not allow other companies just to copy and paste without effort. Should they be able to do it for everything and forever? Probably not and that is what I was referring to.
cantfindaname2take t1_j88yd4n wrote
Reply to comment by cajmorgans in [D] Can Google sue OpenAI for using the Transformer in their products? by t0t0t4t4
I dunno if a company paid for r&d then they should be entitled to exclusively make money from it for some time. I think the problem is around the what and the when and not the why.
cantfindaname2take t1_j3pvh1r wrote
cantfindaname2take t1_j2rch7q wrote
Reply to comment by DJ_laundry_list in [P] An old fashioned statistician is looking for other ways to analyse survival data - Is machine learning an option? by lattecoffeegirl
On that note there is a library that extends xgboost's parameters with survival analysis capabilities specifically. Here is a tutorial: https://loft-br.github.io/xgboost-survival-embeddings/how_xgbse_works.html
cantfindaname2take t1_j2810dh wrote
Reply to comment by elbiot in Laptop for Machine Learning by sifarsafar
This. I have the same setup of laptop + work station. It's been great. No model training kills my laptop anymore. More memory for Chrome.
cantfindaname2take t1_izjow8q wrote
Reply to comment by blablanonymous in [D] Product Recommendation Algorithm by RstarPhoneix
Theoretically yes, but it sounds like it requires more coding skills since you need to create the network structure by hand instead of relying on solutions that were specifically built for this. Don't get me wrong, there is nothing bad with what you described, but sounds to me OP needs a tool out of the box.
cantfindaname2take t1_izj23g5 wrote
Reply to comment by blablanonymous in [D] Product Recommendation Algorithm by RstarPhoneix
This already looks over complicated compared to the simply deployable and explainable alternatives that have been proven over time to be better at this problem for small datasets.
cantfindaname2take t1_iym5ipx wrote
Reply to comment by no_witty_username in [Discussion] - "data sourcing will be more important than model building in the era of foundational model fine-tuning" by fourcornerclub
Is it though? One thing that comes back up again is the comparison to human learning. Do humans get clean training samples? I like to think not that. Instead humans learn how to separate signal from noise much better, and also learn how to model hidden causes.
cantfindaname2take t1_iy01a0p wrote
Reply to [P] Memory Profiling for Pandas and Python by thapasaan
How does it compare to filprofiler? Apart from the IDE integration of course. Does it work with parallelized or threaded code? Threading support is especially interesting for numpy.
cantfindaname2take t1_iwuxv6x wrote
Foundations of Data Science is very good if you are looking for something that is pure math/stats
https://www.cambridge.org/core/books/foundations-of-data-science/6A43CE830DE83BED6CC5171E62B0AA9E
cantfindaname2take t1_iu9uyhv wrote
Reply to comment by biggieshiba in [D] What are the bottlenecks in your ML project lifecycle? What tools would you like to see more widely used? by Fine-Topic-6127
In my experience it's much easier to serve models through either AWS Lambda with containers or just upload your model in a container to an EC2 instance.
cantfindaname2take t1_is7b59h wrote
Reply to [R] Clustering a set of graphs by No_Performer203
If your graphs are usually small then you might want to look into Graph Kernels: https://ysig.github.io/GraKeL/0.1a8/documentation/introduction.html
cantfindaname2take t1_ir2vwe7 wrote
Reply to comment by bilby_- in [discussion] Is the future of ML development in low code tools? by bilby_-
IMHO low code is fine for not resource intensive tasks. Once an analysts (without too much programming skills) starts running into performance or memory problems then that person will have a hard time adjusting their pipeline because the low code tools usually offer little flexibility.
cantfindaname2take t1_jb42qee wrote
Reply to comment by ggdupont in To RL or Not to RL? [D] by vidul7498
Isn't it extensively used in robotics??