Submitted by Xeiristotle t3_10kyuhn in MachineLearning
[removed]
Submitted by Xeiristotle t3_10kyuhn in MachineLearning
[removed]
Do you know if any of those other models have APIs for queries/inference or finetuning? I've looked into GPT alternatives and most seem closed off.
Nothing that would beat Open AI's stuff (Google's stuff) is open for inference or finetuning from the public.
I think the best Open source alternative is this
https://github.com/THUDM/GLM-130B
https://huggingface.co/spaces/THUDM/GLM-130B
But it's not finetuned for instruction so you have to prompt/approach it like a text completer. And also you'll need a 4x3090 to get it running locally.
The best open source instruction finetuned models are the flan t5 models
https://huggingface.co/google/flan-t5-xxl
If you're not necessarily looking for open source but still actual alternatives that aren't just an API wraparound of GPT, you can try cohere
Good thing is that it's completely free for non commercial or non production use
or alephalpha
Not free but the pricing is decent and they have a visual language model as well. Something like flamingo
https://www.deepmind.com/blog/tackling-multiple-tasks-with-a-single-visual-language-model
Not at this time. Google says they're going to release some kind of LLM-based product this year though.
You're probably right, but has anyone built an updated set of benchmarks to compare chatgpt with Google's publicly released numbers? (Maybe yes? Maybe I'm out of the loop?) Chatgpt is sufficiently different than gpt3.5 that I think we'd need to rerun benchmarks to compare.
(And, of course, even if we did, there are open questions of potential data leakage--always a concern, but maybe an extra concern here, since it is unclear whether OpenAI would have prioritized that issue in chatgpt build out. Certainly would have been low on my list, personally.)
[removed]
Yum Lecum
why did you put that evil in my mind.
The willpower needed not to use that at every possible juncture might be a bit much for me.
[removed]
[removed]
https://blog.google/technology/ai/lamda/
It's supposed to be ahead of OpenAI's current GPT. Seems only logical to me since it was Google that invented and open-sourced the Transformer model and they likely have much more and much higher quality data than OpenAI for training.
chatGPT is probably a very good model on the task it had to solve (to be a great conversational agent based on openAI data), but there are better models regarding the broad task of language understanding. You could adapt these models to be conversational agents, and they could probably beat chatGPT if they had access to the same dataset. But it would still be this specific task of being a great conversational agent. It's not the task of "thinking by itself like humans".
So it depends on what "more advanced" means. There are probably more "advanced" tasks towards AGI. But towards being a great conversational agent perhaps openAI has the best task-dataset combo today. At least I'm quite sure that there aren't systems which would be "significantly" more advanced than that, because I think the current limit is that it's "just" a very good conversational agent.
This is a joke, but Visual Mod on the Wallstreet bets sub has been looking like a human for quite a while. It's shockingly good.
> Lecum
Dude wth
He's a very famous AI pioneer from rance.
Google definitely does. Chat GPT is based off of GPT3-175B, and google has put out several models that outperform it. Like 4 or 5 I think and each of those significantly outperforms it’s predecessor
ChatGPT is good at syntax, but it's worse than pretty much any rule-based system at logic or arithmetic. So depending on your task, something like IBM Watson could be considered more advanced because it has dedicated rule-based reasoning. All it takes for MS or Google to make a "more advanced" system is just couple a large language model with a logic engine.
Integrating wolfram with chat gpt api works pretty well but your point stands particularly when it comes to logic ; the models might differentiate integrate , work on combinatorics , graph theory ,hell alphatensor even found new linear algebra algorithms but none of em can do any true logic based activities like coming up with original proofs
Wait, has someone actually integrated Wolfram with ChatGPT? I thought it was still in the "would be cool" stage.
Google supposedly has better models based on benchmarks, but few people outside of Google has used them (and those that have used it don't seem to be giving good reviews).
AnthropicAI's Claude model seems promising as a ChatGPT competitor.
[deleted]
*Lecun. And their Galactica was subject of so much ridicule that after pompous launch it was in-launched 48 hours later. OPT-175B is a clone of OpenAI’s GPT3, but performs worth and is essentially a massive pain in the ass cyber-security and phishing/desinformation.
Lecun always was into CovNets for machine vision - text-to-text is Hinton, Bengio, and Sutskever.
So far it looks like Baidu and Google have bigger transformer-based models that could perform better, but only Google’s PaLM is architecturally different enough to potentially perform better.
There are also augmented variants of Transformer-based model that are capable of more factual response, but they tend to be less conversational.
At what?
Meh - now OpenAI has the best chat dataset available to mankind. Not sure if saying one is more advanced now will make any difference in a few years.
NOW it does, after a massive public beta-test :)
MysteryInc152 t1_j5tits4 wrote
Google has few systems that would beat current public SOTA models. PALM/Minerva/Med Palm is the best but Flamingo, Chinchilla/Sparrow would also best chatGPT.
Dunno about anything from meta. They have open source GPT models released but they're not as good as Open AI's stuff.